From 2a4977a102775883ef4ea6f91a7b0ef403beee80 Mon Sep 17 00:00:00 2001
From: auphelia <jakobapk@web.de>
Date: Thu, 28 May 2020 11:20:54 +0100
Subject: [PATCH] [Transformation] Ignore "dw-layers" during hls conversion

---
 src/finn/transformation/fpgadataflow/convert_to_hls_layers.py | 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)

diff --git a/src/finn/transformation/fpgadataflow/convert_to_hls_layers.py b/src/finn/transformation/fpgadataflow/convert_to_hls_layers.py
index 3ff86cab4..f986cce1f 100644
--- a/src/finn/transformation/fpgadataflow/convert_to_hls_layers.py
+++ b/src/finn/transformation/fpgadataflow/convert_to_hls_layers.py
@@ -45,7 +45,7 @@ class InferConvInpGen(Transformation):
         graph_modified = False
         for n in graph.node:
             node_ind += 1
-            if n.op_type == "Im2Col":
+            if n.op_type == "Im2Col" and not getCustomOp(n).get_nodeattr("dw"):
                 i2c_input = n.input[0]
                 i2c_output = n.output[0]
                 i2c_in_shape = model.get_tensor_shape(i2c_input)
@@ -282,7 +282,7 @@ class InferQuantizedStreamingFCLayer(Transformation):
         graph_modified = False
         for n in graph.node:
             node_ind += 1
-            if n.op_type == "MatMul":
+            if n.op_type == "MatMul" and model.get_tensor_sparsity(n.input[1]) is None:
                 mm_input = n.input[0]
                 mm_weight = n.input[1]
                 mm_output = n.output[0]
-- 
GitLab