From 2a4977a102775883ef4ea6f91a7b0ef403beee80 Mon Sep 17 00:00:00 2001 From: auphelia <jakobapk@web.de> Date: Thu, 28 May 2020 11:20:54 +0100 Subject: [PATCH] [Transformation] Ignore "dw-layers" during hls conversion --- src/finn/transformation/fpgadataflow/convert_to_hls_layers.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/src/finn/transformation/fpgadataflow/convert_to_hls_layers.py b/src/finn/transformation/fpgadataflow/convert_to_hls_layers.py index 3ff86cab4..f986cce1f 100644 --- a/src/finn/transformation/fpgadataflow/convert_to_hls_layers.py +++ b/src/finn/transformation/fpgadataflow/convert_to_hls_layers.py @@ -45,7 +45,7 @@ class InferConvInpGen(Transformation): graph_modified = False for n in graph.node: node_ind += 1 - if n.op_type == "Im2Col": + if n.op_type == "Im2Col" and not getCustomOp(n).get_nodeattr("dw"): i2c_input = n.input[0] i2c_output = n.output[0] i2c_in_shape = model.get_tensor_shape(i2c_input) @@ -282,7 +282,7 @@ class InferQuantizedStreamingFCLayer(Transformation): graph_modified = False for n in graph.node: node_ind += 1 - if n.op_type == "MatMul": + if n.op_type == "MatMul" and model.get_tensor_sparsity(n.input[1]) is None: mm_input = n.input[0] mm_weight = n.input[1] mm_output = n.output[0] -- GitLab