From 14fd167fe4b5554b0d8eab09d48b53915e570769 Mon Sep 17 00:00:00 2001 From: auphelia <jakobapk@web.de> Date: Wed, 6 May 2020 16:30:41 +0100 Subject: [PATCH] [Transformation] Change line to write packed shape for driver.py template --- src/finn/transformation/fpgadataflow/make_pynq_driver.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/src/finn/transformation/fpgadataflow/make_pynq_driver.py b/src/finn/transformation/fpgadataflow/make_pynq_driver.py index c5b8d35db..049ede506 100644 --- a/src/finn/transformation/fpgadataflow/make_pynq_driver.py +++ b/src/finn/transformation/fpgadataflow/make_pynq_driver.py @@ -70,8 +70,8 @@ class MakePYNQDriver(Transformation): # extract HLSCustomOp instances to get folded i/o shapes first_node = getCustomOp(model.find_consumer(i_tensor_name)) last_node = getCustomOp(model.find_producer(o_tensor_name)) - i_tensor_shape_folded = first_node.get_folded_input_shape() - o_tensor_shape_folded = last_node.get_folded_output_shape() + i_tensor_shape_folded = tuple(first_node.get_folded_input_shape()) + o_tensor_shape_folded = tuple(last_node.get_folded_output_shape()) # generate dummy folded i/o tensors and their packed versions i_tensor_dummy_folded = gen_finn_dt_tensor(i_tensor_dt, i_tensor_shape_folded) o_tensor_dummy_folded = gen_finn_dt_tensor(o_tensor_dt, o_tensor_shape_folded) -- GitLab