From 14fd167fe4b5554b0d8eab09d48b53915e570769 Mon Sep 17 00:00:00 2001
From: auphelia <jakobapk@web.de>
Date: Wed, 6 May 2020 16:30:41 +0100
Subject: [PATCH] [Transformation] Change line to write packed shape for
 driver.py template

---
 src/finn/transformation/fpgadataflow/make_pynq_driver.py | 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)

diff --git a/src/finn/transformation/fpgadataflow/make_pynq_driver.py b/src/finn/transformation/fpgadataflow/make_pynq_driver.py
index c5b8d35db..049ede506 100644
--- a/src/finn/transformation/fpgadataflow/make_pynq_driver.py
+++ b/src/finn/transformation/fpgadataflow/make_pynq_driver.py
@@ -70,8 +70,8 @@ class MakePYNQDriver(Transformation):
         # extract HLSCustomOp instances to get folded i/o shapes
         first_node = getCustomOp(model.find_consumer(i_tensor_name))
         last_node = getCustomOp(model.find_producer(o_tensor_name))
-        i_tensor_shape_folded = first_node.get_folded_input_shape()
-        o_tensor_shape_folded = last_node.get_folded_output_shape()
+        i_tensor_shape_folded = tuple(first_node.get_folded_input_shape())
+        o_tensor_shape_folded = tuple(last_node.get_folded_output_shape())
         # generate dummy folded i/o tensors and their packed versions
         i_tensor_dummy_folded = gen_finn_dt_tensor(i_tensor_dt, i_tensor_shape_folded)
         o_tensor_dummy_folded = gen_finn_dt_tensor(o_tensor_dt, o_tensor_shape_folded)
-- 
GitLab