From 12ea89772c039adb19ab27d4c084da557bbe0f5b Mon Sep 17 00:00:00 2001 From: Hendrik Borras <hendrikborras@web.de> Date: Fri, 1 Oct 2021 10:18:15 +0100 Subject: [PATCH] Resolved UserWarnings after QONNX to FINN conversion. --- src/finn/transformation/convert_qonnx_to_finn.py | 10 ++++++++-- 1 file changed, 8 insertions(+), 2 deletions(-) diff --git a/src/finn/transformation/convert_qonnx_to_finn.py b/src/finn/transformation/convert_qonnx_to_finn.py index eea00b7fa..814c7f059 100644 --- a/src/finn/transformation/convert_qonnx_to_finn.py +++ b/src/finn/transformation/convert_qonnx_to_finn.py @@ -37,6 +37,7 @@ from finn.custom_op.registry import getCustomOp from finn.transformation.base import Transformation from finn.transformation.infer_datatypes import InferDataTypes from finn.transformation.infer_shapes import InferShapes +from finn.util.basic import get_by_name class ConvertQONNXtoFINN(Transformation): @@ -54,8 +55,6 @@ class ConvertQONNXtoFINN(Transformation): model = model.transform(FoldQuantWeights()) # Convert activations model = model.transform(ConvertQuantActToMultiThreshold()) - # Some datatypes have changed - model = model.transform(InferDataTypes()) return (model, False) @@ -344,6 +343,13 @@ class QuantActBaseHandler(ABC): graph.node.insert(running_node_index, outp_trans_node) running_node_index += 1 + # Unset the FINN datatype + qnt_annotations = model._model_proto.graph.quantization_annotation + ret = get_by_name(qnt_annotations, n.output[0], "tensor_name") + ret.Clear() + # ToDo: This should be supported by finn-base, by calling the following: + # model.set_tensor_datatype(n.output[0], None) + # Insert Add node if adder_bias.shape == (1,): adder_bias = adder_bias[0] -- GitLab