Can finn accelerate 4w4a or 8w8a models ? #578
Unanswered
YangNuoCheng
asked this question in
Q&A
Replies: 1 comment 1 reply
-
|
here are my exported onnx and notebook based on BNN-PYNQ |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Can finn accelerate 4w4a or 8w8a models ?
I noticed there no
CNV_4w4aandCNV_8w8adata in TABLEis the FINN didn't support 4bit accelerate?I got the accelerate idea from FINN-R.
I modified
BNN_PYNQinbrevitas_examplesto getCNV_2w2a、CNV_4w4aandCNV_8w8amodules and use following code for export.I'm tring to accelerate the quantized model on PYNQ-Z2 board.
When I using
cnv=get_test_model_trained("CNV",2,2))or my export CNV_2w2a.onnx there have same bugs as '#437' (InferQuantizedStreamingFCLayer: Signed output requres actval < 0) inLowering and Streamliningpart.But I didn't modified the construction of CNV.
How can I get the same CNV_2w2a result from TABLE?
Beta Was this translation helpful? Give feedback.
All reactions