Skip to content

Questions about layer input/output quantization in FINN #397

Answered by maltanar
hleblevec asked this question in Q&A
Discussion options

You must be logged in to vote

Hi Hugo,

Brevitas is intended to be a general-purpose research tool for quantization-aware training, and there are many networks/quantization configurations that can be represented in Brevitas which won't export to or otherwise go through the FINN compiler. We don't have full documentation on what is and isn't supported in the FINN compiler, see also related FAQ item here: https://finn.readthedocs.io/en/latest/faq.html#can-i-deploy-custom-nns-with-arbitrary-precisions-and-layers-using-finn

  • Regarding quantizing layer inputs/outputs, we typically do this with an explicit quantized activation layer (e.g. QuantIdentity, QuantReLU that comes before/after the layer itself.
  • Regarding choosing t…

Replies: 1 comment 1 reply

Comment options

You must be logged in to vote
1 reply
@hleblevec
Comment options

Answer selected by hleblevec
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants