Skip to content

Transformer-based models #57

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
ChloeCarbonniere opened this issue Apr 14, 2025 · 3 comments
Open

Transformer-based models #57

ChloeCarbonniere opened this issue Apr 14, 2025 · 3 comments

Comments

@ChloeCarbonniere
Copy link

Hello,

Is it possible to implement transforme-based models on the STM32N6 ?

@RSERSTM
Copy link

RSERSTM commented Apr 14, 2025

Hello ChloeCarbonniere,
What do you mean by transforme-based models ? Transformer-based models ?
Attention layers ? Encoder-only computer vision models ? Generative models ?
Thank you,

@ChloeCarbonniere
Copy link
Author

Mi final goal is to implement a DETR on the STM32N6.

@RSERSTM
Copy link

RSERSTM commented Apr 14, 2025

Ok I understand.
For info STM32N6-DK has 128MB of Flash to store weights and RAM (internal 4.2MB + external 32MB) to store activations(the data) https://www.st.com/en/evaluation-tools/stm32n6570-dk.html.
In other words regarding these kind of Transformers (DETR usually takes GBytes of space) for the moment we don't have these in the model zoo because they are often to big for STM32N6.
But it should be possible to implement very small Transformers with quantized (8bit) Attention layers if the weights + activations fit in memory.
Regards,

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants