Distilled model for high-throughput Danish NLP model #200
KennethEnevoldsen
started this conversation in
Missing pieces for Danish NLP
Replies: 1 comment 1 reply
-
|
What about quantization? Is that an undesirable approach for optimization
of processing speed in your view?
…On Fri, 8 Dec 2023 at 16.52, Kenneth Enevoldsen ***@***.***> wrote:
Statement of need
It is important to have highly performant models for Danish NLP as these
are useful for processing large amount of text on limited compute budgets.
Current status
Not in development
Approach
- My guess is that distilling a larger model such as the
dfm-encoder-large-v1 would be the best option, though it might be a better
approach to simply train a small model from scratch.
- We have not looked into what the best distillation approach is.
If you wish to take on the project feel free to start a discussion here.
—
Reply to this email directly, view it on GitHub
<#200>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AJAFWNLRBSZVGBVGTAQDU4LYIMZUXAVCNFSM6AAAAABAM3DBRGVHI2DSMVQWIX3LMV43ERDJONRXK43TNFXW4OZVHEZTSMJQG4>
.
You are receiving this because you are subscribed to this thread.Message
ID:
<centre-for-humanities-computing/danish-foundation-models/repo-discussions/200
@github.com>
|
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Statement of need
It is important to have highly performant models for Danish NLP as these are useful for processing large amount of text on limited compute budgets.
Current status
Not in development
Approach
If you wish to take on the project feel free to start a discussion here.
Beta Was this translation helpful? Give feedback.
All reactions