Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Provide sensible defaults for Model settings #40

Open
fbricon opened this issue Sep 23, 2024 · 1 comment
Open

Provide sensible defaults for Model settings #40

fbricon opened this issue Sep 23, 2024 · 1 comment
Assignees

Comments

@fbricon
Copy link
Collaborator

fbricon commented Sep 23, 2024

Continue supports fine-tuned model configuration.
Screenshot 2024-09-20 at 19 46 53

Screenshot 2024-09-20 at 19 54 34 Screenshot 2024-09-23 at 10 59 14

We should be able to provide proper defaults for each model size. @jamescho72 can you help here?

@jamescho72
Copy link

Soon as we build the performance test against and get baselines we will fine-tune with these configurations and the settings that the granite team recommended.
for code tasks,
completionOptions": {
"temperature": 0.2 or 0.3 (for higher precision, more deterministic)
"topP": 0.9 or 1
"topK": 40
"presencePenalty": 0.0
"frequencyPenalty": 0.1
"stop": null,
"maxTokens": (start small, test and expand)
}
e.g. start maxTokens at 2K or 3K , i.e. the maximum output length, that leaves plenty of room for inputs (over 120K+) but to minimize hallucination, we need to regulate both input and output size , and work to find a balance ... it all depends on the capability of the model

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants