Open
Description
Great job with this repo. I was able to finetune Llama-2 and it certainly seems to have an effect.
Unfortunately the finetune silently accepts all inputs and the documentation states that you simply truncate inputs to max length. But it's not specified anywhere what's LLama-2's max length. Originally Meta released it with a bug that caused max length to be 2048 while the native max length seems to be 4096. So which is it?
Also, I tested my finetune model's completion code with inputs as big as 12,000 tokens and it still makes a completion. So I assume you truncate there as well? Only taking the tail of the prompt, presumably?
tldr:
- What is llama-2's max token length?
- Is there anything we can do to effect this or get better visibility into how the input got tokenized, etc?
Metadata
Metadata
Assignees
Labels
No labels