Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
For S3 Model loading, we clone all files except weights files to a local, temporary, directory.
We do it using ignore pattern of
["*.pt", "*.safetensors", "*.bin"]
.This raise a problem, users may upload large files to the S3 (For example, like in this issue: run-ai/runai-model-streamer#25 (comment))
The case is very valid, downloading all model repo from HF and directly upload it to S3 (Including
.git
which is a large dir)Most of them are not needed.
One option is to still pull all files, but it can take time if the files are large (like the
.git
dir)Option 2 - pull only relevant files, tokenizer files, however there is no clear list of which file names are relevant and may change in the future.
Option 3 - Pull only json files, from the assumption that if the filenames will change in the future most likely all the tokenizer metadata file will still be jsons.
Option 4 - ignore
.git
dir only