Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Moved export_meta_llama_bin.py to new export.py #329

Merged
merged 3 commits into from
Aug 21, 2023

Conversation

atamurad
Copy link
Contributor

Follow up to #326

Summary:

  • Tested updated export.py with Meta LLama2-7B model.
  • Minor note: moved float32 conversion in serialize_fp32 from numpy to torch, because numpy does not support BFloat16 dtype (Meta model uses it)
  • Let argparse handle one of input model arguments required logic & error message:
usage: export.py [-h] [--version VERSION] (--checkpoint CHECKPOINT | --meta-llama META_LLAMA | --hf HF) filepath
export.py: error: one of the arguments --checkpoint --meta-llama --hf is required

@karpathy
Copy link
Owner

yay!! 🎉 much nicer.

@karpathy karpathy merged commit 8a3ea7b into karpathy:master Aug 21, 2023
6 checks passed
vinhtran2611 pushed a commit to vinhtran2611/llama2.c that referenced this pull request Jan 20, 2024
Moved export_meta_llama_bin.py to new export.py
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants