You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The config of llama3.2 sets tie_word_embeddings to True, which means that LLama2.2 3B share the weight matrix of embeddings andlm_head. Please refer to discuss.huggingface.co and transformers for detail.
Hello, I find that theres no lm_head weight in model checkpoints(.safetensors).
How does model load weight for the Linear Layer of lm_head ?
The text was updated successfully, but these errors were encountered: