Skip to content

Commit 34c5dee

Browse files
author
Danqing Wang (MPK)
committed
address comment
1 parent 8419a74 commit 34c5dee

File tree

2 files changed

+2
-7
lines changed

2 files changed

+2
-7
lines changed

examples/models/llama/export_llama_lib.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -94,7 +94,7 @@
9494
"static_llama",
9595
"qwen2_5",
9696
"phi-4-mini",
97-
"smollm",
97+
"smolllm2",
9898
]
9999
TORCHTUNE_DEFINED_MODELS = ["llama3_2_vision"]
100100

examples/models/smollm/convert_weights.py

+1-6
Original file line numberDiff line numberDiff line change
@@ -42,11 +42,6 @@ def smollm_tune_to_meta(state_dict: Dict[str, torch.Tensor]) -> Dict[str, torch.
4242
new_key = get_mapped_key(key, inverted_mapping_dict)
4343
converted_state_dict[new_key] = value
4444

45-
# Input and output embeddings are tied.
46-
converted_state_dict["output.weight"] = converted_state_dict[
47-
"tok_embeddings.weight"
48-
]
49-
5045
return converted_state_dict
5146

5247

@@ -68,7 +63,7 @@ def main():
6863
checkpoint_dir=args.input_dir,
6964
checkpoint_files=["model.safetensors"],
7065
output_dir=".",
71-
model_type="MISTRAL",
66+
model_type="LLAMA",
7267
)
7368

7469
print("Loading checkpoint...")

0 commit comments

Comments
 (0)