Skip to content

Which evo2 model to select for finetuning #1315

@MidAtBest

Description

@MidAtBest

Description

We are currently finetuning evo2 on token level classification. Thanks to your help here #1170 we've managed to move along and finetune the evo2/1b-8k:1.0 model.

However, we noticed there are two checkpoints available for the Evo2 base model — evo2/1b-8k:1.0 and evo2/1b-8k-bf16:1.0. From the descriptions, it seems that the first is the original NeMo2-format conversion of arcinstitute/savanna_evo2_1b_base, while the second is a fine-tuned variant optimized for BF16 precision. For the fairest comparison to Evo2, could you advise which checkpoint and training precision you recommend using?

SHA

495ba03

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions