Skip to content

[FEAT] Integrate LoRETTA: Low-Rank Economic Tensor-Train Adaptation for Ultra-Low-Parameter Fine-Tuning of Large Language Models #2813

@mbaddar1

Description

@mbaddar1

Feature request

Paper link : https://aclanthology.org/2024.naacl-long.174.pdf
Code link : https://github.com/yifanycc/loretta

Motivation : Tensor-Trains are one of the promising low-rank structures heavily applied for parameter-efficient machine learning, esp. LLMs

LoRETTA takes LoRa to the next level by re-parameterizing the updates for pretrained weights to set of two tensor-trains

Image Image

Performance : The paper claims that loretta leads to significant parameters reduction wrt LoRa

Image

Also the paper claims LoRETTA achieves comparable results to other BERT family models

Image

Note
I have contacted authors about this FR
yifanycc/loretta#7

Your contribution

If it is OK, I would love to start a PR on that

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions