Multi-modal projections for Llama-Adapter #2420
efraimdahl
started this conversation in
Ideas
Replies: 1 comment
-
AFAICT, this is not possible. @yeoedward added the method to PEFT, maybe they can clarify. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Just want to make sure I'm not missing anything. Currently the implementation of Llama-Adapter/AdaptionPrompt does not include any way we can add additional input that is added onto the prefix? In the Llama-Adapters paper (implementation here) they include a section on unlocking multimodal reasoning by projecting CLIP embeddings onto the prefix. Are there other PEFT-methods or a configuration of AdaptionPrompt that already implement this functionality?
Beta Was this translation helpful? Give feedback.
All reactions