Train multilingual pipeline with LLaMA embeddings #12790
-
Hello ! |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 3 replies
-
That's a good question! Large language models like LLaMA and GPT-NeoX are generally used as generative models, i.e., models that accept a prompt as input and generate a completion for it. But architecturally, they are similar to other Transformer models such as BERT and can theoretically be used to produce dense representations/embeddings for downstream tasks such as tagging, parsing, entity recognition, etc. Currently, we do not support their direct usage in spaCy pipelines outside of |
Beta Was this translation helpful? Give feedback.
That's a good question! Large language models like LLaMA and GPT-NeoX are generally used as generative models, i.e., models that accept a prompt as input and generate a completion for it. But architecturally, they are similar to other Transformer models such as BERT and can theoretically be used to produce dense representations/embeddings for downstream tasks such as tagging, parsing, entity recognition, etc.
Currently, we do not support their direct usage in spaCy pipelines outside of
spacy-llm
, which - as you correctly concluded - is a prompting component. However, we do have a couple of new libraries in development that we hope to release in the near future. These will serve as a good …