-
Notifications
You must be signed in to change notification settings - Fork 17.5k
Support for Rotary Embeddings for Llama #1885
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
I got the recently merged embeddings pr of llama working. Outputting a 5121 element long vector. What would be the best way to go about plugging this output to langchain? Take one of the existing embedding classes and base it of that or am I missing something? I mean this works but obviously it's terrible
|
How to convert Llama structure into Faster transformer sturcture ? |
Hi, @bradleat! I'm Dosu, and I'm here to help the LangChain team manage their backlog. I wanted to let you know that we are marking this issue as stale. From what I understand, the issue is about adding support for Rotary Embeddings in Llama. You mentioned that you wanted to discuss how to incorporate these embeddings into langchain and that you are willing to assist, although your knowledge is limited. In the comments, user "pachacamac" mentioned that they got the recently merged embeddings working and asked for guidance on how to plug the output into langchain. User "A-ML-ER" also asked about converting Llama structure into Faster transformer structure. Based on the comments, it seems that progress has been made towards resolving the issue. User "pachacamac" mentioned that they got the recently merged embeddings working. Before we close this issue, we wanted to check with you if it is still relevant to the latest version of the LangChain repository. If it is, please let us know by commenting on the issue. Otherwise, feel free to close the issue yourself, or the issue will be automatically closed in 7 days. Thank you for your understanding and contribution to the LangChain project! |
It looks like Llama uses an unsupported embedding scheme:
https://nn.labml.ai/transformers/rope/index.html
I'm opening this thread so we can have a conversation about how to support these embeddings within langchain. I'm happy to help, but my knowledge is limited.
The text was updated successfully, but these errors were encountered: