-
Notifications
You must be signed in to change notification settings - Fork 35
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
how to make it faster #6
Comments
I have a branch that moves more of the processing into native code, I believe it should bring a noticable performance improvement. You can also try 3B models with this version, which should also be much faster. Feel free to try. |
Hi dsd, it works with the apk you provided , but I failed to run it from your forked source. |
It's my first time developing Android apps but feel free to share info about the failure to run from source and I will let you know if I have any ideas. I did not do any work to retain Mac compatibility but I think this is what needs to be done: #12 (comment) |
Is this app using both CPU and GPU of smartphones? |
llama.cpp is used as the backend, so you would need to check if llama.cpp supports your GPU, and if it is usable on 4GB RAM with the model you are interested in. |
Does it support any mobile gpu like mali or adreno? |
i installed the latest version and its a cool app but it so slow I'm running the vicuna 7b is there a way to make faster i have an 8gb ram phone and what other models does support and please link me to them
The text was updated successfully, but these errors were encountered: