-
-
Notifications
You must be signed in to change notification settings - Fork 3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG]: "QNN Engine is offline." when using a Snapdragon X Plus laptop #2962
Comments
Same issue on snapdragon x elite for me |
Is this after downloading a model? Also have you tried a reboot post-download of the model? Second, @barealek - I just got confirmation that we can run Elite compiled models on Plus chipsets, so we will patch that and re-release 1.7.2 |
I downloaded and tried to run a model, choosing the qualcomm LLM provider and NPU embedder, but it came up with the error. It failed to work after fully rebooting the app and restarting my computer, then tried all of the same things after uninstalling and reinstalling the app, which still didn't work. I haven't done any additional setup of the NPU or anything outside of AnythingLLM, so I'm wondering if there is some driver(s) I'm missing? I'll let the experts figure it out. |
@lachlanharrisdev - we just pushed a new build for arm64 Also what device + chipset are you on? Plus, Elite, etc |
@timothycarambat I've just installed the new build and it's still failing but it's behaving differently. After I upgraded to the new version and sent a chat, it came up with the error What I noticed is that when booting up AnythingLLM, right before the loading screen switches to the home UI, I can see a task pop up for a split second in task manager called "AnythingLLMQnnEngine", but it seems to end itself very quickly. Same task also pops up after I send a chat, after the QNN Engine "boots", but then again it quickly closes itself. I'm currently on a Surface Laptop 7 15", running the X elite X1E-80-100. |
@lachlanharrisdev I wrote this up to debug the engine directly (app should be closed) I have the same chipset on a Dell Latitude, |
@timothycarambat yep, that found the issue
If it's relevant, this was using llama 3.1 8b, not 3.2 3b. |
@lachlanharrisdev Now this is a very different issue from other then. If you run the command as administrator does it still fail to initialize? I am wondering how/why you would require admin to execute the LLM engine, but someone else had success with that and I have to determine why that would ever be the case for anyone since that should not be required to start the QNN LLM API. |
From the recent patch that seemed to solve most issues people had (most Plus support was not enabled) but this is certainly something different |
@timothycarambat nope, running it as admin now works and I do see QNN running on localhost.
I tried running AnythingLLM as administrator and, after QNN engine boots, I can successfully chat. This works for me, but I'm more than happy to keep testing things out for you, I'd love to contribute in any way I can. Should we create a new issue and continue there? |
Build 1.7.2-r2-arm64 seems to be working well while running in Administrator mode. Happy hunting, everyone! |
@timothycarambat I've just restarted my PC and now it seems to no longer work even with administrator mode... I'm guessing the same QNN Engine instance stayed online from the instructinos in the google doc, and AnythingLLM used that instance instead of booting another one (if that's even possible, I know barely anything about AI and Qualcomm). Hopefull that clears up any confusion. @AlphaEcho11 interesting, what device are you using? Just wondering if this is only a surface laptop thing |
@lachlanharrisdev - Surface Pro 11 here, on the X Elite. After several device reboots and AnythingLLM refreshes, it's been working without issue. |
I am still having issues, even when launching as an administrative account. It seems like it's starting up now, I get a message that roughly says "QNN is still booting, please wait", but then it just crashes and the QNN engine goes offline. Here's my logs: |
Thank you for the logs! Yes, seeing the QNN engine fail to get online here; going to check one more area and see if another variable is at play. |
Can you reattempt this with the 8B model as well? Following @timothycarambat 's previous recommendations and tweaking:
Let us know the results! |
How are you running AnythingLLM?
AnythingLLM desktop app
What happened?
When trying to inference on any QNN model on a Snapdragon X Plus laptop, the issue below occurs.
The logs specifies that the required CPU/NPU is not found:
Starting AnythingLLM and reproducing the error, the full log looks like this:
Are there known steps to reproduce?
No response
The text was updated successfully, but these errors were encountered: