-
Notifications
You must be signed in to change notification settings - Fork 1.6k
Issues: mlc-ai/mlc-llm
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[Question] How to export MLCChat .apk with weight bundled/included?
question
Question about the usage
#3013
opened Nov 6, 2024 by
lifelongeeek
[Bug] Speculative decoding doesn't work on Vulkan (AMD iGPU)
bug
Confirmed bugs
#3011
opened Nov 4, 2024 by
SkyHeroesS
[Question] Android app issue
question
Question about the usage
#3010
opened Nov 4, 2024 by
j0h0k0i0m
[Bug] Llama-3.1-70B-Instruct-q3f16_1-MLC model running across two GPUs with tensor_parallel_shards=2
bug
Confirmed bugs
#3004
opened Oct 31, 2024 by
shahizat
[Bug] Misalignment of Llama3.2 chat template
bug
Confirmed bugs
#3002
opened Oct 31, 2024 by
Hzfengsy
[Question] Error running prep_emcc_deps.sh - 'tvm/runtime/object.h' file not found
question
Question about the usage
#3001
opened Oct 30, 2024 by
Big-Boy-420
[Question] Which models do you recommend for compiling on Mac Intel chip, metal gpu?
question
Question about the usage
#2999
opened Oct 28, 2024 by
RINO-GAELICO
[Bug] Llama 3.2 3B and 1B on MLC are significantly slower than Llama 3.1 8B (L40s, fp16)
bug
Confirmed bugs
#2997
opened Oct 26, 2024 by
chrisreese-if
[Question] TVM error on Mac Intel chip, Metal accelerator
question
Question about the usage
#2995
opened Oct 24, 2024 by
RINO-GAELICO
[Bug] Concurrent requests are being run sequentially on AMD MI60 (gfx906)
bug
Confirmed bugs
#2992
opened Oct 22, 2024 by
Said-Akbar
Problems encountered while doing Android SDK
question
Question about the usage
#2986
opened Oct 18, 2024 by
whyshifht
[Bug] some questions about build Android app with mlc
bug
Confirmed bugs
#2985
opened Oct 17, 2024 by
ecccccsgo
[Bug] Running Quick Start Example in Windows gives Error: Confirmed bugs
'MLCEngine' object has no attribute '_ffi'
bug
#2979
opened Oct 15, 2024 by
jackuh105
[Bug] Llama example inference using Vulkan gives build error
bug
Confirmed bugs
#2977
opened Oct 14, 2024 by
asfarkTii
[Question] Fail to pip install mlc-llm
question
Question about the usage
#2974
opened Oct 13, 2024 by
xiboliyaxiangjiaojun
Error with LLVM Configuration on Windows for GPU Inference in mlc-llm
bug
Confirmed bugs
#2969
opened Oct 9, 2024 by
sreejith-ios
[Question] Why phi3.5v inference image is much longer than HF
question
Question about the usage
#2967
opened Oct 8, 2024 by
Liuuuu54
InternalError when running llava model
question
Question about the usage
#2966
opened Oct 7, 2024 by
plufz
[Bug] ValueError: Unknown model type: rwkv
bug
Confirmed bugs
#2963
opened Oct 7, 2024 by
haohenggang
[Question] How do you abort in the event of a memory pressure warning on iOS?
question
Question about the usage
#2930
opened Sep 22, 2024 by
bryan1anderson
Previous Next
ProTip!
Find all open issues with in progress development work with linked:pr.