-
Notifications
You must be signed in to change notification settings - Fork 1.4k
Issues: mlc-ai/mlc-llm
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[Bug] iOS | mlc_llm package not working
bug
Confirmed bugs
#2477
opened May 30, 2024 by
iOSDevCodiste
[Doc] benchmark on different hardware
documentation
Improvements or additions to documentation
#2475
opened May 30, 2024 by
louis030195
[Doc] Request for suggested build-from-source options + explanation of added functionality
documentation
Improvements or additions to documentation
#2473
opened May 30, 2024 by
BuildBackBuehler
Compiling WebAssembly library with debug symbols/source map to aid in debugging
question
Question about the usage
#2472
opened May 30, 2024 by
slash-under
mlc_llm serve fails on concurrent users - Llama3 70B parameter hosting
bug
Confirmed bugs
#2462
opened May 29, 2024 by
swamysrivathsan
[Question] Does OpenCL on Adreno GPU support OpenCL ML SDK
question
Question about the usage
#2461
opened May 29, 2024 by
xiaoxiaoyuwen
'ChatGLMTokenizer' object has no attribute 'backend_tokenizer'
bug
Confirmed bugs
#2460
opened May 29, 2024 by
lihaofd
qwen1.5-0.5B-chat : lm_head.weight
question
Question about the usage
#2458
opened May 29, 2024 by
viaowp
Unable to serve Mistral-7B-Instruct-v0.3
bug
Confirmed bugs
#2447
opened May 28, 2024 by
swamysrivathsan
[Doc] Python API KV/memory reset details absent
documentation
Improvements or additions to documentation
#2426
opened May 26, 2024 by
federicoparra
[Feature Request] phi-3 small realeased -> performs two times ebtter then Phi-3 mini
feature request
New feature or request
#2420
opened May 26, 2024 by
sebastienbo
Phi-2 q4f16_1 runs faster when compiled without Confirmed bugs
tvm.relax.transform.FuseOps()
and tvm.relax.transform.FuseTIR()
transformations
bug
#2405
opened May 24, 2024 by
MMuzzammil1
Fail to build tvm-unity from source on orin[Bug]
bug
Confirmed bugs
#2389
opened May 23, 2024 by
Louym
[Bug] java.lang.NullPointerException: Attempt to invoke virtual method 'org.apache.tvm.TVMValue org.apache.tvm.Function.invoke()' on a null object reference
bug
Confirmed bugs
#2366
opened May 21, 2024 by
View999888
[Question] Single forward pass through ChatModule
question
Question about the usage
#2354
opened May 17, 2024 by
caenopy
[Feature Request] Implement AttentionStore
feature request
New feature or request
#2353
opened May 16, 2024 by
kripper
[Question] mlc_llm serve fails with --speculative-mode, does it require certain hardware?
question
Question about the usage
#2350
opened May 16, 2024 by
0xDEADFED5
[Question] Can MLC quantize multimodal models?
question
Question about the usage
#2349
opened May 16, 2024 by
LJ-Hao
[Question] Deployment of Pruned Models
question
Question about the usage
#2338
opened May 14, 2024 by
qianjyM
Previous Next
ProTip!
What’s not been updated in a month: updated:<2024-04-30.