-
## 🐛 Bug
I use the jetson-containers of MLC and use Meta-Llama-3-8B-Instruct model . after I run ```
python3 -m mlc_llm.build \
--model Meta-Llama-3-8B-Instruct-hf \
--quantization q4f16_…
-
### Describe the issue
I am trying building TVM Execution Provider from this document:
https://onnxruntime.ai/docs/execution-providers/community-maintained/TVM-ExecutionProvider.html
however, a…
-
see [here](https://github.com/apache/tvm)
-
## 🐛 Bug
I am seeing `illegal instruction` error when running mlc-llm with H100 (Driver Version: 550.54.1 CUDA Version: 12.4). It works fine on another A100 machine (Driver Version: 535.161.0…
-
hi, this is great lib, i wonder how would you describe the difference or different goals between candle and [mlc-llm](https://github.com/mlc-ai/mlc-llm)/[tvm](https://github.com/apache/tvm/tree/main)?
-
### What Happened
When I attempt to build TVM-Unity I run into errors with the ACL Lib. I believe/hope it is because of this change not being reflected within the TVM ACL CMake Config.
The 2 er…
-
## 🐛 Bug
I'm serving Llama3 70B model in a g5.12xlarge aws EC2 instance. Below is the version of mlc_llm installed.
mlc-ai-nightly-cu122 0.15.dev389
mlc-llm-nightly-cu122 0.1.dev1320
…
-
TVM Upgrades [2023.07](https://docs.ton.org/learn/tvm-instructions/tvm-upgrade-2023-07) and [2024.04](https://docs.ton.org/learn/tvm-instructions/fee-calculation-instructions) introduced many useful T…
-
-
## 🐛 Bug
I found [this](https://huggingface.co/bayley/Midnight-Miqu-70B-v1.5-q4f16_1-MLC) repo on Huggingface, kindly publicly shared by @bayley who also provided the commands for serving, but upon…