-
Hello,
I'm trying to add some more augmentations not yet available in DALI using DLTensorPythonFunction.
I made some tests with tfa.image.mean_filter2d on cpu and the speed is more than 10 lower t…
-
## 🐛 Bug
crash log
```
libc++abi: terminating due to uncaught exception of type tvm::runtime::InternalError: [13:57:25] /Users/xxx/project/mlc-llm/3rdparty/tvm/src/runtime/relax_vm/lm_support.cc:…
-
I deploy and execute locally under Windows/ scripts/prep_ Deps.sh, error message displayed as follows:
$ ./scripts/prep_deps.sh
+ emcc --version
emcc (Emscripten gcc/clang-like replacement + link…
-
# Failing Tests
> Please see the failing tests divided into sections below. Click on each section to expand. Feel free to get assigned to an issue by following the instructions [here](https://unify.ai…
-
I am running python backend on CPU and call to another GPU model, how to effective convert output to CPU without import `torch` GPU:
```
infer_response = inference_request.exec()
if infer_resp…
-
### Describe the issue
/workspace/dev/openlibs/onnxruntime/build/Linux/Release/tensorboard/compat/proto/config.pb.cc:702:6: error: ‘::descriptor_table_tensorboard_2fcompat_2fproto_2frewriter_5fconfig…
-
## 🐛 Bug
I have dual rtx 3090. Compiled model with command:
mlc_llm compile Llama-3-70B-Instruct-q4f16_1-MLC/mlc-chat-config.json --device cuda --overrides "tensor_parallel_shards=2" -o Llama-3-70…
pw-k updated
5 months ago
-
## 🐛 Bug
hello!when I running mlc_chat compile ./dist/Llama-2-7b-chat-hf-q4f16_1-MLC/mlc-chat-config.json --device vulkan -o dist/libs/Llama-2-7b-chat-hf-q4f16_1-vulkan.dll,I get the following erro…
-
## 🐛 Bug
When I compile models according to this link https://llm.mlc.ai/docs/compilation/compile_models.html, I encounter the following error when running the compiled model and library:
```
T…
-
Hi all,
I am using the latest version of the rapids.ai docker. 21.06 and in Juptyer notebook this code works with no issue:
```
from cuml.common.sparsefuncs import csr_row_normalize_l2
def…