-
### Problem Description
runTrace.sh the vLLM benchmark failed
### Operating System
Ubuntu22.04 in the docker image rocm/vllm-dev:20241025-tuned
### CPU
AMD EPYC 9654 96-Core Processor
### GPU
A…
-
### Version
3.2.33
### Reproduction link
[sfc.vuejs.org/](https://sfc.vuejs.org/#eyJBcHAudnVlIjoiPHNjcmlwdCBzZXR1cD5cbmltcG9ydCB7IHJlZiB9IGZyb20gJ3Z1ZSdcblxuY29uc3QgZmlyc3RPcHRpb24gPSByZWYodHJ1ZS…
-
#### Summary
Hi again! I encountered a bug while playing with attention and sharding in JAX. The issue occurs with specific sharding setups and fails under certain core configurations.
#### Steps …
-
I am runing the kpm pipeline on my deeplabcut analyzed data with 10 individuals' position. When I run the generate grid movie function I encounter such error:
```
---------------------------------…
-
## Bug Description
I'm trying to serve torch-tensorrt optimized model to Nvidia Triton server based on the provided tutorial
https://pytorch.org/TensorRT/tutorials/serving_torch_tensorrt_with_t…
-
## Describe the bug
### My environment
Windows 11 Pro, Docker Desktop, WSL2 Ubuntu Engine, latest nvidia driver
### Cuda test
I made sure the Docker WSL2 Cuda implementation works correctly by…
-
[Bayesian optimisation of functions on graphs](https://proceedings.neurips.cc/paper_files/paper/2023/hash/86419aba4e5eafd2b1009a2e3c540bb0-Abstract-Conference.html)
```bib
@article{wan2023bayesian,
…
-
Here is a trace from my Intel Arc A770 via Docker:
```
$ ollama run deepseek-coder-v2
>>> write fizzbuzz
"""""""""""""""""""""""""""""""
```
And here is an trace from Arch linux running on …
-
**Is your feature request related to a problem? Please describe.**
The current deployed version of instructlab requires llama_cpp version 0.2.79, and I want to run the new IBM Granite architecture, w…
-
[Blog](https://ai.meta.com/blog/llama-3-2-connect-2024-vision-edge-mobile-devices/)