-
### Your current environment
```text
The output of `python collect_env.py`
Collecting environment information...
WARNING 08-05 17:30:24 _custom_ops.py:15] Failed to import from vllm._C with Module…
-
I am getting segmentation fault when trying to compile model jinaai/jina-embeddings-v2-base-en with neuronx.
Example script:
```python
from transformers import AutoTokenizer, AutoModel
import …
-
-
We are trying to run the [e5 model](https://huggingface.co/intfloat/e5-large-v2) on an inf2 instance. The model compiles fine and analyze reports no unsupported operators but when trying it out on an …
aabbi updated
1 month ago
-
Hi, I find that neuron compiler will crash when compiling my customized model running on a Neuron device. Could you help check this issue? **Thanks so much!**
#### Below is my model sample and the…
-
### Feature request
Llama 3.1 is out and should be compatible with Neuron, however, it requires `transformers==4.43.1`, and `optimum-neuron` has pinned `transformers` to `4.41.1`.
Notes that sin…
-
-
I encountered an issue when trying to export the facebook/m2m100_418M model using the optimum-cli tool. The error message indicates that the m2m-100-encoder is not supported, despite m2m-100 being lis…
-
When running the notebook for inference using [Llama3](https://github.com/aws-neuron/aws-neuron-samples/blob/master/torch-neuronx/transformers-neuronx/inference/meta-llama-2-13b-sampling.ipynb)
```…
-