-
Hi Turbo,
I have a fresh pull of the dev branch, and I deleted the ~/.cache/torch_extensions folder.
I am trying to convert Phi-3-small using this command:
```
python convert.py \
-i /med…
-
### Checked other resources
- [X] I added a very descriptive title to this issue.
- [X] I searched the LangChain documentation with the integrated search.
- [X] I used the GitHub search to find a…
-
### feature
Could you please support Llama3 in Llava ?
-
Hi all, this issue will track the feature requests you've made to TensorRT-LLM & provide a place to see what TRT-LLM is currently working on.
Last update: `Jan 14th, 2024`
🚀 = in development
#…
-
### Describe the bug
`onnxruntime` using CUDA is packaged with both `onnxruntime_USE_CUDA` and `onnxruntime_DISABLE_CONTRIB_OPS` which effectively disables CUDA and leads to following error on runtim…
anpin updated
4 months ago
-
# URL
- https://arxiv.org/abs/2404.14219
# Affiliations
- Marah Abdin, N/A
- Sam Ade Jacobs, N/A
- Ammar Ahmad Awan, N/A
- Jyoti Aneja, N/A
- Ahmed Awadallah, N/A
- Hany Awadalla, N/A
- …
-
Hi,
How can I create a tokenizer based on [microsoft/Phi-3.5-mini-instruct](https://huggingface.co/microsoft/Phi-3.5-mini-instruct) model?
Should I load [tokenzier.model](https://huggingface.co/micr…
-
### What is the issue?
When I start the Conversation in German Language, Phi-3 Mini and Medium working fine. But after some Conversations, the Models starting producing slowly Gibberish and Nonsens a…
-
### Software
Desktop Application
### Operating System / Platform
macOS
### Your Pieces OS Version
10.0.0
### Early Access Program
- [ ] Yes, this is related to an Early Access Program feature.
…
-
The GPT4All program crashes every time I attempt to load a model. My laptop should have the necessary specs to handle the models, so I believe there might be a bug or compatibility issue.
Steps to …