-
**Describe the bug**
I am getting a function error when testing the connection to Ollama. It seems I get a error staying. To me it seems the because of the model and Ollama results it's not working w…
-
I'd like to know if the code in this repository is complete. Has anyone tried pre-training this model from scratch?
-
Adding some support for 8bit quantization would be a good idea because it can fill the gap for people with less GPU VRAM to work with, So I think it would be a good idea if possible.
thank you.
-
As described, trying to spin up a `mistralai/Mistral-7B-v0.1` using the examples in the README. This is on an EC2 `g5.xlarge`.
```
import mii
client = mii.serve("mistralai/Mistral-7B-v0.1")
resp…
-
```python
from transformers import AutoTokenizer, TextStreamer
from intel_extension_for_transformers.transformers import AutoModelForCausalLM, WeightOnlyQuantConfig
model_name = "Intel/neural-chat-…
-
The `h2ogpt` linux installation method as [given here](https://github.com/h2oai/h2ogpt?tab=readme-ov-file#get-started) is as follows:
### A. Variable export instructions:
`export PIP_EXTRA_INDEX_…
-
### Python -VV
```shell
Python 3.10.12 (main, Jul 29 2024, 16:56:48) [GCC 11.4.0]
```
### Pip Freeze
```shell
mistral_common==1.3.3
```
### Reproduction Steps
Example code:
```
from mistral_c…
-
I don't understand how to use it, now I'll try to describe what exactly doesn't work.
I downloaded the ollama model (tried several different ones, including llama3), installed everything, everythin…
-
Why Phi 3.5 mini 3.8b require mor vram then Mistral 7b v0.3 for android?
-
# Prerequisites
Version 0.2.84 or 0.2.85 and using create_chat_completion method.
Tried different GGUF models.
Please answer the following questions for yourself before submitting an issue.
…