-
Congrats on Flash Attention in the latest version, or to be precise, in having your storage limit increased on Pypi.org so you could upload the release that was weeks ago. Here are some benchmarks fo…
-
# Post Updates and Resources:
## Official Website for KDD 2024: [KDD'24](https://www.biendata.xyz/kdd2024/) and [Guidelines-ACA](https://www.biendata.xyz/competition/aqa_kdd_2024/)
## LLM RoadMa…
-
Is there a specific version of openai that is aligned with the OpenAI interfaces offered by neuralchat? I am currently testing using the current **1.12.0** but encountering a **422 Unprocessable Entit…
-
### Checklist
- [X] I am sure the error is coming from Pyrogram's code and not elsewhere
- [X] I have searched in the issue tracker for similar bug reports, including closed ones
- [X] I ran `pip3 in…
-
```python
from transformers import AutoTokenizer, TextStreamer
from intel_extension_for_transformers.transformers import AutoModelForCausalLM, WeightOnlyQuantConfig
model_name = "Intel/neural-chat-…
-
# 1. Ollama
## 1. use Ollama CLI:
```
ollama serve
ollama run llama2:7b, llama3, llama3:70b, mistral, dophin-phi, phi, neural-chat, codellama, llama2:13b, llama2:70b
ollama list
ollama show
…
-
### System Info
gpu:
```nvidia-smi
Mon Apr 22 17:00:40 2024
+---------------------------------------------------------------------------------------+
| NVIDIA-SMI 535.161.08 …
-
We have seen a significant difference in performance drop with the env created with the latest repo for vllm serving for the neural-chat model as compared to the old env built with the old repo. With …
-
Why this project has been deprecated? The raise of recent advancements of browser WebGPU and neural network Chat LLM models, could be a motivation to resurrect this amazing project!
Also mentioned in …
-
### Describe the bug
When attempting to run "interpreter --local" and choosing jan.ai as the llm provider, the model choice function crashes interpreter.
LM_Studio runs as expected. (I'm assumi…