-
Currently, [Aspire supports connecting to OpenAI / Azure OpenAI services](https://github.com/dotnet/aspire/tree/main/src/Components/Aspire.Azure.AI.OpenAI).
Add component to support different AI m…
-
This issue is to explain how to host locally the LLM model.
For all the solutions listed below, `ngrok.com` (or any similar tool) can be used to share the local AI server to other people.
We ha…
-
The new version is much better, I mean it started working easily and with Vulcan offloading.
The local mode returns a warning:
`[node-llama-cpp] Using this model ("~/.humanifyjs/models/Phi-3.1-min…
-
With limited memory on most of phones, there's community requests on supporting a model with a smaller size like Phi-3 mini. It may be supported out of box, but need to verification, evaluation and pr…
-
### What is the issue?
I setted up a modelFile that loads PHI-3-mini-instruct, and whatever input I give it it starts to generate infinite response related to coding as shown in the screenshot.
Here…
-
### What happened?
While flash attention works well for my python ui (https://github.com/curvedinf/dir-assistant/) on an nvidia system, it produces bad results on my AMD system. My AMD system has a…
-
### **Background:**
TT-Buda, developed by Tenstorrent, is a growing collection of model demos showcasing the capabilities of AI models running on Tenstorrent hardware. These demonstrations cover a wi…
-
**Describe the bug**
The tutorial on the front README of this repo says to run the following command:
```
huggingface-cli download microsoft/Phi-3-mini-4k-instruct-onnx --include cpu_and_mobile/c…
-
### Model description
Can you please add support for the Phi-2 and the Ohi-3 mini models.
The Phi-3 supports onnx, so I'm hoping it should work:
https://huggingface.co/microsoft/Phi-3-mini-128k-in…
-
### Checked other resources
- [X] I added a very descriptive title to this issue.
- [X] I searched the LangChain documentation with the integrated search.
- [X] I used the GitHub search to find a…