-
### Is your feature request related to a problem? Please describe.
Re: [Conversation between Eric and I](https://github.com/microsoft/autogen/pull/2375#issuecomment-2091056580) We should have the abi…
-
### Feature Description
Is it possible for `node-llama-cpp` to support React Native projects?
I know that this project depends on many node standard libraries, but is it possible to make changes a…
-
Dear Clement,
I am recently trying to run your experiment. First of all, thanks for your great work! I really appreciate it. Yet I have two small questions:
1. I saw there are several Slurm scri…
-
### Summary
Although it is possible to run some LLM tasks with the RK3588 NPU, the toolchain released by Rockchip is currently closed-source, and its license is incompatible with our project. Here ar…
-
### Your current environment
```text
Collecting environment information...
PyTorch version: N/A
Is debug build: N/A
CUDA used to build PyTorch: N/A
ROCM used to build PyTorch: N/A
OS: Ubuntu …
-
By leveraging technologies like GPUs, Heroku can provide customers with faster and more efficient AI computing capabilities all with the Heroku DX to provide the “Heroku magic” customer experience whi…
-
We are building a voice-interactive chatbot that leverages cutting-edge technologies such as Speech-to-Text (STT), Text-to-Speech (TTS), and local Large Language Models (LLMs), with a focus on Ollama'…
-
### Describe the bug
[HuggingFaceEndpoint](libs/partners/huggingface/langchain_huggingface/llms/huggingface_endpoint.py) in the partners libs **HuggingFace** in the **Langchain librairy** cannot be u…
-
## ⚙️ Request New Models
- Link to an existing implementation (e.g. Hugging Face/Github): https://huggingface.co/apple/OpenELM-3B-Instruct
- Is this model architecture supported by MLC-LLM? (the …
-
## 🚀 Feature
I know there is OpenCL backend on Android platform.
But for many Android Devices, the GPU has been used by other subsystem like display.
So we need to use CPU to run LLM.