-
### Describe the bug
Trying to serve a basic LLaMa 13B model, I encounter an ImportError
### To reproduce
Two ways:
1. `pip` install
pip install "openllm[llama, vllm, fine-tune]==0.2.27"
…
-
### Self Checks
- [X] I have searched for existing issues [search for existing issues](https://github.com/langgenius/dify/issues), including closed ones.
- [X] I confirm that I am using English to…
-
### Describe the bug
Starting from a clean setup (Python 3.10), trying to start a LLaMa 13B results in a `ModuleNotFoundError` which, when corrected (by installing SciPy), results in nothing much h…
-
### Description of the new feature / enhancement
本地部署版的通义千问需要能被配置在模型注册中
### Scenario when this would be used?
这将会提高用户本地资源利用率
### Supporting information
_No response_
-
### Describe the bug
When I ping the endpoint started for an int8-quantized Llama-2, I get the following error:
```
File "/opt/conda/lib/python3.9/site-packages/openllm/models/llama/modeling_l…
-
I have been using Zep for a while and I have found it to be fantastic.
Setup:
Conversational Local LLM through Langchain
Local Embedding on Zep
OpenAI gpt 3.5 for Zep's LLM.
However, I have no…
-
### Issue with current documentation:
For example, in scenarios where chain_type=refine, LLM calls are made multiple times.
How do you control the intervals between LLM calls?
### Idea or request …
-
I've been exploring libraries that provide wrappers for LLM (large language models) and came across both `gpt4all` and `OpenLLM`. On a first glance, both seem to offer wrappers for a wide array of LLM…
-
Leveraging open source AI rather than propriety AI is quite important but for the sake of this feature request lets just say more operator choice.
Does Windmill AI use fine training or do prom…
-
### Feature request
Create openai.Model.list() method to fetch the running model from the endpoint.
### Motivation
Requested by Chaoyu
### Other
_No response_