-
1、按要求安装了vs(安装了C++)、cuda
2、torch 2.0 cuda(True),可用
3、模型rwkv-7b 10v
4、显卡3090
5、pip 安装了 rwkv ninja
6、参数没有改动
serving on 0.0.0.0:17860 view at http://127.0.0.1:17860
No sentence-transformers mode…
-
Hi,
I'm using agx xavier, L4T: 35.4.1, JP:5.1.2, started: riva_quickstart_arm64_v2.12.0. (IP: 192.168.0.40)
and I have one orin agx, JP 6. running docker: dustynv/local_llm:r36.2.0 (IP: 192.168.…
-
On machines with either 8x A100-80GB or 8x H100, I'm getting many tactic out of memory issues during the build.
The tactic says it requesting 530000 MB while the GPU has 80GB, yet I only observe ~1…
-
Is it possible to fine tune a model that I pull from ollama? What would be the general process for that?
-
Hey there,
thanks for langchain! It's super awesome! 👍
I am currently trying to write a simple REST API but i am getting somewhat random errors. Sometimes (about 1 in 15 runs) it's this:
```…
-
### Self Checks
- [X] This is only for bug report, if you would like to ask a question, please head to [Discussions](https://github.com/langgenius/dify/discussions/categories/general).
- [X] I have s…
-
I have a 7900XT and would definitely love to have ROCm support. It seems like it might be coming with https://github.com/jmorganca/ollama/pull/667?
I couldn't find a dedicated issue for this so I'm…
-
Hi!
I'm using Ollama on a local server RTX A6000 ADA running Mixtral 8x7B.
I run Ollama locally and expose an API endpoint for multiple user to connect and use the LLM in a chat powered by Chainli…
-
error whlie attempting to bind on address ('127.0.0.1',20001):address already in use
how to deal with this problem?
![image](https://github.com/chatchat-space/Langchain-Chatchat/assets/85067013/b6e…
-
As we are transitioning to the SLM flow, we plan to phase out the old compilation flow in [mlc_llm](https://github.com/mlc-ai/mlc-llm/tree/main/mlc_llm) folder. The new flow would enable a single unif…