NVIDIA / GenerativeAIExamples

Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.
Apache License 2.0
2.45k stars 527 forks source link

chain-server? docker desktop Mac #205

Open tvuongp opened 1 month ago

tvuongp commented 1 month ago

[+] Running 5/5 ✔ Container milvus-etcd Started 0.3s ✔ Container milvus-minio Started 0.3s ✔ Container chain-server Started 0.2s ✔ Container milvus-standalone Started 0.4s ✔ Container rag-playground Started 0.3s ❯ docker ps --format "table {{.ID}}\t{{.Names}}\t{{.Status}}" CONTAINER ID NAMES STATUS 3a447f023622 rag-playground Up 8 minutes 96a6762b0113 chain-server Up 8 minutes ff4ebdfa15b6 milvus-standalone Up 8 minutes 662cc72b09d8 milvus-minio Up 8 minutes (unhealthy) f3fbd91ef0e5 milvus-etcd Up 8 minutes (healthy) ed18e9aa80c6 buildx_buildkit_builder0 Up 24 hours

ERROR:: (in web interface chat) "Failed to get response from /generate endpoint of chain-server. Check if the fastapi server in chain-server is up. Refer to chain-server logs for details."

docker logs 96a6762b01134ab4d62fc07009f3d9e38cc797784d4fb354b974d466f673058c INFO: Started server process [1] INFO: Waiting for application startup. INFO:faiss.loader:Loading faiss. INFO:faiss.loader:Successfully loaded faiss. INFO:RAG.src.chain_server.utils:Using nvidia-ai-endpoints as model engine and nvidia/nv-embedqa-e5-v5 and model for embeddings INFO:RAG.src.chain_server.utils:Using embedding model nvidia/nv-embedqa-e5-v5 hosted at api catalog INFO:RAG.src.chain_server.utils:Using milvus collection: nvidia_api_catalog INFO:RAG.src.chain_server.utils:Vector store created and saved. INFO: Application startup complete. INFO: Uvicorn running on http://0.0.0.0:8081 (Press CTRL+C to quit) INFO:example:Using llm to generate response directly without knowledge base. INFO:RAG.src.chain_server.utils:Using nvidia-ai-endpoints as model engine for llm. Model name: meta/llama3-70b-instruct WARNING:RAG.src.chain_server.utils:The following parameters from kwargs are not supported: ['stop'] for nvidia-ai-endpoints INFO:RAG.src.chain_server.utils:Using llm model meta/llama3-70b-instruct from api catalog INFO:example:Prompt used for response generation: System: You are a helpful, respectful and honest assistant. Always answer as helpfully as possible, while being safe. Please ensure that your responses are positive in nature.

Human:

Question: what is rag

INFO: 172.19.0.5:59680 - "POST /generate HTTP/1.1" 200 OK ERROR: Exception in ASGI application Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/starlette/responses.py", line 264, in call await wrap(partial(self.listen_for_disconnect, receive)) File "/usr/local/lib/python3.10/dist-packages/starlette/responses.py", line 260, in wrap await func() File "/usr/local/lib/python3.10/dist-packages/starlette/responses.py", line 237, in listen_for_disconnect message = await receive() File "/usr/local/lib/python3.10/dist-packages/uvicorn/protocols/http/httptools_impl.py", line 580, in receive await self.message_event.wait() File "/usr/lib/python3.10/asyncio/locks.py", line 214, in wait await fut asyncio.exceptions.CancelledError: Cancelled by cancel scope fffee886fb80

During handling of the above exception, another exception occurred:

Human:

Question: rag

INFO: 172.19.0.5:53786 - "POST /generate HTTP/1.1" 200 OK ERROR: Exception in ASGI application Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/starlette/responses.py", line 264, in call await wrap(partial(self.listen_for_disconnect, receive)) File "/usr/local/lib/python3.10/dist-packages/starlette/responses.py", line 260, in wrap await func() File "/usr/local/lib/python3.10/dist-packages/starlette/responses.py", line 237, in listen_for_disconnect message = await receive() File "/usr/local/lib/python3.10/dist-packages/uvicorn/protocols/http/httptools_impl.py", line 580, in receive await self.message_event.wait() File "/usr/lib/python3.10/asyncio/locks.py", line 214, in wait await fut asyncio.exceptions.CancelledError: Cancelled by cancel scope fffee7355bd0

During handling of the above exception, another exception occurred:

Human:

Question: rag

INFO: 172.19.0.5:49858 - "POST /generate HTTP/1.1" 200 OK ERROR: Exception in ASGI application Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/starlette/responses.py", line 264, in call await wrap(partial(self.listen_for_disconnect, receive)) File "/usr/local/lib/python3.10/dist-packages/starlette/responses.py", line 260, in wrap await func() File "/usr/local/lib/python3.10/dist-packages/starlette/responses.py", line 237, in listen_for_disconnect message = await receive() File "/usr/local/lib/python3.10/dist-packages/uvicorn/protocols/http/httptools_impl.py", line 580, in receive await self.message_event.wait() File "/usr/lib/python3.10/asyncio/locks.py", line 214, in wait await fut asyncio.exceptions.CancelledError: Cancelled by cancel scope fffe8716d810

During handling of the above exception, another exception occurred: