Open Suiji12 opened 3 months ago
I met a error:C:\Users\jiaojiaxing.conda\envs\localgpt\python.exe E:\jiaojiaxing\GenerativeAIExamples\RetrievalAugmentedGeneration\examples\developer_rag\chains.py C:\Users\jiaojiaxing.conda\envs\localgpt\lib\site-packages\langchain_nvidia_ai_endpoints_common.py:172: UserWarning: An API key is required for the hosted NIM. This will become an error in the future. warnings.warn( C:\Users\jiaojiaxing.conda\envs\localgpt\lib\site-packages\langchain_nvidia_ai_endpoints_common.py:172: UserWarning: An API key is required for the hosted NIM. This will become an error in the future. warnings.warn( Traceback (most recent call last): File "E:\jiaojiaxing\GenerativeAIExamples\RetrievalAugmentedGeneration\examples\developer_rag\chains.py", line 40, in available_models
Process finished with exit code 1
My setting about rag-app-text-chatbot.yaml is services: jupyter-server: container_name: notebook-server image: notebook-server:${TAG:-latest} build: context: ../../ dockerfile: ./notebooks/Dockerfile.notebooks # replace GPU enabled Dockerfile ./notebooks/Dockerfile.gpu_notebook ports:
"8888" deploy: resources: reservations: devices:
chain-server: container_name: chain-server image: chain-server:${TAG:-latest} build: context: ../../ dockerfile: ./RetrievalAugmentedGeneration/Dockerfile args: EXAMPLE_NAME: developer_rag command: --port 8081 --host 0.0.0.0 environment: APP_VECTORSTORE_URL: "http://milvus:19530" APP_VECTORSTORE_NAME: "milvus" APP_EMBEDDINGS_MODELNAME: ${APP_EMBEDDINGS_MODELNAME:-G:/jjx/moxing/snowflake-arctic-embed-l} APP_EMBEDDINGS_MODELENGINE: ${APP_EMBEDDINGS_MODELENGINE:-local} APP_EMBEDDINGS_SERVERURL: ${APP_EMBEDDINGS_SERVERURL:-""} APP_LLM_SERVERURL: ${APP_LLM_SERVERURL:-""} APP_LLM_MODELNAME: ${APP_LLM_MODELNAME:-"G:/jjx/moxing/llama-2-13b-chat-hf"} APP_LLM_MODELENGINE: ${APP_LLM_MODELENGINE:-local} NVIDIA_API_KEY: ${NVIDIA_API_KEY} POSTGRES_PASSWORD: ${POSTGRES_PASSWORD:-password} POSTGRES_USER: ${POSTGRES_USER:-postgres} POSTGRES_DB: ${POSTGRES_DB:-api} COLLECTION_NAME: ${COLLECTION_NAME:-developer_rag} APP_RETRIEVER_TOPK: 4 APP_RETRIEVER_SCORETHRESHOLD: 0.25 OTEL_EXPORTER_OTLP_ENDPOINT: http://otel-collector:4317 OTEL_EXPORTER_OTLP_PROTOCOL: grpc ENABLE_TRACING: false APP_TEXTSPLITTER_MODELNAME: Snowflake/snowflake-arctic-embed-l APP_TEXTSPLITTER_CHUNKSIZE: 506 APP_TEXTSPLITTER_CHUNKOVERLAP: 200 LOGLEVEL: ${LOGLEVEL:-INFO} ports:
"8081" shm_size: 5gb deploy: resources: reservations: devices:
rag-playground: container_name: rag-playground image: rag-playground:${TAG:-latest} build: context: ../.././RetrievalAugmentedGeneration/frontend/ dockerfile: Dockerfile command: --port 8090 environment: APP_SERVERURL: http://chain-server APP_SERVERPORT: 8081 APP_MODELNAME: ${APP_LLM_MODELNAME:-"meta/llama3-8b-instruct"} OTEL_EXPORTER_OTLP_ENDPOINT: http://otel-collector:4317 OTEL_EXPORTER_OTLP_PROTOCOL: grpc ENABLE_TRACING: false RIVA_API_URI: ${RIVA_API_URI:-} RIVA_API_KEY: ${RIVA_API_KEY:-} RIVA_FUNCTION_ID: ${RIVA_FUNCTION_ID:-} TTS_SAMPLE_RATE: ${TTS_SAMPLE_RATE:-48000} ports:
networks: default: name: nvidia-rag What should I do?