microsoft / LLaVA-Med

Large Language-and-Vision Assistant for Biomedicine, built towards multimodal GPT-4 level capabilities.
Other
1.29k stars 148 forks source link

ConnectionRefusedError: [Errno 111] Connection refused #21

Open atultiwari opened 7 months ago

atultiwari commented 7 months ago

Hi, When I try to the Serving commands either launching a model worker, sending test message or launch gradio webserver, I always get ConnectionRefused Error 111. I have tried both on a rented A100 GPU via cloud service (jarvislabs.ai) and Google Colab Pro+. I have tried following commands in both -

!python -m llava.serve.gradio_web_server --controller http://localhost:10000
!python -m llava.serve.test_message --model-name /content/drive/MyDrive/nlp/llava_med_model --controller http://localhost:10000
!python -m llava.serve.model_worker --host 0.0.0.0 --controller http://localhost:10000 --port 40000 --worker http://localhost:40000 --model-path /content/drive/MyDrive/nlp/llava_med_model --multi-modal

Following is the log from launching model worker -

2023-11-17 12:57:04.254739: E tensorflow/compiler/xla/stream_executor/cuda/cuda_dnn.cc:9342] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered
2023-11-17 12:57:04.254799: E tensorflow/compiler/xla/stream_executor/cuda/cuda_fft.cc:609] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered
2023-11-17 12:57:04.254877: E tensorflow/compiler/xla/stream_executor/cuda/cuda_blas.cc:1518] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered
2023-11-17 12:57:05.515822: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-11-17 12:57:07 | INFO | model_worker | args: Namespace(host='0.0.0.0', port=40000, worker_address='http://localhost:40000/', controller_address='http://localhost:10000/', model_path='/content/drive/MyDrive/nlp/llava_med_model', model_name=None, multi_modal=True, keep_aspect_ratio=False, num_gpus=1, limit_model_concurrency=5, stream_interval=2, no_register=False)
2023-11-17 12:57:07 | WARNING | model_worker | Multimodal mode is automatically detected with model name, please make sure `llava` is included in the model path.
2023-11-17 12:57:07 | INFO | model_worker | Loading the model llava_med_model on worker 2b442e ...
(…)t-large-patch14/resolve/main/config.json:   0% 0.00/4.52k [00:00<?, ?B/s]
(…)t-large-patch14/resolve/main/config.json: 100% 4.52k/4.52k [00:00<00:00, 16.7MB/s]
2023-11-17 12:57:14 | ERROR | stderr | 
model.safetensors:   0% 0.00/1.71G [00:00<?, ?B/s]
model.safetensors:   1% 21.0M/1.71G [00:00<00:09, 184MB/s]
model.safetensors:   3% 52.4M/1.71G [00:00<00:06, 254MB/s]
model.safetensors:   5% 83.9M/1.71G [00:00<00:06, 265MB/s]
model.safetensors:   7% 115M/1.71G [00:00<00:06, 229MB/s]
model.safetensors:   9% 147M/1.71G [00:00<00:07, 221MB/s]
model.safetensors:  10% 178M/1.71G [00:00<00:07, 216MB/s]
model.safetensors:  12% 210M/1.71G [00:00<00:06, 231MB/s]
model.safetensors:  14% 241M/1.71G [00:01<00:06, 238MB/s]
model.safetensors:  16% 273M/1.71G [00:01<00:05, 256MB/s]
model.safetensors:  18% 304M/1.71G [00:01<00:05, 258MB/s]
model.safetensors:  20% 346M/1.71G [00:01<00:04, 278MB/s]
model.safetensors:  22% 377M/1.71G [00:01<00:05, 226MB/s]
model.safetensors:  24% 409M/1.71G [00:01<00:05, 239MB/s]
model.safetensors:  26% 440M/1.71G [00:01<00:04, 255MB/s]
model.safetensors:  28% 472M/1.71G [00:01<00:05, 248MB/s]
model.safetensors:  29% 503M/1.71G [00:02<00:05, 225MB/s]
model.safetensors:  31% 535M/1.71G [00:02<00:05, 218MB/s]
model.safetensors:  34% 577M/1.71G [00:02<00:04, 245MB/s]
model.safetensors:  36% 619M/1.71G [00:02<00:04, 260MB/s]
model.safetensors:  39% 661M/1.71G [00:02<00:03, 275MB/s]
model.safetensors:  41% 703M/1.71G [00:02<00:03, 289MB/s]
model.safetensors:  43% 734M/1.71G [00:02<00:03, 288MB/s]
model.safetensors:  45% 776M/1.71G [00:03<00:03, 298MB/s]
model.safetensors:  47% 807M/1.71G [00:03<00:03, 263MB/s]
model.safetensors:  49% 839M/1.71G [00:03<00:03, 262MB/s]
model.safetensors:  51% 870M/1.71G [00:03<00:03, 259MB/s]
model.safetensors:  53% 902M/1.71G [00:03<00:03, 244MB/s]
model.safetensors:  55% 933M/1.71G [00:03<00:03, 258MB/s]
model.safetensors:  56% 965M/1.71G [00:03<00:02, 261MB/s]
model.safetensors:  58% 996M/1.71G [00:03<00:02, 264MB/s]
model.safetensors:  61% 1.04G/1.71G [00:04<00:02, 290MB/s]
model.safetensors:  63% 1.07G/1.71G [00:04<00:02, 284MB/s]
model.safetensors:  64% 1.10G/1.71G [00:04<00:02, 290MB/s]
model.safetensors:  66% 1.13G/1.71G [00:04<00:02, 235MB/s]
model.safetensors:  69% 1.17G/1.71G [00:04<00:02, 263MB/s]
model.safetensors:  71% 1.22G/1.71G [00:04<00:01, 258MB/s]
model.safetensors:  73% 1.25G/1.71G [00:04<00:01, 248MB/s]
model.safetensors:  75% 1.29G/1.71G [00:05<00:01, 272MB/s]
model.safetensors:  77% 1.32G/1.71G [00:05<00:01, 270MB/s]
model.safetensors:  80% 1.36G/1.71G [00:05<00:01, 286MB/s]
model.safetensors:  82% 1.39G/1.71G [00:05<00:01, 286MB/s]
model.safetensors:  83% 1.43G/1.71G [00:05<00:01, 282MB/s]
model.safetensors:  85% 1.46G/1.71G [00:05<00:00, 272MB/s]
model.safetensors:  87% 1.49G/1.71G [00:05<00:00, 258MB/s]
model.safetensors:  89% 1.52G/1.71G [00:05<00:00, 252MB/s]
model.safetensors:  91% 1.56G/1.71G [00:06<00:00, 273MB/s]
model.safetensors:  93% 1.59G/1.71G [00:06<00:00, 250MB/s]
model.safetensors:  95% 1.63G/1.71G [00:06<00:00, 244MB/s]
model.safetensors:  97% 1.67G/1.71G [00:06<00:00, 259MB/s]
model.safetensors:  99% 1.70G/1.71G [00:06<00:00, 272MB/s]
model.safetensors: 100% 1.71G/1.71G [00:06<00:00, 258MB/s]
2023-11-17 12:57:21 | ERROR | stderr | 
Some weights of the model checkpoint at openai/clip-vit-large-patch14 were not used when initializing CLIPVisionModel: ['text_model.encoder.layers.2.mlp.fc1.weight', 'text_model.encoder.layers.3.self_attn.v_proj.weight', 'text_model.encoder.layers.2.self_attn.out_proj.bias', 'text_model.encoder.layers.4.self_attn.out_proj.bias', 'text_model.final_layer_norm.bias', 'text_model.encoder.layers.10.mlp.fc2.bias', 'text_model.encoder.layers.5.layer_norm1.bias', 'text_model.encoder.layers.2.layer_norm1.bias', 'text_model.encoder.layers.1.self_attn.q_proj.weight', 'text_model.encoder.layers.7.self_attn.k_proj.bias', 'text_model.encoder.layers.3.self_attn.q_proj.bias', 'visual_projection.weight', 'text_model.encoder.layers.3.mlp.fc2.bias', 'text_model.encoder.layers.4.mlp.fc1.weight', 'text_model.encoder.layers.3.layer_norm1.bias', 'text_model.encoder.layers.8.layer_norm2.weight', 'text_model.encoder.layers.3.layer_norm1.weight', 'text_model.encoder.layers.11.self_attn.out_proj.weight', 'text_model.encoder.layers.4.mlp.fc1.bias', 'text_model.encoder.layers.8.self_attn.v_proj.weight', 'text_model.encoder.layers.11.mlp.fc1.weight', 'text_model.encoder.layers.7.mlp.fc2.weight', 'text_model.encoder.layers.3.mlp.fc1.bias', 'text_model.encoder.layers.4.self_attn.k_proj.bias', 'text_model.encoder.layers.4.self_attn.q_proj.bias', 'text_model.encoder.layers.0.self_attn.v_proj.weight', 'text_model.encoder.layers.7.self_attn.q_proj.bias', 'text_model.encoder.layers.0.layer_norm2.weight', 'text_model.encoder.layers.8.self_attn.q_proj.weight', 'text_model.encoder.layers.8.self_attn.out_proj.bias', 'text_model.encoder.layers.6.layer_norm1.weight', 'text_model.encoder.layers.3.self_attn.q_proj.weight', 'text_model.embeddings.position_ids', 'text_model.encoder.layers.11.layer_norm1.bias', 'text_model.encoder.layers.5.self_attn.v_proj.weight', 'text_model.encoder.layers.6.self_attn.q_proj.bias', 'text_model.encoder.layers.6.self_attn.out_proj.weight', 'text_model.encoder.layers.2.self_attn.q_proj.bias', 'text_model.encoder.layers.0.layer_norm2.bias', 'text_model.encoder.layers.4.self_attn.v_proj.weight', 'text_model.embeddings.token_embedding.weight', 'text_model.encoder.layers.9.self_attn.q_proj.bias', 'text_model.encoder.layers.10.layer_norm2.bias', 'text_model.encoder.layers.0.self_attn.k_proj.weight', 'text_model.encoder.layers.2.layer_norm2.weight', 'text_model.encoder.layers.6.mlp.fc2.weight', 'text_model.encoder.layers.2.self_attn.v_proj.bias', 'text_model.encoder.layers.4.mlp.fc2.weight', 'text_model.encoder.layers.9.self_attn.out_proj.bias', 'text_model.encoder.layers.7.mlp.fc1.weight', 'text_model.encoder.layers.0.mlp.fc1.bias', 'text_model.encoder.layers.1.mlp.fc1.weight', 'text_model.encoder.layers.0.mlp.fc1.weight', 'text_model.encoder.layers.11.mlp.fc1.bias', 'text_model.encoder.layers.6.self_attn.k_proj.weight', 'text_model.encoder.layers.1.mlp.fc2.bias', 'text_model.encoder.layers.0.self_attn.q_proj.weight', 'text_model.encoder.layers.10.self_attn.out_proj.weight', 'text_model.encoder.layers.1.self_attn.q_proj.bias', 'text_model.encoder.layers.6.self_attn.v_proj.bias', 'text_model.encoder.layers.8.layer_norm2.bias', 'text_projection.weight', 'text_model.encoder.layers.2.mlp.fc2.bias', 'text_model.encoder.layers.5.mlp.fc2.weight', 'text_model.encoder.layers.6.self_attn.q_proj.weight', 'text_model.encoder.layers.10.layer_norm1.weight', 'text_model.encoder.layers.0.self_attn.out_proj.bias', 'text_model.encoder.layers.6.mlp.fc2.bias', 'text_model.encoder.layers.5.self_attn.k_proj.weight', 'text_model.encoder.layers.0.layer_norm1.bias', 'text_model.encoder.layers.7.mlp.fc2.bias', 'text_model.encoder.layers.10.self_attn.v_proj.weight', 'text_model.encoder.layers.7.self_attn.out_proj.bias', 'text_model.encoder.layers.0.self_attn.q_proj.bias', 'text_model.encoder.layers.8.mlp.fc1.bias', 'text_model.encoder.layers.6.layer_norm2.bias', 'text_model.encoder.layers.5.mlp.fc1.bias', 'text_model.encoder.layers.2.layer_norm1.weight', 'text_model.encoder.layers.3.mlp.fc1.weight', 'text_model.encoder.layers.10.self_attn.q_proj.bias', 'text_model.encoder.layers.3.self_attn.v_proj.bias', 'text_model.encoder.layers.10.mlp.fc1.weight', 'text_model.encoder.layers.7.self_attn.q_proj.weight', 'text_model.encoder.layers.9.mlp.fc2.bias', 'text_model.encoder.layers.4.layer_norm1.weight', 'text_model.encoder.layers.7.self_attn.k_proj.weight', 'text_model.encoder.layers.4.layer_norm2.weight', 'text_model.encoder.layers.8.layer_norm1.bias', 'text_model.encoder.layers.1.self_attn.out_proj.bias', 'text_model.encoder.layers.7.layer_norm1.bias', 'text_model.encoder.layers.8.self_attn.k_proj.weight', 'text_model.encoder.layers.11.layer_norm2.bias', 'text_model.encoder.layers.5.self_attn.k_proj.bias', 'text_model.encoder.layers.0.self_attn.out_proj.weight', 'text_model.encoder.layers.1.layer_norm2.bias', 'text_model.encoder.layers.6.self_attn.k_proj.bias', 'text_model.encoder.layers.5.mlp.fc1.weight', 'text_model.encoder.layers.11.layer_norm2.weight', 'text_model.encoder.layers.2.mlp.fc2.weight', 'text_model.encoder.layers.9.self_attn.v_proj.weight', 'text_model.encoder.layers.8.self_attn.v_proj.bias', 'text_model.encoder.layers.3.layer_norm2.bias', 'text_model.encoder.layers.9.layer_norm1.bias', 'text_model.encoder.layers.3.layer_norm2.weight', 'text_model.encoder.layers.9.self_attn.k_proj.bias', 'text_model.encoder.layers.11.self_attn.k_proj.bias', 'text_model.encoder.layers.5.layer_norm2.bias', 'text_model.encoder.layers.2.mlp.fc1.bias', 'text_model.encoder.layers.5.mlp.fc2.bias', 'text_model.encoder.layers.10.self_attn.k_proj.bias', 'text_model.encoder.layers.6.layer_norm2.weight', 'text_model.encoder.layers.10.self_attn.v_proj.bias', 'text_model.encoder.layers.11.self_attn.q_proj.bias', 'text_model.encoder.layers.5.self_attn.out_proj.bias', 'text_model.encoder.layers.8.self_attn.out_proj.weight', 'text_model.encoder.layers.6.self_attn.v_proj.weight', 'text_model.encoder.layers.3.self_attn.k_proj.weight', 'logit_scale', 'text_model.encoder.layers.0.self_attn.v_proj.bias', 'text_model.encoder.layers.1.layer_norm1.weight', 'text_model.encoder.layers.1.self_attn.k_proj.bias', 'text_model.encoder.layers.11.layer_norm1.weight', 'text_model.encoder.layers.11.mlp.fc2.bias', 'text_model.encoder.layers.9.mlp.fc2.weight', 'text_model.encoder.layers.7.mlp.fc1.bias', 'text_model.encoder.layers.10.self_attn.k_proj.weight', 'text_model.encoder.layers.11.self_attn.v_proj.weight', 'text_model.encoder.layers.2.self_attn.v_proj.weight', 'text_model.encoder.layers.1.self_attn.v_proj.bias', 'text_model.encoder.layers.5.layer_norm1.weight', 'text_model.encoder.layers.3.mlp.fc2.weight', 'text_model.encoder.layers.0.mlp.fc2.bias', 'text_model.encoder.layers.9.self_attn.q_proj.weight', 'text_model.encoder.layers.10.layer_norm2.weight', 'text_model.encoder.layers.9.self_attn.v_proj.bias', 'text_model.encoder.layers.7.self_attn.out_proj.weight', 'text_model.encoder.layers.11.self_attn.k_proj.weight', 'text_model.encoder.layers.8.mlp.fc1.weight', 'text_model.encoder.layers.1.self_attn.v_proj.weight', 'text_model.encoder.layers.1.layer_norm2.weight', 'text_model.encoder.layers.9.mlp.fc1.weight', 'text_model.encoder.layers.6.mlp.fc1.weight', 'text_model.encoder.layers.4.self_attn.q_proj.weight', 'text_model.encoder.layers.2.self_attn.out_proj.weight', 'text_model.final_layer_norm.weight', 'text_model.encoder.layers.10.layer_norm1.bias', 'text_model.encoder.layers.10.mlp.fc1.bias', 'text_model.encoder.layers.0.self_attn.k_proj.bias', 'text_model.encoder.layers.0.mlp.fc2.weight', 'text_model.encoder.layers.4.layer_norm1.bias', 'text_model.encoder.layers.4.layer_norm2.bias', 'text_model.encoder.layers.5.self_attn.v_proj.bias', 'text_model.encoder.layers.1.self_attn.out_proj.weight', 'text_model.encoder.layers.7.layer_norm2.bias', 'text_model.encoder.layers.2.self_attn.q_proj.weight', 'text_model.encoder.layers.3.self_attn.k_proj.bias', 'text_model.encoder.layers.7.self_attn.v_proj.bias', 'text_model.encoder.layers.5.self_attn.q_proj.weight', 'text_model.encoder.layers.1.mlp.fc2.weight', 'text_model.encoder.layers.9.self_attn.k_proj.weight', 'text_model.encoder.layers.8.self_attn.q_proj.bias', 'text_model.encoder.layers.11.mlp.fc2.weight', 'text_model.encoder.layers.8.layer_norm1.weight', 'text_model.encoder.layers.11.self_attn.q_proj.weight', 'text_model.encoder.layers.4.self_attn.k_proj.weight', 'text_model.encoder.layers.6.self_attn.out_proj.bias', 'text_model.encoder.layers.3.self_attn.out_proj.bias', 'text_model.encoder.layers.11.self_attn.v_proj.bias', 'text_model.encoder.layers.4.mlp.fc2.bias', 'text_model.encoder.layers.1.layer_norm1.bias', 'text_model.encoder.layers.1.self_attn.k_proj.weight', 'text_model.encoder.layers.5.layer_norm2.weight', 'text_model.encoder.layers.9.mlp.fc1.bias', 'text_model.encoder.layers.7.layer_norm1.weight', 'text_model.encoder.layers.2.self_attn.k_proj.weight', 'text_model.encoder.layers.2.layer_norm2.bias', 'text_model.encoder.layers.2.self_attn.k_proj.bias', 'text_model.encoder.layers.3.self_attn.out_proj.weight', 'text_model.encoder.layers.8.self_attn.k_proj.bias', 'text_model.encoder.layers.10.mlp.fc2.weight', 'text_model.encoder.layers.1.mlp.fc1.bias', 'text_model.encoder.layers.7.self_attn.v_proj.weight', 'text_model.encoder.layers.8.mlp.fc2.bias', 'text_model.encoder.layers.6.layer_norm1.bias', 'text_model.encoder.layers.11.self_attn.out_proj.bias', 'text_model.encoder.layers.9.layer_norm2.bias', 'text_model.embeddings.position_embedding.weight', 'text_model.encoder.layers.4.self_attn.out_proj.weight', 'text_model.encoder.layers.7.layer_norm2.weight', 'text_model.encoder.layers.6.mlp.fc1.bias', 'text_model.encoder.layers.10.self_attn.q_proj.weight', 'text_model.encoder.layers.4.self_attn.v_proj.bias', 'text_model.encoder.layers.0.layer_norm1.weight', 'text_model.encoder.layers.10.self_attn.out_proj.bias', 'text_model.encoder.layers.5.self_attn.q_proj.bias', 'text_model.encoder.layers.9.self_attn.out_proj.weight', 'text_model.encoder.layers.8.mlp.fc2.weight', 'text_model.encoder.layers.9.layer_norm2.weight', 'text_model.encoder.layers.9.layer_norm1.weight', 'text_model.encoder.layers.5.self_attn.out_proj.weight']
- This IS expected if you are initializing CLIPVisionModel from the checkpoint of a model trained on another task or with another architecture (e.g. initializing a BertForSequenceClassification model from a BertForPreTraining model).
- This IS NOT expected if you are initializing CLIPVisionModel from the checkpoint of a model that you expect to be exactly identical (initializing a BertForSequenceClassification model from a BertForSequenceClassification model).
Loading checkpoint shards:   0% 0/2 [00:00<?, ?it/s]
Loading checkpoint shards:  50% 1/2 [02:48<02:48, 168.95s/it]
Loading checkpoint shards: 100% 2/2 [04:03<00:00, 113.41s/it]
Loading checkpoint shards: 100% 2/2 [04:03<00:00, 121.74s/it]
2023-11-17 13:01:25 | ERROR | stderr | 
(…)14/resolve/main/preprocessor_config.json:   0% 0.00/316 [00:00<?, ?B/s]
(…)14/resolve/main/preprocessor_config.json: 100% 316/316 [00:00<00:00, 1.66MB/s]
2023-11-17 13:01:27 | ERROR | stderr | 
Some weights of the model checkpoint at openai/clip-vit-large-patch14 were not used when initializing CLIPVisionModel: ['text_model.encoder.layers.2.mlp.fc1.weight', 'text_model.encoder.layers.3.self_attn.v_proj.weight', 'text_model.encoder.layers.2.self_attn.out_proj.bias', 'text_model.encoder.layers.4.self_attn.out_proj.bias', 'text_model.final_layer_norm.bias', 'text_model.encoder.layers.10.mlp.fc2.bias', 'text_model.encoder.layers.5.layer_norm1.bias', 'text_model.encoder.layers.2.layer_norm1.bias', 'text_model.encoder.layers.1.self_attn.q_proj.weight', 'text_model.encoder.layers.7.self_attn.k_proj.bias', 'text_model.encoder.layers.3.self_attn.q_proj.bias', 'visual_projection.weight', 'text_model.encoder.layers.3.mlp.fc2.bias', 'text_model.encoder.layers.4.mlp.fc1.weight', 'text_model.encoder.layers.3.layer_norm1.bias', 'text_model.encoder.layers.8.layer_norm2.weight', 'text_model.encoder.layers.3.layer_norm1.weight', 'text_model.encoder.layers.11.self_attn.out_proj.weight', 'text_model.encoder.layers.4.mlp.fc1.bias', 'text_model.encoder.layers.8.self_attn.v_proj.weight', 'text_model.encoder.layers.11.mlp.fc1.weight', 'text_model.encoder.layers.7.mlp.fc2.weight', 'text_model.encoder.layers.3.mlp.fc1.bias', 'text_model.encoder.layers.4.self_attn.k_proj.bias', 'text_model.encoder.layers.4.self_attn.q_proj.bias', 'text_model.encoder.layers.0.self_attn.v_proj.weight', 'text_model.encoder.layers.7.self_attn.q_proj.bias', 'text_model.encoder.layers.0.layer_norm2.weight', 'text_model.encoder.layers.8.self_attn.q_proj.weight', 'text_model.encoder.layers.8.self_attn.out_proj.bias', 'text_model.encoder.layers.6.layer_norm1.weight', 'text_model.encoder.layers.3.self_attn.q_proj.weight', 'text_model.embeddings.position_ids', 'text_model.encoder.layers.11.layer_norm1.bias', 'text_model.encoder.layers.5.self_attn.v_proj.weight', 'text_model.encoder.layers.6.self_attn.q_proj.bias', 'text_model.encoder.layers.6.self_attn.out_proj.weight', 'text_model.encoder.layers.2.self_attn.q_proj.bias', 'text_model.encoder.layers.0.layer_norm2.bias', 'text_model.encoder.layers.4.self_attn.v_proj.weight', 'text_model.embeddings.token_embedding.weight', 'text_model.encoder.layers.9.self_attn.q_proj.bias', 'text_model.encoder.layers.10.layer_norm2.bias', 'text_model.encoder.layers.0.self_attn.k_proj.weight', 'text_model.encoder.layers.2.layer_norm2.weight', 'text_model.encoder.layers.6.mlp.fc2.weight', 'text_model.encoder.layers.2.self_attn.v_proj.bias', 'text_model.encoder.layers.4.mlp.fc2.weight', 'text_model.encoder.layers.9.self_attn.out_proj.bias', 'text_model.encoder.layers.7.mlp.fc1.weight', 'text_model.encoder.layers.0.mlp.fc1.bias', 'text_model.encoder.layers.1.mlp.fc1.weight', 'text_model.encoder.layers.0.mlp.fc1.weight', 'text_model.encoder.layers.11.mlp.fc1.bias', 'text_model.encoder.layers.6.self_attn.k_proj.weight', 'text_model.encoder.layers.1.mlp.fc2.bias', 'text_model.encoder.layers.0.self_attn.q_proj.weight', 'text_model.encoder.layers.10.self_attn.out_proj.weight', 'text_model.encoder.layers.1.self_attn.q_proj.bias', 'text_model.encoder.layers.6.self_attn.v_proj.bias', 'text_model.encoder.layers.8.layer_norm2.bias', 'text_projection.weight', 'text_model.encoder.layers.2.mlp.fc2.bias', 'text_model.encoder.layers.5.mlp.fc2.weight', 'text_model.encoder.layers.6.self_attn.q_proj.weight', 'text_model.encoder.layers.10.layer_norm1.weight', 'text_model.encoder.layers.0.self_attn.out_proj.bias', 'text_model.encoder.layers.6.mlp.fc2.bias', 'text_model.encoder.layers.5.self_attn.k_proj.weight', 'text_model.encoder.layers.0.layer_norm1.bias', 'text_model.encoder.layers.7.mlp.fc2.bias', 'text_model.encoder.layers.10.self_attn.v_proj.weight', 'text_model.encoder.layers.7.self_attn.out_proj.bias', 'text_model.encoder.layers.0.self_attn.q_proj.bias', 'text_model.encoder.layers.8.mlp.fc1.bias', 'text_model.encoder.layers.6.layer_norm2.bias', 'text_model.encoder.layers.5.mlp.fc1.bias', 'text_model.encoder.layers.2.layer_norm1.weight', 'text_model.encoder.layers.3.mlp.fc1.weight', 'text_model.encoder.layers.10.self_attn.q_proj.bias', 'text_model.encoder.layers.3.self_attn.v_proj.bias', 'text_model.encoder.layers.10.mlp.fc1.weight', 'text_model.encoder.layers.7.self_attn.q_proj.weight', 'text_model.encoder.layers.9.mlp.fc2.bias', 'text_model.encoder.layers.4.layer_norm1.weight', 'text_model.encoder.layers.7.self_attn.k_proj.weight', 'text_model.encoder.layers.4.layer_norm2.weight', 'text_model.encoder.layers.8.layer_norm1.bias', 'text_model.encoder.layers.1.self_attn.out_proj.bias', 'text_model.encoder.layers.7.layer_norm1.bias', 'text_model.encoder.layers.8.self_attn.k_proj.weight', 'text_model.encoder.layers.11.layer_norm2.bias', 'text_model.encoder.layers.5.self_attn.k_proj.bias', 'text_model.encoder.layers.0.self_attn.out_proj.weight', 'text_model.encoder.layers.1.layer_norm2.bias', 'text_model.encoder.layers.6.self_attn.k_proj.bias', 'text_model.encoder.layers.5.mlp.fc1.weight', 'text_model.encoder.layers.11.layer_norm2.weight', 'text_model.encoder.layers.2.mlp.fc2.weight', 'text_model.encoder.layers.9.self_attn.v_proj.weight', 'text_model.encoder.layers.8.self_attn.v_proj.bias', 'text_model.encoder.layers.3.layer_norm2.bias', 'text_model.encoder.layers.9.layer_norm1.bias', 'text_model.encoder.layers.3.layer_norm2.weight', 'text_model.encoder.layers.9.self_attn.k_proj.bias', 'text_model.encoder.layers.11.self_attn.k_proj.bias', 'text_model.encoder.layers.5.layer_norm2.bias', 'text_model.encoder.layers.2.mlp.fc1.bias', 'text_model.encoder.layers.5.mlp.fc2.bias', 'text_model.encoder.layers.10.self_attn.k_proj.bias', 'text_model.encoder.layers.6.layer_norm2.weight', 'text_model.encoder.layers.10.self_attn.v_proj.bias', 'text_model.encoder.layers.11.self_attn.q_proj.bias', 'text_model.encoder.layers.5.self_attn.out_proj.bias', 'text_model.encoder.layers.8.self_attn.out_proj.weight', 'text_model.encoder.layers.6.self_attn.v_proj.weight', 'text_model.encoder.layers.3.self_attn.k_proj.weight', 'logit_scale', 'text_model.encoder.layers.0.self_attn.v_proj.bias', 'text_model.encoder.layers.1.layer_norm1.weight', 'text_model.encoder.layers.1.self_attn.k_proj.bias', 'text_model.encoder.layers.11.layer_norm1.weight', 'text_model.encoder.layers.11.mlp.fc2.bias', 'text_model.encoder.layers.9.mlp.fc2.weight', 'text_model.encoder.layers.7.mlp.fc1.bias', 'text_model.encoder.layers.10.self_attn.k_proj.weight', 'text_model.encoder.layers.11.self_attn.v_proj.weight', 'text_model.encoder.layers.2.self_attn.v_proj.weight', 'text_model.encoder.layers.1.self_attn.v_proj.bias', 'text_model.encoder.layers.5.layer_norm1.weight', 'text_model.encoder.layers.3.mlp.fc2.weight', 'text_model.encoder.layers.0.mlp.fc2.bias', 'text_model.encoder.layers.9.self_attn.q_proj.weight', 'text_model.encoder.layers.10.layer_norm2.weight', 'text_model.encoder.layers.9.self_attn.v_proj.bias', 'text_model.encoder.layers.7.self_attn.out_proj.weight', 'text_model.encoder.layers.11.self_attn.k_proj.weight', 'text_model.encoder.layers.8.mlp.fc1.weight', 'text_model.encoder.layers.1.self_attn.v_proj.weight', 'text_model.encoder.layers.1.layer_norm2.weight', 'text_model.encoder.layers.9.mlp.fc1.weight', 'text_model.encoder.layers.6.mlp.fc1.weight', 'text_model.encoder.layers.4.self_attn.q_proj.weight', 'text_model.encoder.layers.2.self_attn.out_proj.weight', 'text_model.final_layer_norm.weight', 'text_model.encoder.layers.10.layer_norm1.bias', 'text_model.encoder.layers.10.mlp.fc1.bias', 'text_model.encoder.layers.0.self_attn.k_proj.bias', 'text_model.encoder.layers.0.mlp.fc2.weight', 'text_model.encoder.layers.4.layer_norm1.bias', 'text_model.encoder.layers.4.layer_norm2.bias', 'text_model.encoder.layers.5.self_attn.v_proj.bias', 'text_model.encoder.layers.1.self_attn.out_proj.weight', 'text_model.encoder.layers.7.layer_norm2.bias', 'text_model.encoder.layers.2.self_attn.q_proj.weight', 'text_model.encoder.layers.3.self_attn.k_proj.bias', 'text_model.encoder.layers.7.self_attn.v_proj.bias', 'text_model.encoder.layers.5.self_attn.q_proj.weight', 'text_model.encoder.layers.1.mlp.fc2.weight', 'text_model.encoder.layers.9.self_attn.k_proj.weight', 'text_model.encoder.layers.8.self_attn.q_proj.bias', 'text_model.encoder.layers.11.mlp.fc2.weight', 'text_model.encoder.layers.8.layer_norm1.weight', 'text_model.encoder.layers.11.self_attn.q_proj.weight', 'text_model.encoder.layers.4.self_attn.k_proj.weight', 'text_model.encoder.layers.6.self_attn.out_proj.bias', 'text_model.encoder.layers.3.self_attn.out_proj.bias', 'text_model.encoder.layers.11.self_attn.v_proj.bias', 'text_model.encoder.layers.4.mlp.fc2.bias', 'text_model.encoder.layers.1.layer_norm1.bias', 'text_model.encoder.layers.1.self_attn.k_proj.weight', 'text_model.encoder.layers.5.layer_norm2.weight', 'text_model.encoder.layers.9.mlp.fc1.bias', 'text_model.encoder.layers.7.layer_norm1.weight', 'text_model.encoder.layers.2.self_attn.k_proj.weight', 'text_model.encoder.layers.2.layer_norm2.bias', 'text_model.encoder.layers.2.self_attn.k_proj.bias', 'text_model.encoder.layers.3.self_attn.out_proj.weight', 'text_model.encoder.layers.8.self_attn.k_proj.bias', 'text_model.encoder.layers.10.mlp.fc2.weight', 'text_model.encoder.layers.1.mlp.fc1.bias', 'text_model.encoder.layers.7.self_attn.v_proj.weight', 'text_model.encoder.layers.8.mlp.fc2.bias', 'text_model.encoder.layers.6.layer_norm1.bias', 'text_model.encoder.layers.11.self_attn.out_proj.bias', 'text_model.encoder.layers.9.layer_norm2.bias', 'text_model.embeddings.position_embedding.weight', 'text_model.encoder.layers.4.self_attn.out_proj.weight', 'text_model.encoder.layers.7.layer_norm2.weight', 'text_model.encoder.layers.6.mlp.fc1.bias', 'text_model.encoder.layers.10.self_attn.q_proj.weight', 'text_model.encoder.layers.4.self_attn.v_proj.bias', 'text_model.encoder.layers.0.layer_norm1.weight', 'text_model.encoder.layers.10.self_attn.out_proj.bias', 'text_model.encoder.layers.5.self_attn.q_proj.bias', 'text_model.encoder.layers.9.self_attn.out_proj.weight', 'text_model.encoder.layers.8.mlp.fc2.weight', 'text_model.encoder.layers.9.layer_norm2.weight', 'text_model.encoder.layers.9.layer_norm1.weight', 'text_model.encoder.layers.5.self_attn.out_proj.weight']
- This IS expected if you are initializing CLIPVisionModel from the checkpoint of a model trained on another task or with another architecture (e.g. initializing a BertForSequenceClassification model from a BertForPreTraining model).
- This IS NOT expected if you are initializing CLIPVisionModel from the checkpoint of a model that you expect to be exactly identical (initializing a BertForSequenceClassification model from a BertForSequenceClassification model).
2023-11-17 13:01:34 | INFO | model_worker | Register to controller
2023-11-17 13:01:34 | ERROR | stderr | Traceback (most recent call last):
2023-11-17 13:01:34 | ERROR | stderr |   File "/usr/local/lib/python3.10/dist-packages/urllib3/connection.py", line 203, in _new_conn
2023-11-17 13:01:34 | ERROR | stderr |     sock = connection.create_connection(
2023-11-17 13:01:34 | ERROR | stderr |   File "/usr/local/lib/python3.10/dist-packages/urllib3/util/connection.py", line 85, in create_connection
2023-11-17 13:01:34 | ERROR | stderr |     raise err
2023-11-17 13:01:34 | ERROR | stderr |   File "/usr/local/lib/python3.10/dist-packages/urllib3/util/connection.py", line 73, in create_connection
2023-11-17 13:01:34 | ERROR | stderr |     sock.connect(sa)
2023-11-17 13:01:34 | ERROR | stderr | ConnectionRefusedError: [Errno 111] Connection refused
2023-11-17 13:01:34 | ERROR | stderr | 
2023-11-17 13:01:34 | ERROR | stderr | The above exception was the direct cause of the following exception:
2023-11-17 13:01:34 | ERROR | stderr | 
2023-11-17 13:01:34 | ERROR | stderr | Traceback (most recent call last):
2023-11-17 13:01:34 | ERROR | stderr |   File "/usr/local/lib/python3.10/dist-packages/urllib3/connectionpool.py", line 791, in urlopen
2023-11-17 13:01:34 | ERROR | stderr |     response = self._make_request(
2023-11-17 13:01:34 | ERROR | stderr |   File "/usr/local/lib/python3.10/dist-packages/urllib3/connectionpool.py", line 497, in _make_request
2023-11-17 13:01:34 | ERROR | stderr |     conn.request(
2023-11-17 13:01:34 | ERROR | stderr |   File "/usr/local/lib/python3.10/dist-packages/urllib3/connection.py", line 395, in request
2023-11-17 13:01:34 | ERROR | stderr |     self.endheaders()
2023-11-17 13:01:34 | ERROR | stderr |   File "/usr/lib/python3.10/http/client.py", line 1278, in endheaders
2023-11-17 13:01:34 | ERROR | stderr |     self._send_output(message_body, encode_chunked=encode_chunked)
2023-11-17 13:01:34 | ERROR | stderr |   File "/usr/lib/python3.10/http/client.py", line 1038, in _send_output
2023-11-17 13:01:34 | ERROR | stderr |     self.send(msg)
2023-11-17 13:01:34 | ERROR | stderr |   File "/usr/lib/python3.10/http/client.py", line 976, in send
2023-11-17 13:01:34 | ERROR | stderr |     self.connect()
2023-11-17 13:01:34 | ERROR | stderr |   File "/usr/local/lib/python3.10/dist-packages/urllib3/connection.py", line 243, in connect
2023-11-17 13:01:34 | ERROR | stderr |     self.sock = self._new_conn()
2023-11-17 13:01:34 | ERROR | stderr |   File "/usr/local/lib/python3.10/dist-packages/urllib3/connection.py", line 218, in _new_conn
2023-11-17 13:01:34 | ERROR | stderr |     raise NewConnectionError(
2023-11-17 13:01:34 | ERROR | stderr | urllib3.exceptions.NewConnectionError: <urllib3.connection.HTTPConnection object at 0x79ef4d33bcd0>: Failed to establish a new connection: [Errno 111] Connection refused
2023-11-17 13:01:34 | ERROR | stderr | 
2023-11-17 13:01:34 | ERROR | stderr | The above exception was the direct cause of the following exception:
2023-11-17 13:01:34 | ERROR | stderr | 
2023-11-17 13:01:34 | ERROR | stderr | Traceback (most recent call last):
2023-11-17 13:01:34 | ERROR | stderr |   File "/usr/local/lib/python3.10/dist-packages/requests/adapters.py", line 486, in send
2023-11-17 13:01:34 | ERROR | stderr |     resp = conn.urlopen(
2023-11-17 13:01:34 | ERROR | stderr |   File "/usr/local/lib/python3.10/dist-packages/urllib3/connectionpool.py", line 845, in urlopen
2023-11-17 13:01:34 | ERROR | stderr |     retries = retries.increment(
2023-11-17 13:01:34 | ERROR | stderr |   File "/usr/local/lib/python3.10/dist-packages/urllib3/util/retry.py", line 515, in increment
2023-11-17 13:01:34 | ERROR | stderr |     raise MaxRetryError(_pool, url, reason) from reason  # type: ignore[arg-type]
2023-11-17 13:01:34 | ERROR | stderr | urllib3.exceptions.MaxRetryError: HTTPConnectionPool(host='localhost', port=10000): Max retries exceeded with url: /register_worker (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x79ef4d33bcd0>: Failed to establish a new connection: [Errno 111] Connection refused'))
2023-11-17 13:01:34 | ERROR | stderr | 
2023-11-17 13:01:34 | ERROR | stderr | During handling of the above exception, another exception occurred:
2023-11-17 13:01:34 | ERROR | stderr | 
2023-11-17 13:01:34 | ERROR | stderr | Traceback (most recent call last):
2023-11-17 13:01:34 | ERROR | stderr |   File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main
2023-11-17 13:01:34 | ERROR | stderr |     return _run_code(code, main_globals, None,
2023-11-17 13:01:34 | ERROR | stderr |   File "/usr/lib/python3.10/runpy.py", line 86, in _run_code
2023-11-17 13:01:34 | ERROR | stderr |     exec(code, run_globals)
2023-11-17 13:01:34 | ERROR | stderr |   File "/content/LLaVA-Med/llava/serve/model_worker.py", line 362, in <module>
2023-11-17 13:01:34 | ERROR | stderr |     worker = ModelWorker(args.controller_address,
2023-11-17 13:01:34 | ERROR | stderr |   File "/content/LLaVA-Med/llava/serve/model_worker.py", line 125, in __init__
2023-11-17 13:01:34 | ERROR | stderr |     self.register_to_controller()
2023-11-17 13:01:34 | ERROR | stderr |   File "/content/LLaVA-Med/llava/serve/model_worker.py", line 139, in register_to_controller
2023-11-17 13:01:34 | ERROR | stderr |     r = requests.post(url, json=data)
2023-11-17 13:01:34 | ERROR | stderr |   File "/usr/local/lib/python3.10/dist-packages/requests/api.py", line 115, in post
2023-11-17 13:01:34 | ERROR | stderr |     return request("post", url, data=data, json=json, **kwargs)
2023-11-17 13:01:34 | ERROR | stderr |   File "/usr/local/lib/python3.10/dist-packages/requests/api.py", line 59, in request
2023-11-17 13:01:34 | ERROR | stderr |     return session.request(method=method, url=url, **kwargs)
2023-11-17 13:01:34 | ERROR | stderr |   File "/usr/local/lib/python3.10/dist-packages/requests/sessions.py", line 589, in request
2023-11-17 13:01:34 | ERROR | stderr |     resp = self.send(prep, **send_kwargs)
2023-11-17 13:01:34 | ERROR | stderr |   File "/usr/local/lib/python3.10/dist-packages/requests/sessions.py", line 703, in send
2023-11-17 13:01:34 | ERROR | stderr |     r = adapter.send(request, **kwargs)
2023-11-17 13:01:34 | ERROR | stderr |   File "/usr/local/lib/python3.10/dist-packages/requests/adapters.py", line 519, in send
2023-11-17 13:01:34 | ERROR | stderr |     raise ConnectionError(e, request=request)
2023-11-17 13:01:34 | ERROR | stderr | requests.exceptions.ConnectionError: HTTPConnectionPool(host='localhost', port=10000): Max retries exceeded with url: /register_worker (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x79ef4d33bcd0>: Failed to establish a new connection: [Errno 111] Connection refused'))

The complete error log when trying to send a test message is -

2023-11-17 12:28:47.190995: E tensorflow/compiler/xla/stream_executor/cuda/cuda_dnn.cc:9342] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered
2023-11-17 12:28:47.191055: E tensorflow/compiler/xla/stream_executor/cuda/cuda_fft.cc:609] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered
2023-11-17 12:28:47.191120: E tensorflow/compiler/xla/stream_executor/cuda/cuda_blas.cc:1518] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered
2023-11-17 12:28:48.382592: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
Traceback (most recent call last):
  File "/usr/local/lib/python3.10/dist-packages/urllib3/connection.py", line 203, in _new_conn
    sock = connection.create_connection(
  File "/usr/local/lib/python3.10/dist-packages/urllib3/util/connection.py", line 85, in create_connection
    raise err
  File "/usr/local/lib/python3.10/dist-packages/urllib3/util/connection.py", line 73, in create_connection
    sock.connect(sa)
ConnectionRefusedError: [Errno 111] Connection refused

The above exception was the direct cause of the following exception:

Traceback (most recent call last):
  File "/usr/local/lib/python3.10/dist-packages/urllib3/connectionpool.py", line 791, in urlopen
    response = self._make_request(
  File "/usr/local/lib/python3.10/dist-packages/urllib3/connectionpool.py", line 497, in _make_request
    conn.request(
  File "/usr/local/lib/python3.10/dist-packages/urllib3/connection.py", line 395, in request
    self.endheaders()
  File "/usr/lib/python3.10/http/client.py", line 1278, in endheaders
    self._send_output(message_body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.10/http/client.py", line 1038, in _send_output
    self.send(msg)
  File "/usr/lib/python3.10/http/client.py", line 976, in send
    self.connect()
  File "/usr/local/lib/python3.10/dist-packages/urllib3/connection.py", line 243, in connect
    self.sock = self._new_conn()
  File "/usr/local/lib/python3.10/dist-packages/urllib3/connection.py", line 218, in _new_conn
    raise NewConnectionError(
urllib3.exceptions.NewConnectionError: <urllib3.connection.HTTPConnection object at 0x7f8a280a1a80>: Failed to establish a new connection: [Errno 111] Connection refused

The above exception was the direct cause of the following exception:

Traceback (most recent call last):
  File "/usr/local/lib/python3.10/dist-packages/requests/adapters.py", line 486, in send
    resp = conn.urlopen(
  File "/usr/local/lib/python3.10/dist-packages/urllib3/connectionpool.py", line 845, in urlopen
    retries = retries.increment(
  File "/usr/local/lib/python3.10/dist-packages/urllib3/util/retry.py", line 515, in increment
    raise MaxRetryError(_pool, url, reason) from reason  # type: ignore[arg-type]
urllib3.exceptions.MaxRetryError: HTTPConnectionPool(host='localhost', port=10000): Max retries exceeded with url: /refresh_all_workers (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7f8a280a1a80>: Failed to establish a new connection: [Errno 111] Connection refused'))

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main
    return _run_code(code, main_globals, None,
  File "/usr/lib/python3.10/runpy.py", line 86, in _run_code
    exec(code, run_globals)
  File "/content/LLaVA-Med/llava/serve/test_message.py", line 62, in <module>
    main()
  File "/content/LLaVA-Med/llava/serve/test_message.py", line 14, in main
    ret = requests.post(controller_addr + "/refresh_all_workers")
  File "/usr/local/lib/python3.10/dist-packages/requests/api.py", line 115, in post
    return request("post", url, data=data, json=json, **kwargs)
  File "/usr/local/lib/python3.10/dist-packages/requests/api.py", line 59, in request
    return session.request(method=method, url=url, **kwargs)
  File "/usr/local/lib/python3.10/dist-packages/requests/sessions.py", line 589, in request
    resp = self.send(prep, **send_kwargs)
  File "/usr/local/lib/python3.10/dist-packages/requests/sessions.py", line 703, in send
    r = adapter.send(request, **kwargs)
  File "/usr/local/lib/python3.10/dist-packages/requests/adapters.py", line 519, in send
    raise ConnectionError(e, request=request)
requests.exceptions.ConnectionError: HTTPConnectionPool(host='localhost', port=10000): Max retries exceeded with url: /refresh_all_workers (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7f8a280a1a80>: Failed to establish a new connection: [Errno 111] Connection refused'))

However when I try to launch a controller, it seems to work, however when I try to stop the execution of the code block, it takes some time to stop. The command I execuated in Google Colab is - !python -m llava.serve.controller --host 0.0.0.0 --port 10000

The complete log is -

2023-11-17 12:50:40.181860: E tensorflow/compiler/xla/stream_executor/cuda/cuda_dnn.cc:9342] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered
2023-11-17 12:50:40.181908: E tensorflow/compiler/xla/stream_executor/cuda/cuda_fft.cc:609] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered
2023-11-17 12:50:40.181976: E tensorflow/compiler/xla/stream_executor/cuda/cuda_blas.cc:1518] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered
2023-11-17 12:50:41.422420: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-11-17 12:50:43 | INFO | controller | args: Namespace(host='0.0.0.0', port=10000, dispatch_method='shortest_queue')
2023-11-17 12:50:43 | INFO | controller | Init controller
2023-11-17 12:50:43 | ERROR | stderr | INFO:     Started server process [7142]
2023-11-17 12:50:43 | ERROR | stderr | INFO:     Waiting for application startup.
2023-11-17 12:50:43 | ERROR | stderr | INFO:     Application startup complete.
2023-11-17 12:50:43 | ERROR | stderr | INFO:     Uvicorn running on http://0.0.0.0:10000/ (Press CTRL+C to quit)
2023-11-17 12:52:07 | ERROR | stderr | INFO:     Shutting down
2023-11-17 12:52:07 | ERROR | stderr | INFO:     Finished server process [7142]
2023-11-17 12:52:07 | ERROR | stderr | ERROR:    Traceback (most recent call last):
2023-11-17 12:52:07 | ERROR | stderr |   File "/usr/local/lib/python3.10/dist-packages/starlette/routing.py", line 686, in lifespan
2023-11-17 12:52:07 | ERROR | stderr |     await receive()
2023-11-17 12:52:07 | ERROR | stderr |   File "/usr/local/lib/python3.10/dist-packages/uvicorn/lifespan/on.py", line 137, in receive
2023-11-17 12:52:07 | ERROR | stderr |     return await self.receive_queue.get()
2023-11-17 12:52:07 | ERROR | stderr |   File "/usr/lib/python3.10/asyncio/queues.py", line 159, in get
2023-11-17 12:52:07 | ERROR | stderr |     await getter
2023-11-17 12:52:07 | ERROR | stderr | asyncio.exceptions.CancelledError
2023-11-17 12:52:07 | ERROR | stderr |

Edit - My Google Colab Notebook link for reference - https://colab.research.google.com/drive/1cSmBm4Jtm62zjdxvLC-xXFS400vvVpL5?usp=sharing

I am unable to figure out the source of error. Any help would be great. It has been days, I wanted to test it out, but even after spending so much money on renting paid gpus and colab pro+ just for knowlege, but I am continuously hitting roadblock.

Thank you Regards, Atul

Zoe-TAN commented 5 months ago

Hi, I successfully ran the code with our dataset on Linux just following the instructions. It seems that the error you posted is related to the connection with the Colab rather than the code. I am not that familiar with the Colab (Sorry for that) so I suggest you run it locally.