Closed johnomage closed 1 year ago
Can you please share the link of this notebook (If it's a public one)? Also, try latest 3.3.1 release, maybe you will have better luck with this version
Can you please share the link of this notebook (If it's a public one)? Also, try latest 3.3.1 release, maybe you will have better luck with this version
Hi Thanks for helping.
Here's the link to the notebook
I encountered the same error when I ran !pip install super-gradients==3.3.1
.
To resolve this, I installed the packages emerged in the error message.
!pip install pytube --upgrade
!pip install fastapi
!pip install kaleido
!pip install python-multipart
!pip install uvicorn
!pip install pyparsing==2.4.7
!pip install numpy==1.23.5
When I ran code below,
from super_gradients.training import Trainer
CHECKPOINT_DIR = 'CHECKPOINT-PATH'
trainer = Trainer(experiment_name='yolonas_run', ckpt_root_dir=CHECKPOINT_DIR)
I also got an error and installed packages in the error messages as in the previous step.
!pip install boto3
!pip install deprecated
!pip install coverage
!pip install sphinx-rtd-theme
!pip install torchmetrics
!pip install hydra-core
!pip install omegaconf
!pip install onnxruntime
!pip install onnx
!pip install einops
!pip install treelib
!pip install stringcase
!pip install rapidfuzz
!pip install json-tricks
!pip install onnx-simplifier
!pip install data-gradients
And it works for me! I hope it works for you too.
If you'd like to take a look my notebook, here's the link yolonas-dataverse - 1
I encountered the same error when I ran
!pip install super-gradients==3.3.1
.To resolve this, I installed the packages emerged in the error message.
!pip install pytube --upgrade !pip install fastapi !pip install kaleido !pip install python-multipart !pip install uvicorn !pip install pyparsing==2.4.7 !pip install numpy==1.23.5
When I ran code below,
from super_gradients.training import Trainer CHECKPOINT_DIR = 'CHECKPOINT-PATH' trainer = Trainer(experiment_name='yolonas_run', ckpt_root_dir=CHECKPOINT_DIR)
I also got an error and installed packages in the error messages as in the previous step.
!pip install boto3 !pip install deprecated !pip install coverage !pip install sphinx-rtd-theme !pip install torchmetrics !pip install hydra-core !pip install omegaconf !pip install onnxruntime !pip install onnx !pip install einops !pip install treelib !pip install stringcase !pip install rapidfuzz !pip install json-tricks !pip install onnx-simplifier !pip install data-gradients
And it works for me! I hope it works for you too.
If you'd like to take a look my notebook, here's the link yolonas-dataverse - 1
Thanks mate. This worked.
.train() is running smoothly.
I've discovered a new and simple solution. After downloading the super-gradients package, it's necessary to restart the notebook. This is mentioned in the YOLO-NAS notebook on Kaggle in the installation process. If you'd like to take a look, here's the link: Intro to SuperGradients + YOLONAS Starter Notebook
Hi sir can you help me ,DreamBooth_Stable_Diffusion.ipynb can't working
ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts. lida 0.0.10 requires kaleido, which is not installed. llmx 0.0.15a0 requires cohere, which is not installed. llmx 0.0.15a0 requires openai, which is not installed. llmx 0.0.15a0 requires tiktoken, which is not installed. tensorflow-probability 0.22.0 requires typing-extensions<4.6.0, but you have typing-extensions 4.8.0 which is incompatible.
WARNING[XFORMERS]: xFormers can't load C++/CUDA extensions. xFormers was built for:
PyTorch 2.1.0+cu121 with CUDA 1201 (you have 2.1.0+cu118)
Python 3.10.13 (you have 3.10.12)
Please reinstall xformers (see https://github.com/facebookresearch/xformers#installing-xformers)
Memory-efficient attention, SwiGLU, sparse and more won't be available.
Set XFORMERS_MORE_DETAILS=1 for more details
2023-11-06 05:57:58.909288: E tensorflow/compiler/xla/stream_executor/cuda/cuda_dnn.cc:9342] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered
2023-11-06 05:57:58.909344: E tensorflow/compiler/xla/stream_executor/cuda/cuda_fft.cc:609] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered
2023-11-06 05:57:58.909391: E tensorflow/compiler/xla/stream_executor/cuda/cuda_blas.cc:1518] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered
2023-11-06 05:58:00.548575: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_hf_folder.py:98: UserWarning: A token has been found in /root/.huggingface/token
. This is the old path where tokens were stored. The new location is /root/.cache/huggingface/token
which is configurable using HF_HOME
environment variable. Your token has been copied to this new location. You can now safely delete the old token file manually or use huggingface-cli logout
.
warnings.warn(
Downloading (…)lve/main/config.json: 100% 547/547 [00:00<00:00, 2.91MB/s]
Downloading (…)ch_model.safetensors: 100% 335M/335M [00:04<00:00, 72.4MB/s]
Downloading (…)p16/model_index.json: 100% 543/543 [00:00<00:00, 3.32MB/s]
text_encoder/model.safetensors not found
Fetching 15 files: 0% 0/15 [00:00<?, ?it/s]
Downloading pytorch_model.bin: 0% 0.00/246M [00:00<?, ?B/s]
Downloading (…)_encoder/config.json: 100% 636/636 [00:00<00:00, 3.80MB/s]
Downloading (…)tokenizer/merges.txt: 0% 0.00/525k [00:00<?, ?B/s]
Downloading (…)rocessor_config.json: 100% 342/342 [00:00<00:00, 2.08MB/s] Fetching 15 files: 7% 1/15 [00:00<00:11, 1.20it/s]
Downloading (…)_checker/config.json: 100% 4.70k/4.70k [00:00<00:00, 25.3MB/s]
Downloading (…)cheduler_config.json: 100% 307/307 [00:00<00:00, 2.06MB/s]
Downloading (…)cial_tokens_map.json: 100% 472/472 [00:00<00:00, 3.31MB/s]
Downloading pytorch_model.bin: 0% 0.00/608M [00:00<?, ?B/s]
Downloading pytorch_model.bin: 2% 10.5M/608M [00:00<00:05, 99.9MB/s]
Downloading (…)okenizer_config.json: 100% 822/822 [00:00<00:00, 3.81MB/s]
Downloading (…)tokenizer/merges.txt: 100% 525k/525k [00:00<00:00, 2.58MB/s]
Downloading pytorch_model.bin: 4% 10.5M/246M [00:00<00:05, 43.2MB/s]
Downloading pytorch_model.bin: 5% 31.5M/608M [00:00<00:03, 150MB/s]
Downloading (…)on_pytorch_model.bin: 0% 0.00/1.72G [00:00<?, ?B/s]
Downloading (…)tokenizer/vocab.json: 0% 0.00/1.06M [00:00<?, ?B/s]
Downloading (…)cc2/unet/config.json: 100% 806/806 [00:00<00:00, 3.36MB/s]
Downloading pytorch_model.bin: 13% 31.5M/246M [00:00<00:02, 89.2MB/s]
Downloading pytorch_model.bin: 10% 62.9M/608M [00:00<00:02, 202MB/s]
Downloading (…)tokenizer/vocab.json: 100% 1.06M/1.06M [00:00<00:00, 7.69MB/s]
Downloading (…)9cc2/vae/config.json: 100% 609/609 [00:00<00:00, 2.78MB/s]
Downloading (…)on_pytorch_model.bin: 1% 10.5M/1.72G [00:00<00:38, 44.7MB/s] Downloading pytorch_model.bin: 21% 52.4M/246M [00:00<00:01, 97.9MB/s]
Downloading pytorch_model.bin: 14% 83.9M/608M [00:00<00:03, 150MB/s]
Downloading (…)on_pytorch_model.bin: 1% 21.0M/1.72G [00:00<00:26, 62.9MB/s] Downloading pytorch_model.bin: 30% 73.4M/246M [00:00<00:01, 104MB/s]
Downloading (…)on_pytorch_model.bin: 0% 0.00/167M [00:00<?, ?B/s]
Downloading pytorch_model.bin: 17% 105M/608M [00:00<00:03, 134MB/s]
Downloading (…)on_pytorch_model.bin: 2% 41.9M/1.72G [00:00<00:19, 84.7MB/s] Downloading pytorch_model.bin: 38% 94.4M/246M [00:00<00:01, 116MB/s]
Downloading pytorch_model.bin: 21% 126M/608M [00:00<00:03, 136MB/s]
Downloading (…)on_pytorch_model.bin: 4% 62.9M/1.72G [00:00<00:16, 103MB/s]
Downloading (…)on_pytorch_model.bin: 6% 10.5M/167M [00:00<00:04, 39.2MB/s] Downloading pytorch_model.bin: 47% 115M/246M [00:01<00:01, 110MB/s]
Downloading (…)on_pytorch_model.bin: 13% 21.0M/167M [00:00<00:02, 57.9MB/s]
Downloading pytorch_model.bin: 24% 147M/608M [00:01<00:03, 119MB/s]
Downloading (…)on_pytorch_model.bin: 5% 83.9M/1.72G [00:00<00:15, 107MB/s] Downloading pytorch_model.bin: 55% 136M/246M [00:01<00:01, 96.9MB/s]
Downloading (…)on_pytorch_model.bin: 25% 41.9M/167M [00:00<00:01, 69.9MB/s]
Downloading (…)on_pytorch_model.bin: 6% 105M/1.72G [00:01<00:17, 93.8MB/s]
Downloading pytorch_model.bin: 28% 168M/608M [00:01<00:04, 95.1MB/s] Downloading pytorch_model.bin: 60% 147M/246M [00:01<00:01, 89.8MB/s]
Downloading (…)on_pytorch_model.bin: 31% 52.4M/167M [00:00<00:01, 65.3MB/s]
Downloading (…)on_pytorch_model.bin: 7% 115M/1.72G [00:01<00:19, 80.7MB/s] Downloading pytorch_model.bin: 64% 157M/246M [00:01<00:01, 84.5MB/s]
Downloading (…)on_pytorch_model.bin: 38% 62.9M/167M [00:00<00:01, 65.0MB/s]
Downloading (…)on_pytorch_model.bin: 7% 126M/1.72G [00:01<00:19, 80.0MB/s]
Downloading pytorch_model.bin: 31% 189M/608M [00:01<00:05, 78.6MB/s] Downloading pytorch_model.bin: 68% 168M/246M [00:01<00:01, 75.5MB/s]
Downloading (…)on_pytorch_model.bin: 8% 136M/1.72G [00:01<00:21, 74.0MB/s]
Downloading (…)on_pytorch_model.bin: 44% 73.4M/167M [00:01<00:01, 59.8MB/s]
Downloading pytorch_model.bin: 33% 199M/608M [00:01<00:05, 76.2MB/s] Downloading pytorch_model.bin: 72% 178M/246M [00:02<00:00, 69.7MB/s]
Downloading (…)on_pytorch_model.bin: 50% 83.9M/167M [00:01<00:01, 62.1MB/s]
Downloading (…)on_pytorch_model.bin: 9% 147M/1.72G [00:01<00:22, 70.4MB/s]
Downloading pytorch_model.bin: 34% 210M/608M [00:02<00:05, 72.9MB/s] Downloading pytorch_model.bin: 77% 189M/246M [00:02<00:00, 68.6MB/s]
Downloading (…)on_pytorch_model.bin: 56% 94.4M/167M [00:01<00:01, 63.6MB/s]
Downloading (…)on_pytorch_model.bin: 9% 157M/1.72G [00:01<00:22, 70.7MB/s]
Downloading pytorch_model.bin: 36% 220M/608M [00:02<00:05, 70.6MB/s] Downloading pytorch_model.bin: 81% 199M/246M [00:02<00:00, 67.1MB/s]
Downloading (…)on_pytorch_model.bin: 10% 168M/1.72G [00:02<00:21, 70.7MB/s]
Downloading (…)on_pytorch_model.bin: 63% 105M/167M [00:01<00:00, 63.9MB/s]
Downloading pytorch_model.bin: 38% 231M/608M [00:02<00:06, 61.8MB/s] Downloading pytorch_model.bin: 85% 210M/246M [00:02<00:00, 45.6MB/s]
Downloading (…)on_pytorch_model.bin: 10% 178M/1.72G [00:02<00:32, 47.9MB/s]
Downloading pytorch_model.bin: 40% 241M/608M [00:02<00:06, 53.1MB/s]
Downloading (…)on_pytorch_model.bin: 69% 115M/167M [00:02<00:01, 43.4MB/s]
Downloading (…)on_pytorch_model.bin: 11% 189M/1.72G [00:02<00:28, 54.1MB/s]
Downloading pytorch_model.bin: 41% 252M/608M [00:03<00:06, 53.3MB/s]
Downloading (…)on_pytorch_model.bin: 75% 126M/167M [00:02<00:00, 45.6MB/s] Downloading pytorch_model.bin: 89% 220M/246M [00:03<00:00, 42.1MB/s]
Downloading (…)on_pytorch_model.bin: 12% 199M/1.72G [00:02<00:25, 58.8MB/s]
Downloading (…)on_pytorch_model.bin: 12% 210M/1.72G [00:02<00:22, 67.5MB/s] Downloading pytorch_model.bin: 94% 231M/246M [00:03<00:00, 46.0MB/s]
Downloading pytorch_model.bin: 43% 262M/608M [00:03<00:06, 50.7MB/s]
Downloading (…)on_pytorch_model.bin: 13% 220M/1.72G [00:03<00:20, 71.7MB/s]
Downloading pytorch_model.bin: 45% 273M/608M [00:03<00:06, 49.3MB/s]
Downloading (…)on_pytorch_model.bin: 88% 147M/167M [00:02<00:00, 45.6MB/s] Downloading pytorch_model.bin: 98% 241M/246M [00:03<00:00, 44.4MB/s]
Downloading (…)on_pytorch_model.bin: 13% 231M/1.72G [00:03<00:24, 61.0MB/s] Downloading pytorch_model.bin: 100% 246M/246M [00:03<00:00, 43.6MB/s]
Downloading pytorch_model.bin: 100% 246M/246M [00:03<00:00, 66.4MB/s]
Downloading (…)on_pytorch_model.bin: 94% 157M/167M [00:02<00:00, 49.0MB/s]
Downloading (…)on_pytorch_model.bin: 14% 241M/1.72G [00:03<00:23, 62.3MB/s]
Downloading pytorch_model.bin: 48% 294M/608M [00:03<00:05, 60.8MB/s]
Downloading (…)on_pytorch_model.bin: 100% 167M/167M [00:03<00:00, 55.8MB/s]
Downloading (…)on_pytorch_model.bin: 100% 167M/167M [00:03<00:00, 54.4MB/s]
Downloading (…)on_pytorch_model.bin: 16% 273M/1.72G [00:03<00:15, 93.3MB/s]
Downloading pytorch_model.bin: 52% 315M/608M [00:03<00:03, 77.5MB/s]
Downloading (…)on_pytorch_model.bin: 17% 294M/1.72G [00:03<00:13, 109MB/s]
Downloading pytorch_model.bin: 55% 336M/608M [00:04<00:02, 93.0MB/s]
Downloading (…)on_pytorch_model.bin: 18% 315M/1.72G [00:03<00:12, 117MB/s]
Downloading pytorch_model.bin: 59% 357M/608M [00:04<00:02, 107MB/s]
Downloading pytorch_model.bin: 62% 377M/608M [00:04<00:01, 116MB/s]
Downloading (…)on_pytorch_model.bin: 20% 336M/1.72G [00:04<00:11, 122MB/s]
Downloading (…)on_pytorch_model.bin: 21% 357M/1.72G [00:04<00:10, 132MB/s]
Downloading pytorch_model.bin: 66% 398M/608M [00:04<00:01, 122MB/s]
Downloading (…)on_pytorch_model.bin: 22% 377M/1.72G [00:04<00:09, 136MB/s]
Downloading pytorch_model.bin: 69% 419M/608M [00:04<00:01, 128MB/s]
Downloading (…)on_pytorch_model.bin: 23% 398M/1.72G [00:04<00:09, 138MB/s]
Downloading pytorch_model.bin: 72% 440M/608M [00:04<00:01, 135MB/s]
Downloading (…)on_pytorch_model.bin: 24% 419M/1.72G [00:04<00:09, 136MB/s]
Downloading pytorch_model.bin: 76% 461M/608M [00:04<00:01, 132MB/s]
Downloading (…)on_pytorch_model.bin: 26% 440M/1.72G [00:04<00:08, 143MB/s]
Downloading pytorch_model.bin: 79% 482M/608M [00:05<00:00, 129MB/s]
Downloading (…)on_pytorch_model.bin: 27% 461M/1.72G [00:05<00:08, 141MB/s]
Downloading pytorch_model.bin: 83% 503M/608M [00:05<00:00, 137MB/s]
Downloading (…)on_pytorch_model.bin: 28% 482M/1.72G [00:05<00:08, 144MB/s]
Downloading pytorch_model.bin: 86% 524M/608M [00:05<00:00, 140MB/s]
Downloading (…)on_pytorch_model.bin: 29% 503M/1.72G [00:05<00:08, 144MB/s]
Downloading pytorch_model.bin: 90% 545M/608M [00:05<00:00, 144MB/s]
Downloading (…)on_pytorch_model.bin: 30% 524M/1.72G [00:05<00:08, 139MB/s]
Downloading pytorch_model.bin: 93% 566M/608M [00:05<00:00, 123MB/s]
Downloading pytorch_model.bin: 97% 587M/608M [00:05<00:00, 134MB/s]
Downloading (…)on_pytorch_model.bin: 32% 545M/1.72G [00:05<00:09, 120MB/s]
Downloading pytorch_model.bin: 100% 608M/608M [00:06<00:00, 141MB/s]
Downloading pytorch_model.bin: 100% 608M/608M [00:06<00:00, 100MB/s] Fetching 15 files: 27% 4/15 [00:06<00:19, 1.81s/it]
Downloading (…)on_pytorch_model.bin: 34% 587M/1.72G [00:05<00:07, 143MB/s]
Downloading (…)on_pytorch_model.bin: 36% 619M/1.72G [00:06<00:06, 165MB/s]
Downloading (…)on_pytorch_model.bin: 38% 650M/1.72G [00:06<00:05, 182MB/s]
Downloading (…)on_pytorch_model.bin: 40% 682M/1.72G [00:06<00:05, 191MB/s]
Downloading (…)on_pytorch_model.bin: 41% 703M/1.72G [00:06<00:05, 173MB/s]
Downloading (…)on_pytorch_model.bin: 42% 724M/1.72G [00:06<00:05, 173MB/s]
Downloading (…)on_pytorch_model.bin: 44% 755M/1.72G [00:06<00:04, 195MB/s]
Downloading (…)on_pytorch_model.bin: 46% 786M/1.72G [00:06<00:04, 208MB/s]
Downloading (…)on_pytorch_model.bin: 48% 818M/1.72G [00:07<00:04, 220MB/s]
Downloading (…)on_pytorch_model.bin: 49% 849M/1.72G [00:07<00:03, 222MB/s]
Downloading (…)on_pytorch_model.bin: 51% 881M/1.72G [00:07<00:03, 225MB/s]
Downloading (…)on_pytorch_model.bin: 53% 912M/1.72G [00:07<00:03, 225MB/s]
Downloading (…)on_pytorch_model.bin: 55% 944M/1.72G [00:07<00:03, 227MB/s]
Downloading (…)on_pytorch_model.bin: 57% 975M/1.72G [00:07<00:03, 233MB/s]
Downloading (…)on_pytorch_model.bin: 59% 1.01G/1.72G [00:07<00:03, 233MB/s]
Downloading (…)on_pytorch_model.bin: 60% 1.04G/1.72G [00:07<00:02, 228MB/s]
Downloading (…)on_pytorch_model.bin: 62% 1.07G/1.72G [00:08<00:02, 238MB/s]
Downloading (…)on_pytorch_model.bin: 64% 1.10G/1.72G [00:08<00:02, 241MB/s]
Downloading (…)on_pytorch_model.bin: 66% 1.13G/1.72G [00:08<00:02, 249MB/s]
Downloading (…)on_pytorch_model.bin: 68% 1.16G/1.72G [00:08<00:02, 249MB/s]
Downloading (…)on_pytorch_model.bin: 70% 1.20G/1.72G [00:08<00:02, 242MB/s]
Downloading (…)on_pytorch_model.bin: 71% 1.23G/1.72G [00:08<00:02, 218MB/s]
Downloading (…)on_pytorch_model.bin: 73% 1.26G/1.72G [00:08<00:02, 215MB/s]
Downloading (…)on_pytorch_model.bin: 75% 1.29G/1.72G [00:09<00:01, 217MB/s]
Downloading (…)on_pytorch_model.bin: 77% 1.32G/1.72G [00:09<00:01, 222MB/s]
Downloading (…)on_pytorch_model.bin: 79% 1.35G/1.72G [00:09<00:01, 227MB/s]
Downloading (…)on_pytorch_model.bin: 81% 1.38G/1.72G [00:09<00:01, 222MB/s]
Downloading (…)on_pytorch_model.bin: 82% 1.42G/1.72G [00:09<00:01, 228MB/s]
Downloading (…)on_pytorch_model.bin: 84% 1.45G/1.72G [00:09<00:01, 235MB/s]
Downloading (…)on_pytorch_model.bin: 86% 1.48G/1.72G [00:09<00:01, 235MB/s]
Downloading (…)on_pytorch_model.bin: 88% 1.51G/1.72G [00:09<00:00, 243MB/s]
Downloading (…)on_pytorch_model.bin: 90% 1.54G/1.72G [00:10<00:00, 254MB/s]
Downloading (…)on_pytorch_model.bin: 91% 1.57G/1.72G [00:10<00:00, 267MB/s]
Downloading (…)on_pytorch_model.bin: 93% 1.60G/1.72G [00:10<00:00, 271MB/s]
Downloading (…)on_pytorch_model.bin: 95% 1.64G/1.72G [00:10<00:00, 266MB/s]
Downloading (…)on_pytorch_model.bin: 97% 1.67G/1.72G [00:10<00:00, 273MB/s]
Downloading (…)on_pytorch_model.bin: 100% 1.72G/1.72G [00:10<00:00, 160MB/s]
Fetching 15 files: 100% 15/15 [00:11<00:00, 1.26it/s]
/usr/local/lib/python3.10/dist-packages/transformers/models/clip/feature_extraction_clip.py:28: FutureWarning: The class CLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please use CLIPImageProcessor instead.
warnings.warn(
You have disabled the safety checker for <class 'diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline'> by passing safety_checker=None
. Ensure that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered results in services or applications open to the public. Both the diffusers team and Hugging Face strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling it only for use-cases that involve analyzing network behavior or auditing its results. For more information, please have a look at https://github.com/huggingface/diffusers/pull/254 .
Traceback (most recent call last):
File "/content/train_dreambooth.py", line 869, in memory_efficient_attention_forward
with inputs:
query : shape=(1, 2, 1, 40) (torch.float32)
key : shape=(1, 2, 1, 40) (torch.float32)
value : shape=(1, 2, 1, 40) (torch.float32)
attn_bias : <class 'NoneType'>
p : 0.0
decoderF
is not supported because:
xFormers wasn't build with CUDA support
attn_bias type is <class 'NoneType'>
operator wasn't built - see python -m xformers.info
for more info
flshattF@0.0.0
is not supported because:
xFormers wasn't build with CUDA support
requires device with capability > (8, 0) but your GPU has capability (7, 5) (too old)
dtype=torch.float32 (supported: {torch.float16, torch.bfloat16})
operator wasn't built - see python -m xformers.info
for more info
tritonflashattF
is not supported because:
xFormers wasn't build with CUDA support
requires device with capability > (8, 0) but your GPU has capability (7, 5) (too old)
dtype=torch.float32 (supported: {torch.float16, torch.bfloat16})
operator wasn't built - see python -m xformers.info
for more info
triton is not available
requires GPU with sm80 minimum compute capacity, e.g., A100/H100/L4
Only work on pre-MLIR triton for now
cutlassF
is not supported because:
xFormers wasn't build with CUDA support
operator wasn't built - see python -m xformers.info
for more info
smallkF
is not supported because:
max(query.shape[-1] != value.shape[-1]) > 32
xFormers wasn't build with CUDA support
operator wasn't built - see python -m xformers.info
for more info
unsupported embed per head: 40
🐛 Describe the bug
Error installing super-gradients (or super_gradients) on colab.
I got these errors/warnings while pip-installing SuperGradients library:
!pip install super_gradients==3.2.0
ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts.
lida 0.0.10 requires fastapi, which is not installed.
lida 0.0.10 requires kaleido, which is not installed.
lida 0.0.10 requires python-multipart, which is not installed.
lida 0.0.10 requires uvicorn, which is not installed.
tensorflow 2.14.0 requires numpy>=1.23.5, but you have numpy 1.23.0 which is incompatible.
Successfully installed Deprecated-1.2.14 antlr4-python3-runtime-4.9.3 boto3-1.28.73 botocore-1.31.73 coloredlogs-15.0.1 coverage-5.3.1 docutils-0.17.1 einops-0.3.2 humanfriendly-10.0 hydra-core-1.3.2 jmespath-1.0.1 json-tricks-3.16.1 numpy-1.23.0 omegaconf-2.3.0 onnx-1.13.0 onnx-simplifier-0.4.35 onnxruntime-1.13.1 pyDeprecate-0.3.2 pycocotools-2.0.6 pyparsing-2.4.5 rapidfuzz-3.4.0 s3transfer-0.7.0 sphinx-4.0.3 sphinx-rtd-theme-1.3.0 sphinxcontrib-applehelp-1.0.4 sphinxcontrib-devhelp-1.0.2 sphinxcontrib-htmlhelp-2.0.1 sphinxcontrib-jquery-4.1 sphinxcontrib-qthelp-1.0.3 sphinxcontrib-serializinghtml-1.1.5 stringcase-1.2.0 super_gradients-3.2.0 termcolor-1.1.0 torchmetrics-0.8.0 treelib-1.6.1
WARNING: The following packages were previously imported in this runtime:
[numpy,pydevd_plugins,pyparsing,sphinxcontrib]
You must restart the runtime in order to use newly installed versions.
[2023-10-29 02:25:13] INFO - crash_tips_setup.py - Crash tips is enabled. You can set your environment variable to CRASH_HANDLER=FALSE to disable it
The console stream is logged into /root/sg_logs/console.log
[2023-10-29 02:25:15] WARNING - init.py - Failed to import pytorch_quantization
[2023-10-29 02:25:17] INFO - utils.py - NumExpr defaulting to 2 threads.
/usr/local/lib/python3.10/dist-packages/_distutils_hack/init.py:33: UserWarning: Setuptools is replacing distutils.
warnings.warn("Setuptools is replacing distutils.")
[2023-10-29 02:25:19] WARNING - calibrator.py - Failed to import pytorch_quantization
[2023-10-29 02:25:19] WARNING - export.py - Failed to import pytorch_quantization
[2023-10-29 02:25:19] WARNING - selective_quantization_utils.py - Failed to import pytorch_quantization
[2023-10-29 02:25:19] WARNING - env_sanity_check.py - Failed to verify installed packages: boto3 required but not found
[2023-10-29 02:25:19] WARNING - env_sanity_check.py - Failed to verify installed packages: deprecated required but not found
[2023-10-29 02:25:19] WARNING - env_sanity_check.py - Failed to verify installed packages: coverage required but not found
[2023-10-29 02:25:19] WARNING - env_sanity_check.py - Failed to verify installed packages: sphinx-rtd-theme required but not found
[2023-10-29 02:25:19] WARNING - env_sanity_check.py - Failed to verify installed packages: torchmetrics required but not found
[2023-10-29 02:25:19] WARNING - env_sanity_check.py - Failed to verify installed packages: hydra-core required but not found
[2023-10-29 02:25:19] WARNING - env_sanity_check.py - Failed to verify installed packages: omegaconf required but not found
[2023-10-29 02:25:19] WARNING - env_sanity_check.py - Failed to verify installed packages: onnxruntime required but not found
[2023-10-29 02:25:19] WARNING - env_sanity_check.py - Failed to verify installed packages: onnx required but not found
[2023-10-29 02:25:19] WARNING - env_sanity_check.py - Failed to verify installed packages: einops required but not found
[2023-10-29 02:25:19] WARNING - env_sanity_check.py - Failed to verify installed packages: treelib required but not found
[2023-10-29 02:25:19] WARNING - env_sanity_check.py - Failed to verify installed packages: stringcase required but not found
[2023-10-29 02:25:19] WARNING - env_sanity_check.py - Failed to verify installed packages: rapidfuzz required but not found
[2023-10-29 02:25:19] WARNING - env_sanity_check.py - Failed to verify installed packages: json-tricks required but not found
[2023-10-29 02:25:19] WARNING - env_sanity_check.py - Failed to verify installed packages: onnx-simplifier required but not found.
Also for:
i got: ValueError: 'accuracy' is not in list.
Anyone else got these?
Versions
--2023-10-29 02:44:04-- https://raw.githubusercontent.com/pytorch/pytorch/main/torch/utils/collect_env.py Resolving raw.githubusercontent.com (raw.githubusercontent.com)... 185.199.108.133, 185.199.109.133, 185.199.110.133, ... Connecting to raw.githubusercontent.com (raw.githubusercontent.com)|185.199.108.133|:443... connected. HTTP request sent, awaiting response... 200 OK Length: 21737 (21K) [text/plain] Saving to: ‘collect_env.py’
collect_env.py 100%[===================>] 21.23K --.-KB/s in 0s
2023-10-29 02:44:04 (67.3 MB/s) - ‘collect_env.py’ saved [21737/21737]
Collecting environment information... PyTorch version: 2.1.0+cu118 Is debug build: False CUDA used to build PyTorch: 11.8 ROCM used to build PyTorch: N/A
OS: Ubuntu 22.04.2 LTS (x86_64) GCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0 Clang version: 14.0.0-1ubuntu1.1 CMake version: version 3.27.7 Libc version: glibc-2.35
Python version: 3.10.12 (main, Jun 11 2023, 05:26:28) [GCC 11.4.0] (64-bit runtime) Python platform: Linux-5.15.120+-x86_64-with-glibc2.35 Is CUDA available: True CUDA runtime version: 11.8.89 CUDA_MODULE_LOADING set to: LAZY GPU models and configuration: GPU 0: Tesla T4 Nvidia driver version: 525.105.17 cuDNN version: Probably one of the following: /usr/lib/x86_64-linux-gnu/libcudnn.so.8.9.0 /usr/lib/x86_64-linux-gnu/libcudnn_adv_infer.so.8.9.0 /usr/lib/x86_64-linux-gnu/libcudnn_adv_train.so.8.9.0 /usr/lib/x86_64-linux-gnu/libcudnn_cnn_infer.so.8.9.0 /usr/lib/x86_64-linux-gnu/libcudnn_cnn_train.so.8.9.0 /usr/lib/x86_64-linux-gnu/libcudnn_ops_infer.so.8.9.0 /usr/lib/x86_64-linux-gnu/libcudnn_ops_train.so.8.9.0 HIP runtime version: N/A MIOpen runtime version: N/A Is XNNPACK available: True
CPU: Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Address sizes: 46 bits physical, 48 bits virtual Byte Order: Little Endian CPU(s): 2 On-line CPU(s) list: 0,1 Vendor ID: GenuineIntel Model name: Intel(R) Xeon(R) CPU @ 2.20GHz CPU family: 6 Model: 79 Thread(s) per core: 2 Core(s) per socket: 1 Socket(s): 1 Stepping: 0 BogoMIPS: 4399.99 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss ht syscall nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopology nonstop_tsc cpuid tsc_known_freq pni pclmulqdq ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch invpcid_single ssbd ibrs ibpb stibp fsgsbase tsc_adjust bmi1 hle avx2 smep bmi2 erms invpcid rtm rdseed adx smap xsaveopt arat md_clear arch_capabilities Hypervisor vendor: KVM Virtualization type: full L1d cache: 32 KiB (1 instance) L1i cache: 32 KiB (1 instance) L2 cache: 256 KiB (1 instance) L3 cache: 55 MiB (1 instance) NUMA node(s): 1 NUMA node0 CPU(s): 0,1 Vulnerability Itlb multihit: Not affected Vulnerability L1tf: Mitigation; PTE Inversion Vulnerability Mds: Vulnerable; SMT Host state unknown Vulnerability Meltdown: Vulnerable Vulnerability Mmio stale data: Vulnerable Vulnerability Retbleed: Vulnerable Vulnerability Spec store bypass: Vulnerable Vulnerability Spectre v1: Vulnerable: __user pointer sanitization and usercopy barriers only; no swapgs barriers Vulnerability Spectre v2: Vulnerable, IBPB: disabled, STIBP: disabled, PBRSB-eIBRS: Not affected Vulnerability Srbds: Not affected Vulnerability Tsx async abort: Vulnerable
Versions of relevant libraries: [pip3] numpy==1.23.0 [pip3] onnx==1.13.0 [pip3] onnx-simplifier==0.4.35 [pip3] onnxruntime==1.13.1 [pip3] torch==2.1.0+cu118 [pip3] torchaudio==2.1.0+cu118 [pip3] torchdata==0.7.0 [pip3] torchmetrics==0.8.0 [pip3] torchsummary==1.5.1 [pip3] torchtext==0.16.0 [pip3] torchvision==0.16.0+cu118 [pip3] triton==2.1.0 [conda] Could not collect