Open obriensystems opened 3 days ago
michael@14900c MINGW64 /c/wse_github/obrienlabsdev/machine-learning (main)
$ ./build.sh
#0 building with "desktop-linux" instance using docker driver
#1 [internal] load build definition from Dockerfile
#1 transferring dockerfile: 486B done
#1 DONE 0.0s
#2 [internal] load metadata for docker.io/tensorflow/tensorflow:2.14.0-gpu
#2 DONE 0.3s
#3 [internal] load .dockerignore
#3 transferring context: 2B done
#3 DONE 0.0s
#4 [1/3] FROM docker.io/tensorflow/tensorflow:2.14.0-gpu@sha256:64602abcd8cc4f4bdd6268ca0abc39e6d37113d700886afd15f6dd151210b206
#4 DONE 0.0s
#5 [internal] load build context
#5 transferring context: 4.04kB done
#5 DONE 0.0s
#6 [2/3] WORKDIR /src
#6 CACHED
#7 [3/3] COPY /src/tflow.py .
#7 DONE 0.0s
#8 exporting to image
#8 exporting layers
#8 exporting layers 0.0s done
#8 writing image sha256:c0831f4a03856d372fedde7d563a2ab6b3983a1f88dd93e34ca348d8b981d586 done
#8 naming to docker.io/library/ml-tensorflow-win done
#8 DONE 0.1s
2024-12-01 02:23:10.019137: I tensorflow/core/util/port.cc:111] oneDNN custom operations are on. You may see slightly different numerical results due to floating-point round-off errors from different computation orders. To turn them off, set the environment variable `TF_ENABLE_ONEDNN_OPTS=0`.
2024-12-01 02:23:10.035434: E tensorflow/compiler/xla/stream_executor/cuda/cuda_dnn.cc:9342] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered
2024-12-01 02:23:10.035467: E tensorflow/compiler/xla/stream_executor/cuda/cuda_fft.cc:609] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered
2024-12-01 02:23:10.035478: E tensorflow/compiler/xla/stream_executor/cuda/cuda_blas.cc:1518] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered
2024-12-01 02:23:10.039315: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 AVX_VNNI FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2024-12-01 02:23:10.827037: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 02:23:10.830004: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 02:23:10.830034: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 02:23:10.830916: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 02:23:10.830939: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 02:23:10.830946: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 02:23:10.930081: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 02:23:10.930118: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 02:23:10.930123: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1977] Could not identify NUMA node of platform GPU id 0, defaulting to 0. Your kernel may not have been built with NUMA support.
2024-12-01 02:23:10.930139: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 02:23:10.930163: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1886] Created device /job:localhost/replica:0/task:0/device:GPU:0 with 46009 MB memory: -> device: 0, name: NVIDIA RTX A6000, pci bus id: 0000:01:00.0, compute capability: 8.6
Downloading data from https://www.cs.toronto.edu/~kriz/cifar-100-python.tar.gz
169001437/169001437 [==============================] - 3s 0us/step
Epoch 1/25
2024-12-01 02:23:22.458429: I tensorflow/compiler/xla/stream_executor/cuda/cuda_dnn.cc:442] Loaded cuDNN version 8600
2024-12-01 02:23:24.912260: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x9f36010 initialized for platform CUDA (this does not guarantee that XLA will be used). Devices:
2024-12-01 02:23:24.912284: I tensorflow/compiler/xla/service/service.cc:176] StreamExecutor device (0): NVIDIA RTX A6000, Compute Capability 8.6
2024-12-01 02:23:24.915227: I tensorflow/compiler/mlir/tensorflow/utils/dump_mlir_util.cc:269] disabling MLIR crash reproducer, set env var `MLIR_CRASH_REPRODUCER_DIRECTORY` to enable.
2024-12-01 02:23:24.966475: I ./tensorflow/compiler/jit/device_compiler.h:186] Compiled cluster using XLA! This line is logged at most once for the lifetime of the process.
9/9 [==============================] - 30s 694ms/step - loss: 5.8745 - accuracy: 0.0215
Epoch 2/25
9/9 [==============================] - 3s 329ms/step - loss: 4.4032 - accuracy: 0.0592
Epoch 3/25
9/9 [==============================] - 3s 329ms/step - loss: 3.9991 - accuracy: 0.0968
Epoch 4/25
9/9 [==============================] - 3s 330ms/step - loss: 3.7315 - accuracy: 0.1350
Epoch 5/25
9/9 [==============================] - 3s 330ms/step - loss: 3.4930 - accuracy: 0.1771
Epoch 6/25
9/9 [==============================] - 3s 330ms/step - loss: 3.2226 - accuracy: 0.2223
Epoch 7/25
9/9 [==============================] - 3s 333ms/step - loss: 3.0065 - accuracy: 0.2659
Epoch 8/25
9/9 [==============================] - 3s 331ms/step - loss: 2.7064 - accuracy: 0.3253
Epoch 9/25
9/9 [==============================] - 3s 332ms/step - loss: 2.4498 - accuracy: 0.3777
Epoch 10/25
9/9 [==============================] - 3s 331ms/step - loss: 2.1219 - accuracy: 0.4489
Epoch 11/25
9/9 [==============================] - 3s 334ms/step - loss: 1.9020 - accuracy: 0.5028
Epoch 12/25
9/9 [==============================] - 3s 334ms/step - loss: 1.6502 - accuracy: 0.5581
Epoch 13/25
9/9 [==============================] - 3s 334ms/step - loss: 1.5115 - accuracy: 0.5869
Epoch 14/25
9/9 [==============================] - 3s 334ms/step - loss: 1.3274 - accuracy: 0.6336
Epoch 15/25
9/9 [==============================] - 3s 334ms/step - loss: 1.1831 - accuracy: 0.6668
Epoch 16/25
9/9 [==============================] - 3s 334ms/step - loss: 0.9772 - accuracy: 0.7211
Epoch 17/25
9/9 [==============================] - 3s 335ms/step - loss: 1.0308 - accuracy: 0.7053
Epoch 18/25
9/9 [==============================] - 3s 337ms/step - loss: 0.8926 - accuracy: 0.7370
Epoch 19/25
9/9 [==============================] - 3s 335ms/step - loss: 0.6769 - accuracy: 0.8030
Epoch 20/25
9/9 [==============================] - 3s 337ms/step - loss: 1.0053 - accuracy: 0.7136
Epoch 21/25
9/9 [==============================] - 3s 336ms/step - loss: 0.8281 - accuracy: 0.7538
Epoch 22/25
9/9 [==============================] - 3s 335ms/step - loss: 0.5696 - accuracy: 0.8357
Epoch 23/25
9/9 [==============================] - 3s 336ms/step - loss: 0.4530 - accuracy: 0.8671
Epoch 24/25
9/9 [==============================] - 3s 335ms/step - loss: 0.3883 - accuracy: 0.8855
Epoch 25/25
9/9 [==============================] - 3s 334ms/step - loss: 0.3016 - accuracy: 0.9138
michael@14900c MINGW64 /c/wse_github/obrienlabsdev/machine-learning (main)
$ ./build.sh
#0 building with "desktop-linux" instance using docker driver
#1 [internal] load build definition from Dockerfile
#1 transferring dockerfile: 486B done
#1 DONE 0.0s
#2 [internal] load metadata for docker.io/tensorflow/tensorflow:2.14.0-gpu
#2 DONE 0.3s
#3 [internal] load .dockerignore
#3 transferring context: 2B done
#3 DONE 0.0s
#4 [1/3] FROM docker.io/tensorflow/tensorflow:2.14.0-gpu@sha256:64602abcd8cc4f4bdd6268ca0abc39e6d37113d700886afd15f6dd151210b206
#4 DONE 0.0s
#5 [internal] load build context
#5 transferring context: 4.04kB 0.0s done
#5 DONE 0.0s
#6 [2/3] WORKDIR /src
#6 CACHED
#7 [3/3] COPY /src/tflow.py .
#7 DONE 0.0s
#8 exporting to image
#8 exporting layers
#8 exporting layers 0.0s done
#8 writing image sha256:922ed6d49d300e184ec9b3d4819e9b759f6b3af17d36ed47735cd73093b85738 done
#8 naming to docker.io/library/ml-tensorflow-win done
#8 DONE 0.1s
2024-12-01 02:25:12.541506: I tensorflow/core/util/port.cc:111] oneDNN custom operations are on. You may see slightly different numerical results due to floating-point round-off errors from different computation orders. To turn them off, set the environment variable `TF_ENABLE_ONEDNN_OPTS=0`.
2024-12-01 02:25:12.559026: E tensorflow/compiler/xla/stream_executor/cuda/cuda_dnn.cc:9342] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered
2024-12-01 02:25:12.559053: E tensorflow/compiler/xla/stream_executor/cuda/cuda_fft.cc:609] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered
2024-12-01 02:25:12.559063: E tensorflow/compiler/xla/stream_executor/cuda/cuda_blas.cc:1518] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered
2024-12-01 02:25:12.562916: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 AVX_VNNI FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2024-12-01 02:25:13.305767: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 02:25:13.308785: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 02:25:13.308836: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 02:25:13.310159: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 02:25:13.310242: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 02:25:13.310393: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 02:25:13.400624: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 02:25:13.400661: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 02:25:13.400666: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1977] Could not identify NUMA node of platform GPU id 0, defaulting to 0. Your kernel may not have been built with NUMA support.
2024-12-01 02:25:13.400682: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 02:25:13.400695: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1886] Created device /job:localhost/replica:0/task:0/device:GPU:0 with 46009 MB memory: -> device: 0, name: NVIDIA RTX A6000, pci bus id: 0000:01:00.0, compute capability: 8.6
Downloading data from https://www.cs.toronto.edu/~kriz/cifar-100-python.tar.gz
169001437/169001437 [==============================] - 3s 0us/step
Epoch 1/25
2024-12-01 02:25:24.924752: I tensorflow/compiler/xla/stream_executor/cuda/cuda_dnn.cc:442] Loaded cuDNN version 8600
2024-12-01 02:25:27.706536: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x7f7e98df02e0 initialized for platform CUDA (this does not guarantee that XLA will be used). Devices:
2024-12-01 02:25:27.706561: I tensorflow/compiler/xla/service/service.cc:176] StreamExecutor device (0): NVIDIA RTX A6000, Compute Capability 8.6
2024-12-01 02:25:27.709657: I tensorflow/compiler/mlir/tensorflow/utils/dump_mlir_util.cc:269] disabling MLIR crash reproducer, set env var `MLIR_CRASH_REPRODUCER_DIRECTORY` to enable.
2024-12-01 02:25:27.757555: I ./tensorflow/compiler/jit/device_compiler.h:186] Compiled cluster using XLA! This line is logged at most once for the lifetime of the process.
7/7 [==============================] - 34s 910ms/step - loss: 6.1072 - accuracy: 0.0174
Epoch 2/25
7/7 [==============================] - 3s 420ms/step - loss: 4.5421 - accuracy: 0.0470
Epoch 3/25
7/7 [==============================] - 3s 421ms/step - loss: 4.1170 - accuracy: 0.0783
Epoch 4/25
7/7 [==============================] - 3s 421ms/step - loss: 3.8726 - accuracy: 0.1141
Epoch 5/25
7/7 [==============================] - 3s 422ms/step - loss: 3.6838 - accuracy: 0.1509
Epoch 6/25
7/7 [==============================] - 3s 421ms/step - loss: 3.4169 - accuracy: 0.1907
Epoch 7/25
7/7 [==============================] - 3s 423ms/step - loss: 3.2222 - accuracy: 0.2297
Epoch 8/25
7/7 [==============================] - 3s 422ms/step - loss: 3.6691 - accuracy: 0.1675
Epoch 9/25
7/7 [==============================] - 3s 424ms/step - loss: 3.7561 - accuracy: 0.1492
Epoch 10/25
7/7 [==============================] - 3s 425ms/step - loss: 3.5491 - accuracy: 0.1760
Epoch 11/25
7/7 [==============================] - 3s 426ms/step - loss: 3.3228 - accuracy: 0.2094
Epoch 12/25
7/7 [==============================] - 3s 425ms/step - loss: 3.1399 - accuracy: 0.2394
Epoch 13/25
7/7 [==============================] - 3s 427ms/step - loss: 2.9667 - accuracy: 0.2725
Epoch 14/25
7/7 [==============================] - 3s 426ms/step - loss: 2.8074 - accuracy: 0.3065
Epoch 15/25
7/7 [==============================] - 3s 426ms/step - loss: 2.6076 - accuracy: 0.3464
Epoch 16/25
7/7 [==============================] - 3s 428ms/step - loss: 2.7522 - accuracy: 0.3272
Epoch 17/25
7/7 [==============================] - 3s 426ms/step - loss: 2.5534 - accuracy: 0.3621
Epoch 18/25
7/7 [==============================] - 3s 427ms/step - loss: 2.3425 - accuracy: 0.4111
Epoch 19/25
7/7 [==============================] - 3s 427ms/step - loss: 2.0997 - accuracy: 0.4621
Epoch 20/25
7/7 [==============================] - 3s 426ms/step - loss: 1.9775 - accuracy: 0.4959
Epoch 21/25
7/7 [==============================] - 3s 428ms/step - loss: 1.7427 - accuracy: 0.5398
Epoch 22/25
7/7 [==============================] - 3s 428ms/step - loss: 1.4871 - accuracy: 0.6050
Epoch 23/25
7/7 [==============================] - 3s 426ms/step - loss: 1.2739 - accuracy: 0.6586
Epoch 24/25
7/7 [==============================] - 3s 426ms/step - loss: 1.0641 - accuracy: 0.7079
Epoch 25/25
7/7 [==============================] - 3s 426ms/step - loss: 0.8945 - accuracy: 0.7526
michael@14900c MINGW64 /c/wse_github/obrienlabsdev/machine-learning (main)
$ ./build.sh
#0 building with "desktop-linux" instance using docker driver
#1 [internal] load build definition from Dockerfile
#1 transferring dockerfile: 486B done
#1 DONE 0.0s
#2 [auth] tensorflow/tensorflow:pull token for registry-1.docker.io
#2 DONE 0.0s
#3 [internal] load metadata for docker.io/tensorflow/tensorflow:2.14.0-gpu
#3 DONE 0.5s
#4 [internal] load .dockerignore
#4 transferring context: 2B done
#4 DONE 0.0s
#5 [1/3] FROM docker.io/tensorflow/tensorflow:2.14.0-gpu@sha256:64602abcd8cc4f4bdd6268ca0abc39e6d37113d700886afd15f6dd151210b206
#5 DONE 0.0s
#6 [internal] load build context
#6 transferring context: 4.04kB 0.0s done
#6 DONE 0.0s
#7 [2/3] WORKDIR /src
#7 CACHED
#8 [3/3] COPY /src/tflow.py .
#8 DONE 0.0s
#9 exporting to image
#9 exporting layers 0.0s done
#9 writing image sha256:6bfe99d6a6992e6d844ab0646e564767d4d37426b0c1550fe97f5944ccc559c7 done
#9 naming to docker.io/library/ml-tensorflow-win done
#9 DONE 0.1s
2024-12-01 02:27:48.937126: I tensorflow/core/util/port.cc:111] oneDNN custom operations are on. You may see slightly different numerical results due to floating-point round-off errors from different computation orders. To turn them off, set the environment variable `TF_ENABLE_ONEDNN_OPTS=0`.
2024-12-01 02:27:48.954591: E tensorflow/compiler/xla/stream_executor/cuda/cuda_dnn.cc:9342] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered
2024-12-01 02:27:48.954624: E tensorflow/compiler/xla/stream_executor/cuda/cuda_fft.cc:609] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered
2024-12-01 02:27:48.954664: E tensorflow/compiler/xla/stream_executor/cuda/cuda_blas.cc:1518] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered
2024-12-01 02:27:48.958926: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 AVX_VNNI FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2024-12-01 02:27:49.701098: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 02:27:49.703305: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 02:27:49.703345: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 02:27:49.704237: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 02:27:49.704267: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 02:27:49.704276: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 02:27:49.801786: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 02:27:49.801822: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 02:27:49.801828: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1977] Could not identify NUMA node of platform GPU id 0, defaulting to 0. Your kernel may not have been built with NUMA support.
2024-12-01 02:27:49.801844: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 02:27:49.801867: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1886] Created device /job:localhost/replica:0/task:0/device:GPU:0 with 46009 MB memory: -> device: 0, name: NVIDIA RTX A6000, pci bus id: 0000:01:00.0, compute capability: 8.6
Downloading data from https://www.cs.toronto.edu/~kriz/cifar-100-python.tar.gz
169001437/169001437 [==============================] - 4s 0us/step
Epoch 1/25
2024-12-01 02:28:02.185247: I tensorflow/compiler/xla/stream_executor/cuda/cuda_dnn.cc:442] Loaded cuDNN version 8600
2024-12-01 02:28:04.332728: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x7fb8e8661c10 initialized for platform CUDA (this does not guarantee that XLA will be used). Devices:
2024-12-01 02:28:04.332749: I tensorflow/compiler/xla/service/service.cc:176] StreamExecutor device (0): NVIDIA RTX A6000, Compute Capability 8.6
2024-12-01 02:28:04.336152: I tensorflow/compiler/mlir/tensorflow/utils/dump_mlir_util.cc:269] disabling MLIR crash reproducer, set env var `MLIR_CRASH_REPRODUCER_DIRECTORY` to enable.
2024-12-01 02:28:04.384715: I ./tensorflow/compiler/jit/device_compiler.h:186] Compiled cluster using XLA! This line is logged at most once for the lifetime of the process.
10/10 [==============================] - 34s 1s/step - loss: 5.5942 - accuracy: 0.0251
Epoch 2/25
10/10 [==============================] - 3s 293ms/step - loss: 4.2972 - accuracy: 0.0699
Epoch 3/25
10/10 [==============================] - 3s 295ms/step - loss: 3.8669 - accuracy: 0.1154
Epoch 4/25
10/10 [==============================] - 3s 295ms/step - loss: 3.5305 - accuracy: 0.1677
Epoch 5/25
10/10 [==============================] - 3s 295ms/step - loss: 3.1965 - accuracy: 0.2263
Epoch 6/25
10/10 [==============================] - 3s 296ms/step - loss: 2.8172 - accuracy: 0.3008
Epoch 7/25
10/10 [==============================] - 3s 296ms/step - loss: 2.3786 - accuracy: 0.3944
Epoch 8/25
10/10 [==============================] - 3s 295ms/step - loss: 1.9369 - accuracy: 0.4957
Epoch 9/25
10/10 [==============================] - 3s 296ms/step - loss: 1.5214 - accuracy: 0.5956
Epoch 10/25
10/10 [==============================] - 3s 297ms/step - loss: 1.1885 - accuracy: 0.6777
Epoch 11/25
10/10 [==============================] - 3s 298ms/step - loss: 0.8498 - accuracy: 0.7654
Epoch 12/25
10/10 [==============================] - 3s 297ms/step - loss: 0.6862 - accuracy: 0.8079
Epoch 13/25
10/10 [==============================] - 3s 298ms/step - loss: 0.4986 - accuracy: 0.8589
Epoch 14/25
10/10 [==============================] - 3s 300ms/step - loss: 0.3895 - accuracy: 0.8877
Epoch 15/25
10/10 [==============================] - 3s 298ms/step - loss: 0.3158 - accuracy: 0.9085
Epoch 16/25
10/10 [==============================] - 3s 301ms/step - loss: 0.2679 - accuracy: 0.9245
Epoch 17/25
10/10 [==============================] - 3s 300ms/step - loss: 0.2197 - accuracy: 0.9366
Epoch 18/25
10/10 [==============================] - 3s 299ms/step - loss: 0.1929 - accuracy: 0.9446
Epoch 19/25
10/10 [==============================] - 3s 300ms/step - loss: 0.1641 - accuracy: 0.9532
Epoch 20/25
10/10 [==============================] - 3s 300ms/step - loss: 0.1469 - accuracy: 0.9581
Epoch 21/25
10/10 [==============================] - 3s 300ms/step - loss: 0.1393 - accuracy: 0.9599
Epoch 22/25
10/10 [==============================] - 3s 301ms/step - loss: 0.1278 - accuracy: 0.9631
Epoch 23/25
10/10 [==============================] - 3s 302ms/step - loss: 0.1229 - accuracy: 0.9633
Epoch 24/25
10/10 [==============================] - 3s 300ms/step - loss: 0.1223 - accuracy: 0.9648
Epoch 25/25
10/10 [==============================] - 3s 300ms/step - loss: 0.1282 - accuracy: 0.9627
michael@14900c MINGW64 /c/wse_github/obrienlabs/benchmark (master)
$ nvidia-smi
Sat Nov 30 21:32:22 2024
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 551.86 Driver Version: 551.86 CUDA Version: 12.4 |
|-----------------------------------------+------------------------+----------------------+
| GPU Name TCC/WDDM | Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. |
| | | MIG M. |
|=========================================+========================+======================|
| 0 NVIDIA RTX A6000 WDDM | 00000000:01:00.0 Off | Off |
| 65% 87C P2 284W / 300W | 46306MiB / 49140MiB | 97% Default |
| | | N/A |
+-----------------------------------------+------------------------+----------------------+
+-----------------------------------------------------------------------------------------+
| Processes: |
| GPU GI CI PID Type Process name GPU Memory |
| ID ID Usage |
|=========================================================================================|
+-----------------------------------------------------------------------------------------+
see #33
FROM tensorflow/tensorflow:2.14.0-gpu
512,256,512,1024,2048,4096,6144,8192,5120