Open obriensystems opened 3 hours ago
256, 512, 1024, 2048, 4096 batch size
micha@p1gen6 MINGW64 /c/wse_github/ObrienlabsDev/machine-learning (main)
$ ./build.sh
#0 building with "desktop-linux" instance using docker driver
#1 [internal] load build definition from Dockerfile
#1 transferring dockerfile: 575B done
#1 DONE 0.0s
#2 [internal] load metadata for docker.io/tensorflow/tensorflow:2.14.0-gpu
#2 DONE 0.5s
#3 [internal] load .dockerignore
#3 transferring context: 2B done
#3 DONE 0.0s
#4 [1/3] FROM docker.io/tensorflow/tensorflow:2.14.0-gpu@sha256:64602abcd8cc4f4bdd6268ca0abc39e6d37113d700886afd15f6dd151210b206
#4 DONE 0.0s
#5 [internal] load build context
#5 transferring context: 3.67kB 0.0s done
#5 DONE 0.0s
#6 [2/3] WORKDIR /src
#6 CACHED
#7 [3/3] COPY /src/tflow.py .
#7 DONE 0.0s
#8 exporting to image
#8 exporting layers 0.0s done
#8 writing image sha256:d820728b9b315c1efdca05a5df834ebee7a4367b7752f51e93e3fb581bccc798 done
#8 naming to docker.io/library/ml-tensorflow-win done
#8 DONE 0.0s
2024-11-30 23:57:26.685484: I tensorflow/core/util/port.cc:111] oneDNN custom operations are on. You may see slightly different numerical results due to floating-point round-off errors from different computation orders. To turn them off, set the environment variable `TF_ENABLE_ONEDNN_OPTS=0`.
2024-11-30 23:57:26.706273: E tensorflow/compiler/xla/stream_executor/cuda/cuda_dnn.cc:9342] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered
2024-11-30 23:57:26.706309: E tensorflow/compiler/xla/stream_executor/cuda/cuda_fft.cc:609] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered
2024-11-30 23:57:26.706321: E tensorflow/compiler/xla/stream_executor/cuda/cuda_blas.cc:1518] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered
2024-11-30 23:57:26.710656: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 AVX_VNNI FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2024-11-30 23:57:27.906699: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-11-30 23:57:27.909251: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-11-30 23:57:27.909301: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-11-30 23:57:27.910960: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-11-30 23:57:27.910997: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-11-30 23:57:27.911009: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-11-30 23:57:28.039226: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-11-30 23:57:28.039275: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-11-30 23:57:28.039281: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1977] Could not identify NUMA node of platform GPU id 0, defaulting to 0. Your kernel may not have been built with NUMA support.
2024-11-30 23:57:28.039301: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-11-30 23:57:28.039338: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1886] Created device /job:localhost/replica:0/task:0/device:GPU:0 with 9581 MB memory: -> device: 0, name: NVIDIA RTX 3500 Ada Generation Laptop GPU, pci bus id: 0000:01:00.0, compute capability: 8.9
Downloading data from https://www.cs.toronto.edu/~kriz/cifar-100-python.tar.gz
169001437/169001437 [==============================] - 9s 0us/step
Epoch 1/25
2024-11-30 23:57:48.472889: I tensorflow/compiler/xla/stream_executor/cuda/cuda_dnn.cc:442] Loaded cuDNN version 8600
2024-11-30 23:57:51.465069: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x7f54603fc950 initialized for platform CUDA (this does not guarantee that XLA will be used). Devices:
2024-11-30 23:57:51.465098: I tensorflow/compiler/xla/service/service.cc:176] StreamExecutor device (0): NVIDIA RTX 3500 Ada Generation Laptop GPU, Compute Capability 8.9
2024-11-30 23:57:51.468427: I tensorflow/compiler/mlir/tensorflow/utils/dump_mlir_util.cc:269] disabling MLIR crash reproducer, set env var `MLIR_CRASH_REPRODUCER_DIRECTORY` to enable.
2024-11-30 23:57:51.525029: I ./tensorflow/compiler/jit/device_compiler.h:186] Compiled cluster using XLA! This line is logged at most once for the lifetime of the process.
196/196 [==============================] - 34s 57ms/step - loss: 4.3636 - accuracy: 0.0868
Epoch 2/25
196/196 [==============================] - 8s 40ms/step - loss: 3.9713 - accuracy: 0.1480
Epoch 3/25
196/196 [==============================] - 8s 40ms/step - loss: 3.7218 - accuracy: 0.1683
Epoch 4/25
196/196 [==============================] - 8s 41ms/step - loss: 3.2891 - accuracy: 0.2331
Epoch 5/25
196/196 [==============================] - 8s 41ms/step - loss: 3.6898 - accuracy: 0.1783
Epoch 6/25
196/196 [==============================] - 8s 42ms/step - loss: 3.3448 - accuracy: 0.2245
Epoch 7/25
196/196 [==============================] - 8s 42ms/step - loss: 3.1134 - accuracy: 0.2653
Epoch 8/25
196/196 [==============================] - 8s 43ms/step - loss: 3.0021 - accuracy: 0.2784
Epoch 9/25
196/196 [==============================] - 8s 43ms/step - loss: 2.7545 - accuracy: 0.3221
Epoch 10/25
196/196 [==============================] - 9s 44ms/step - loss: 2.7476 - accuracy: 0.3218
Epoch 11/25
196/196 [==============================] - 8s 43ms/step - loss: 2.6039 - accuracy: 0.3515
Epoch 12/25
196/196 [==============================] - 9s 44ms/step - loss: 2.3861 - accuracy: 0.3956
Epoch 13/25
196/196 [==============================] - 9s 44ms/step - loss: 2.3027 - accuracy: 0.4242
Epoch 14/25
196/196 [==============================] - 9s 44ms/step - loss: 2.1197 - accuracy: 0.4542
Epoch 15/25
196/196 [==============================] - 9s 44ms/step - loss: 1.8711 - accuracy: 0.5049
Epoch 16/25
196/196 [==============================] - 9s 45ms/step - loss: 1.6557 - accuracy: 0.5570
Epoch 17/25
196/196 [==============================] - 9s 44ms/step - loss: 1.4631 - accuracy: 0.6041
Epoch 18/25
196/196 [==============================] - 9s 45ms/step - loss: 1.3836 - accuracy: 0.6176
Epoch 19/25
196/196 [==============================] - 9s 45ms/step - loss: 1.9901 - accuracy: 0.5074
Epoch 20/25
196/196 [==============================] - 9s 44ms/step - loss: 1.7544 - accuracy: 0.5436
Epoch 21/25
196/196 [==============================] - 9s 45ms/step - loss: 1.8138 - accuracy: 0.5295
Epoch 22/25
196/196 [==============================] - 9s 44ms/step - loss: 1.1861 - accuracy: 0.6827
Epoch 23/25
196/196 [==============================] - 9s 45ms/step - loss: 0.8640 - accuracy: 0.7623
Epoch 24/25
196/196 [==============================] - 9s 43ms/step - loss: 0.6471 - accuracy: 0.8293
Epoch 25/25
196/196 [==============================] - 8s 43ms/step - loss: 0.5260 - accuracy: 0.8597
(venv-cuda)
micha@p1gen6 MINGW64 /c/wse_github/ObrienlabsDev/machine-learning (main)
$ ./build.sh
#0 building with "desktop-linux" instance using docker driver
#1 [internal] load build definition from Dockerfile
#1 transferring dockerfile: 575B done
#1 DONE 0.0s
#2 [auth] tensorflow/tensorflow:pull token for registry-1.docker.io
#2 DONE 0.0s
#3 [internal] load metadata for docker.io/tensorflow/tensorflow:2.14.0-gpu
#3 DONE 0.6s
#4 [internal] load .dockerignore
#4 transferring context: 2B done
#4 DONE 0.0s
#5 [1/3] FROM docker.io/tensorflow/tensorflow:2.14.0-gpu@sha256:64602abcd8cc4f4bdd6268ca0abc39e6d37113d700886afd15f6dd151210b206
#5 DONE 0.0s
#6 [internal] load build context
#6 transferring context: 3.67kB done
#6 DONE 0.0s
#7 [2/3] WORKDIR /src
#7 CACHED
#8 [3/3] COPY /src/tflow.py .
#8 CACHED
#9 exporting to image
#9 exporting layers done
#9 writing image sha256:1d6810df1690137c7ea77998e44557269fbcdfbfcf20ba14748865748bdbeb1a done
#9 naming to docker.io/library/ml-tensorflow-win done
#9 DONE 0.0s
2024-12-01 00:04:25.038485: I tensorflow/core/util/port.cc:111] oneDNN custom operations are on. You may see slightly different numerical results due to floating-point round-off errors from different computation orders. To turn them off, set the environment variable `TF_ENABLE_ONEDNN_OPTS=0`.
2024-12-01 00:04:25.058796: E tensorflow/compiler/xla/stream_executor/cuda/cuda_dnn.cc:9342] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered
2024-12-01 00:04:25.058830: E tensorflow/compiler/xla/stream_executor/cuda/cuda_fft.cc:609] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered
2024-12-01 00:04:25.058841: E tensorflow/compiler/xla/stream_executor/cuda/cuda_blas.cc:1518] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered
2024-12-01 00:04:25.063577: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 AVX_VNNI FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2024-12-01 00:04:26.074299: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:04:26.076994: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:04:26.077051: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:04:26.079172: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:04:26.079201: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:04:26.079211: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:04:26.226348: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:04:26.226451: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:04:26.226492: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1977] Could not identify NUMA node of platform GPU id 0, defaulting to 0. Your kernel may not have been built with NUMA support.
2024-12-01 00:04:26.226564: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:04:26.226595: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1886] Created device /job:localhost/replica:0/task:0/device:GPU:0 with 9581 MB memory: -> device: 0, name: NVIDIA RTX 3500 Ada Generation Laptop GPU, pci bus id: 0000:01:00.0, compute capability: 8.9
Downloading data from https://www.cs.toronto.edu/~kriz/cifar-100-python.tar.gz
169001437/169001437 [==============================] - 6s 0us/step
Epoch 1/25
2024-12-01 00:04:43.941642: I tensorflow/compiler/xla/stream_executor/cuda/cuda_dnn.cc:442] Loaded cuDNN version 8600
2024-12-01 00:04:46.425421: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x7f61bc946980 initialized for platform CUDA (this does not guarantee that XLA will be used). Devices:
2024-12-01 00:04:46.425451: I tensorflow/compiler/xla/service/service.cc:176] StreamExecutor device (0): NVIDIA RTX 3500 Ada Generation Laptop GPU, Compute Capability 8.9
2024-12-01 00:04:46.428552: I tensorflow/compiler/mlir/tensorflow/utils/dump_mlir_util.cc:269] disabling MLIR crash reproducer, set env var `MLIR_CRASH_REPRODUCER_DIRECTORY` to enable.
2024-12-01 00:04:46.484245: I ./tensorflow/compiler/jit/device_compiler.h:186] Compiled cluster using XLA! This line is logged at most once for the lifetime of the process.
98/98 [==============================] - 30s 83ms/step - loss: 4.4019 - accuracy: 0.0753
Epoch 2/25
98/98 [==============================] - 5s 54ms/step - loss: 3.5497 - accuracy: 0.1767
Epoch 3/25
98/98 [==============================] - 5s 54ms/step - loss: 3.1803 - accuracy: 0.2434
Epoch 4/25
98/98 [==============================] - 5s 54ms/step - loss: 3.3044 - accuracy: 0.2308
Epoch 5/25
98/98 [==============================] - 5s 54ms/step - loss: 3.0593 - accuracy: 0.2603
Epoch 6/25
98/98 [==============================] - 5s 54ms/step - loss: 2.6995 - accuracy: 0.3353
Epoch 7/25
98/98 [==============================] - 5s 54ms/step - loss: 2.3841 - accuracy: 0.3945
Epoch 8/25
98/98 [==============================] - 5s 55ms/step - loss: 2.3127 - accuracy: 0.4143
Epoch 9/25
98/98 [==============================] - 5s 54ms/step - loss: 2.0369 - accuracy: 0.4691
Epoch 10/25
98/98 [==============================] - 5s 54ms/step - loss: 1.7624 - accuracy: 0.5369
Epoch 11/25
98/98 [==============================] - 5s 55ms/step - loss: 1.6354 - accuracy: 0.5610
Epoch 12/25
98/98 [==============================] - 5s 55ms/step - loss: 1.3999 - accuracy: 0.6258
Epoch 13/25
98/98 [==============================] - 5s 56ms/step - loss: 3.3840 - accuracy: 0.2818
Epoch 14/25
98/98 [==============================] - 5s 55ms/step - loss: 3.4657 - accuracy: 0.2331
Epoch 15/25
98/98 [==============================] - 5s 56ms/step - loss: 3.0036 - accuracy: 0.3019
Epoch 16/25
98/98 [==============================] - 5s 56ms/step - loss: 2.4746 - accuracy: 0.3717
Epoch 17/25
98/98 [==============================] - 5s 56ms/step - loss: 2.1085 - accuracy: 0.4415
Epoch 18/25
98/98 [==============================] - 5s 56ms/step - loss: 1.8274 - accuracy: 0.5083
Epoch 19/25
98/98 [==============================] - 6s 56ms/step - loss: 1.5464 - accuracy: 0.5782
Epoch 20/25
98/98 [==============================] - 6s 57ms/step - loss: 1.2816 - accuracy: 0.6471
Epoch 21/25
98/98 [==============================] - 6s 57ms/step - loss: 1.0437 - accuracy: 0.7191
Epoch 22/25
98/98 [==============================] - 6s 57ms/step - loss: 0.7747 - accuracy: 0.7823
Epoch 23/25
98/98 [==============================] - 6s 58ms/step - loss: 0.8367 - accuracy: 0.7624
Epoch 24/25
98/98 [==============================] - 6s 58ms/step - loss: 0.6280 - accuracy: 0.8256
Epoch 25/25
98/98 [==============================] - 6s 58ms/step - loss: 0.4482 - accuracy: 0.8822
(venv-cuda)
micha@p1gen6 MINGW64 /c/wse_github/ObrienlabsDev/machine-learning (main)
$ java --version
java 21.0.5 2024-10-15 LTS
Java(TM) SE Runtime Environment (build 21.0.5+9-LTS-239)
Java HotSpot(TM) 64-Bit Server VM (build 21.0.5+9-LTS-239, mixed mode, sharing)
(venv-cuda)
micha@p1gen6 MINGW64 /c/wse_github/ObrienlabsDev/machine-learning (main)
$ ./build.sh
ERROR: listing workers for Build: failed to list workers: Canceled: context canceled
(venv-cuda)
micha@p1gen6 MINGW64 /c/wse_github/ObrienlabsDev/machine-learning (main)
$ ./build.sh
#0 building with "desktop-linux" instance using docker driver
#1 [internal] load build definition from Dockerfile
#1 transferring dockerfile: 575B done
#1 DONE 0.0s
#2 [internal] load metadata for docker.io/tensorflow/tensorflow:2.14.0-gpu
#2 ...
#3 [auth] tensorflow/tensorflow:pull token for registry-1.docker.io
#3 DONE 0.0s
#2 [internal] load metadata for docker.io/tensorflow/tensorflow:2.14.0-gpu
#2 DONE 0.8s
#4 [internal] load .dockerignore
#4 transferring context: 2B done
#4 DONE 0.0s
#5 [1/3] FROM docker.io/tensorflow/tensorflow:2.14.0-gpu@sha256:64602abcd8cc4f4bdd6268ca0abc39e6d37113d700886afd15f6dd151210b206
#5 DONE 0.0s
#6 [internal] load build context
#6 transferring context: 3.68kB 0.0s done
#6 DONE 0.0s
#7 [2/3] WORKDIR /src
#7 CACHED
#8 [3/3] COPY /src/tflow.py .
#8 DONE 0.0s
#9 exporting to image
#9 exporting layers 0.0s done
#9 writing image sha256:bbcea2a57cf72678d126f28fe2eb940774d4bd9acf9b5ecba9f66bbb589e6153 done
#9 naming to docker.io/library/ml-tensorflow-win done
#9 DONE 0.0s
2024-12-01 00:25:31.999107: I tensorflow/core/util/port.cc:111] oneDNN custom operations are on. You may see slightly different numerical results due to floating-point round-off errors from different computation orders. To turn them off, set the environment variable `TF_ENABLE_ONEDNN_OPTS=0`.
2024-12-01 00:25:32.229234: E tensorflow/compiler/xla/stream_executor/cuda/cuda_dnn.cc:9342] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered
2024-12-01 00:25:32.229266: E tensorflow/compiler/xla/stream_executor/cuda/cuda_fft.cc:609] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered
2024-12-01 00:25:32.230800: E tensorflow/compiler/xla/stream_executor/cuda/cuda_blas.cc:1518] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered
2024-12-01 00:25:32.340545: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 AVX_VNNI FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2024-12-01 00:25:34.385224: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:25:34.426195: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:25:34.426241: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:25:34.429134: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:25:34.429191: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:25:34.429204: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:25:34.595716: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:25:34.595763: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:25:34.595767: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1977] Could not identify NUMA node of platform GPU id 0, defaulting to 0. Your kernel may not have been built with NUMA support.
2024-12-01 00:25:34.595799: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:25:34.595810: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1886] Created device /job:localhost/replica:0/task:0/device:GPU:0 with 9581 MB memory: -> device: 0, name: NVIDIA RTX 3500 Ada Generation Laptop GPU, pci bus id: 0000:01:00.0, compute capability: 8.9
Downloading data from https://www.cs.toronto.edu/~kriz/cifar-100-python.tar.gz
169001437/169001437 [==============================] - 18s 0us/step
Epoch 1/25
2024-12-01 00:26:04.454519: I tensorflow/compiler/xla/stream_executor/cuda/cuda_dnn.cc:442] Loaded cuDNN version 8600
2024-12-01 00:26:06.652346: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x7f9a1359f940 initialized for platform CUDA (this does not guarantee that XLA will be used). Devices:
2024-12-01 00:26:06.652378: I tensorflow/compiler/xla/service/service.cc:176] StreamExecutor device (0): NVIDIA RTX 3500 Ada Generation Laptop GPU, Compute Capability 8.9
2024-12-01 00:26:06.659005: I tensorflow/compiler/mlir/tensorflow/utils/dump_mlir_util.cc:269] disabling MLIR crash reproducer, set env var `MLIR_CRASH_REPRODUCER_DIRECTORY` to enable.
2024-12-01 00:26:06.733169: I ./tensorflow/compiler/jit/device_compiler.h:186] Compiled cluster using XLA! This line is logged at most once for the lifetime of the process.
49/49 [==============================] - 32s 201ms/step - loss: 4.6684 - accuracy: 0.0623
Epoch 2/25
49/49 [==============================] - 5s 103ms/step - loss: 3.6396 - accuracy: 0.1590
Epoch 3/25
49/49 [==============================] - 5s 105ms/step - loss: 3.1207 - accuracy: 0.2439
Epoch 4/25
49/49 [==============================] - 5s 105ms/step - loss: 2.8306 - accuracy: 0.3066
Epoch 5/25
49/49 [==============================] - 5s 106ms/step - loss: 2.7187 - accuracy: 0.3275
Epoch 6/25
49/49 [==============================] - 5s 106ms/step - loss: 2.6684 - accuracy: 0.3385
Epoch 7/25
49/49 [==============================] - 5s 106ms/step - loss: 2.4572 - accuracy: 0.3773
Epoch 8/25
49/49 [==============================] - 5s 106ms/step - loss: 2.9407 - accuracy: 0.2756
Epoch 9/25
49/49 [==============================] - 5s 105ms/step - loss: 2.4240 - accuracy: 0.3785
Epoch 10/25
49/49 [==============================] - 5s 106ms/step - loss: 2.0339 - accuracy: 0.4637
Epoch 11/25
49/49 [==============================] - 5s 106ms/step - loss: 1.6721 - accuracy: 0.5452
Epoch 12/25
49/49 [==============================] - 5s 105ms/step - loss: 1.3453 - accuracy: 0.6221
Epoch 13/25
49/49 [==============================] - 5s 106ms/step - loss: 1.1025 - accuracy: 0.6861
Epoch 14/25
49/49 [==============================] - 5s 106ms/step - loss: 0.8672 - accuracy: 0.7436
Epoch 15/25
49/49 [==============================] - 5s 106ms/step - loss: 0.6552 - accuracy: 0.8053
Epoch 16/25
49/49 [==============================] - 5s 105ms/step - loss: 0.5098 - accuracy: 0.8456
Epoch 17/25
49/49 [==============================] - 5s 105ms/step - loss: 0.8505 - accuracy: 0.7490
Epoch 18/25
49/49 [==============================] - 5s 106ms/step - loss: 1.0267 - accuracy: 0.7125
Epoch 19/25
49/49 [==============================] - 5s 105ms/step - loss: 0.9959 - accuracy: 0.7142
Epoch 20/25
49/49 [==============================] - 5s 106ms/step - loss: 0.4886 - accuracy: 0.8532
Epoch 21/25
49/49 [==============================] - 5s 105ms/step - loss: 0.2797 - accuracy: 0.9231
Epoch 22/25
49/49 [==============================] - 5s 105ms/step - loss: 0.1975 - accuracy: 0.9440
Epoch 23/25
49/49 [==============================] - 5s 105ms/step - loss: 0.1371 - accuracy: 0.9600
Epoch 24/25
49/49 [==============================] - 5s 107ms/step - loss: 0.0981 - accuracy: 0.9735
Epoch 25/25
49/49 [==============================] - 5s 106ms/step - loss: 0.0758 - accuracy: 0.9776
(venv-cuda)
micha@p1gen6 MINGW64 /c/wse_github/ObrienlabsDev/machine-learning (main)
$ ./build.sh
#0 building with "desktop-linux" instance using docker driver
#1 [internal] load build definition from Dockerfile
#1 transferring dockerfile: 575B done
#1 DONE 0.0s
#2 [auth] tensorflow/tensorflow:pull token for registry-1.docker.io
#2 DONE 0.0s
#3 [internal] load metadata for docker.io/tensorflow/tensorflow:2.14.0-gpu
#3 DONE 1.0s
#4 [internal] load .dockerignore
#4 transferring context: 2B done
#4 DONE 0.0s
#5 [1/3] FROM docker.io/tensorflow/tensorflow:2.14.0-gpu@sha256:64602abcd8cc4f4bdd6268ca0abc39e6d37113d700886afd15f6dd151210b206
#5 DONE 0.0s
#6 [internal] load build context
#6 transferring context: 3.68kB 0.0s done
#6 DONE 0.0s
#7 [2/3] WORKDIR /src
#7 CACHED
#8 [3/3] COPY /src/tflow.py .
#8 DONE 0.0s
#9 exporting to image
#9 exporting layers 0.0s done
#9 writing image sha256:91662b98c48f03fb3acd6fbc253841ce7aac8e445d0cdb8380442f5e29a648f1 done
#9 naming to docker.io/library/ml-tensorflow-win done
#9 DONE 0.0s
2024-12-01 00:30:21.013693: I tensorflow/core/util/port.cc:111] oneDNN custom operations are on. You may see slightly different numerical results due to floating-point round-off errors from different computation orders. To turn them off, set the environment variable `TF_ENABLE_ONEDNN_OPTS=0`.
2024-12-01 00:30:21.034515: E tensorflow/compiler/xla/stream_executor/cuda/cuda_dnn.cc:9342] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered
2024-12-01 00:30:21.034574: E tensorflow/compiler/xla/stream_executor/cuda/cuda_fft.cc:609] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered
2024-12-01 00:30:21.034588: E tensorflow/compiler/xla/stream_executor/cuda/cuda_blas.cc:1518] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered
2024-12-01 00:30:21.039309: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 AVX_VNNI FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2024-12-01 00:30:22.066704: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:30:22.069331: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:30:22.069380: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:30:22.071225: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:30:22.071251: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:30:22.071259: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:30:22.238991: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:30:22.239037: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:30:22.239044: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1977] Could not identify NUMA node of platform GPU id 0, defaulting to 0. Your kernel may not have been built with NUMA support.
2024-12-01 00:30:22.239062: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:30:22.239076: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1886] Created device /job:localhost/replica:0/task:0/device:GPU:0 with 9581 MB memory: -> device: 0, name: NVIDIA RTX 3500 Ada Generation Laptop GPU, pci bus id: 0000:01:00.0, compute capability: 8.9
Downloading data from https://www.cs.toronto.edu/~kriz/cifar-100-python.tar.gz
169001437/169001437 [==============================] - 6s 0us/step
Epoch 1/25
2024-12-01 00:30:39.622591: I tensorflow/compiler/xla/stream_executor/cuda/cuda_dnn.cc:442] Loaded cuDNN version 8600
2024-12-01 00:30:41.973276: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x7f0dc02065d0 initialized for platform CUDA (this does not guarantee that XLA will be used). Devices:
2024-12-01 00:30:41.973300: I tensorflow/compiler/xla/service/service.cc:176] StreamExecutor device (0): NVIDIA RTX 3500 Ada Generation Laptop GPU, Compute Capability 8.9
2024-12-01 00:30:41.976530: I tensorflow/compiler/mlir/tensorflow/utils/dump_mlir_util.cc:269] disabling MLIR crash reproducer, set env var `MLIR_CRASH_REPRODUCER_DIRECTORY` to enable.
2024-12-01 00:30:42.035687: I ./tensorflow/compiler/jit/device_compiler.h:186] Compiled cluster using XLA! This line is logged at most once for the lifetime of the process.
25/25 [==============================] - 34s 400ms/step - loss: 4.9395 - accuracy: 0.0354
Epoch 2/25
25/25 [==============================] - 5s 207ms/step - loss: 3.9236 - accuracy: 0.1054
Epoch 3/25
25/25 [==============================] - 5s 207ms/step - loss: 3.4189 - accuracy: 0.1878
Epoch 4/25
25/25 [==============================] - 5s 209ms/step - loss: 3.0047 - accuracy: 0.2666
Epoch 5/25
25/25 [==============================] - 5s 208ms/step - loss: 2.6428 - accuracy: 0.3382
Epoch 6/25
25/25 [==============================] - 5s 208ms/step - loss: 2.3087 - accuracy: 0.4076
Epoch 7/25
25/25 [==============================] - 5s 208ms/step - loss: 2.1035 - accuracy: 0.4594
Epoch 8/25
25/25 [==============================] - 5s 209ms/step - loss: 1.8614 - accuracy: 0.5035
Epoch 9/25
25/25 [==============================] - 5s 208ms/step - loss: 1.5357 - accuracy: 0.5820
Epoch 10/25
25/25 [==============================] - 5s 209ms/step - loss: 1.4527 - accuracy: 0.6061
Epoch 11/25
25/25 [==============================] - 5s 209ms/step - loss: 1.0684 - accuracy: 0.6985
Epoch 12/25
25/25 [==============================] - 5s 209ms/step - loss: 0.7071 - accuracy: 0.7972
Epoch 13/25
25/25 [==============================] - 5s 209ms/step - loss: 0.5945 - accuracy: 0.8280
Epoch 14/25
25/25 [==============================] - 5s 208ms/step - loss: 0.5797 - accuracy: 0.8359
Epoch 15/25
25/25 [==============================] - 5s 208ms/step - loss: 0.5358 - accuracy: 0.8441
Epoch 16/25
25/25 [==============================] - 5s 208ms/step - loss: 0.4175 - accuracy: 0.8808
Epoch 17/25
25/25 [==============================] - 5s 209ms/step - loss: 0.3493 - accuracy: 0.8983
Epoch 18/25
25/25 [==============================] - 5s 209ms/step - loss: 0.4485 - accuracy: 0.8755
Epoch 19/25
25/25 [==============================] - 5s 209ms/step - loss: 0.7039 - accuracy: 0.8063
Epoch 20/25
25/25 [==============================] - 5s 209ms/step - loss: 1.0323 - accuracy: 0.7179
Epoch 21/25
25/25 [==============================] - 5s 209ms/step - loss: 0.5772 - accuracy: 0.8379
Epoch 22/25
25/25 [==============================] - 5s 209ms/step - loss: 0.2776 - accuracy: 0.9209
Epoch 23/25
25/25 [==============================] - 5s 208ms/step - loss: 0.2011 - accuracy: 0.9501
Epoch 24/25
25/25 [==============================] - 5s 209ms/step - loss: 0.2049 - accuracy: 0.9618
Epoch 25/25
25/25 [==============================] - 5s 209ms/step - loss: 0.1222 - accuracy: 0.9830
(venv-cuda)
micha@p1gen6 MINGW64 /c/wse_github/ObrienlabsDev/machine-learning (main)
$ ./build.sh
#0 building with "desktop-linux" instance using docker driver
#1 [internal] load build definition from Dockerfile
#1 transferring dockerfile: 575B done
#1 DONE 0.0s
#2 [auth] tensorflow/tensorflow:pull token for registry-1.docker.io
#2 DONE 0.0s
#3 [internal] load metadata for docker.io/tensorflow/tensorflow:2.14.0-gpu
#3 DONE 0.8s
#4 [internal] load .dockerignore
#4 transferring context: 2B done
#4 DONE 0.0s
#5 [1/3] FROM docker.io/tensorflow/tensorflow:2.14.0-gpu@sha256:64602abcd8cc4f4bdd6268ca0abc39e6d37113d700886afd15f6dd151210b206
#5 DONE 0.0s
#6 [internal] load build context
#6 transferring context: 3.67kB 0.0s done
#6 DONE 0.0s
#7 [2/3] WORKDIR /src
#7 CACHED
#8 [3/3] COPY /src/tflow.py .
#8 DONE 0.0s
#9 exporting to image
#9 exporting layers 0.0s done
#9 writing image sha256:a5ee876f43f024c7a064c48b3d61ae7d854230bbec86ef4729f987e0e96aba63 done
#9 naming to docker.io/library/ml-tensorflow-win done
#9 DONE 0.0s
2024-12-01 00:35:19.583397: I tensorflow/core/util/port.cc:111] oneDNN custom operations are on. You may see slightly different numerical results due to floating-point round-off errors from different computation orders. To turn them off, set the environment variable `TF_ENABLE_ONEDNN_OPTS=0`.
2024-12-01 00:35:19.606388: E tensorflow/compiler/xla/stream_executor/cuda/cuda_dnn.cc:9342] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered
2024-12-01 00:35:19.606418: E tensorflow/compiler/xla/stream_executor/cuda/cuda_fft.cc:609] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered
2024-12-01 00:35:19.606429: E tensorflow/compiler/xla/stream_executor/cuda/cuda_blas.cc:1518] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered
2024-12-01 00:35:19.610639: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 AVX_VNNI FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2024-12-01 00:35:21.166299: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:35:21.168924: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:35:21.168970: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:35:21.170887: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:35:21.170914: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:35:21.170923: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:35:21.315538: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:35:21.315611: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:35:21.315620: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1977] Could not identify NUMA node of platform GPU id 0, defaulting to 0. Your kernel may not have been built with NUMA support.
2024-12-01 00:35:21.315645: I tensorflow/compiler/xla/stream_executor/cuda/cuda_gpu_executor.cc:880] could not open file to read NUMA node: /sys/bus/pci/devices/0000:01:00.0/numa_node
Your kernel may have been built without NUMA support.
2024-12-01 00:35:21.315679: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1886] Created device /job:localhost/replica:0/task:0/device:GPU:0 with 9581 MB memory: -> device: 0, name: NVIDIA RTX 3500 Ada Generation Laptop GPU, pci bus id: 0000:01:00.0, compute capability: 8.9
Downloading data from https://www.cs.toronto.edu/~kriz/cifar-100-python.tar.gz
169001437/169001437 [==============================] - 8s 0us/step
Epoch 1/25
2024-12-01 00:35:40.830076: I tensorflow/compiler/xla/stream_executor/cuda/cuda_dnn.cc:442] Loaded cuDNN version 8600
2024-12-01 00:35:44.185017: W tensorflow/tsl/framework/bfc_allocator.cc:296] Allocator (GPU_0_bfc) ran out of memory trying to allocate 5.09GiB with freed_by_count=0. The caller indicates that this is not a failure, but this may mean that there could be performance gains if more memory were available.
2024-12-01 00:35:44.422820: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x7f40b45739e0 initialized for platform CUDA (this does not guarantee that XLA will be used). Devices:
2024-12-01 00:35:44.422842: I tensorflow/compiler/xla/service/service.cc:176] StreamExecutor device (0): NVIDIA RTX 3500 Ada Generation Laptop GPU, Compute Capability 8.9
2024-12-01 00:35:44.425879: I tensorflow/compiler/mlir/tensorflow/utils/dump_mlir_util.cc:269] disabling MLIR crash reproducer, set env var `MLIR_CRASH_REPRODUCER_DIRECTORY` to enable.
2024-12-01 00:35:44.487522: I ./tensorflow/compiler/jit/device_compiler.h:186] Compiled cluster using XLA! This line is logged at most once for the lifetime of the process.
2024-12-01 00:35:51.967788: W tensorflow/tsl/framework/bfc_allocator.cc:296] Allocator (GPU_0_bfc) ran out of memory trying to allocate 4.02GiB with freed_by_count=0. The caller indicates that this is not a failure, but this may mean that there could be performance gains if more memory were available.
2024-12-01 00:35:51.967857: W tensorflow/tsl/framework/bfc_allocator.cc:296] Allocator (GPU_0_bfc) ran out of memory trying to allocate 4.02GiB with freed_by_count=0. The caller indicates that this is not a failure, but this may mean that there could be performance gains if more memory were available.
2024-12-01 00:35:51.967866: W tensorflow/tsl/framework/bfc_allocator.cc:296] Allocator (GPU_0_bfc) ran out of memory trying to allocate 4.14GiB with freed_by_count=0. The caller indicates that this is not a failure, but this may mean that there could be performance gains if more memory were available.
2024-12-01 00:35:52.727086: W tensorflow/tsl/framework/bfc_allocator.cc:296] Allocator (GPU_0_bfc) ran out of memory trying to allocate 2.41GiB with freed_by_count=0. The caller indicates that this is not a failure, but this may mean that there could be performance gains if more memory were available.
2024-12-01 00:35:57.297448: W tensorflow/tsl/framework/bfc_allocator.cc:296] Allocator (GPU_0_bfc) ran out of memory trying to allocate 4.33GiB with freed_by_count=0. The caller indicates that this is not a failure, but this may mean that there could be performance gains if more memory were available.
2024-12-01 00:35:57.297528: W tensorflow/tsl/framework/bfc_allocator.cc:296] Allocator (GPU_0_bfc) ran out of memory trying to allocate 4.33GiB with freed_by_count=0. The caller indicates that this is not a failure, but this may mean that there could be performance gains if more memory were available.
12/13 [==========================>...] - ETA: 0s - loss: 5.3903 - accuracy: 0.02432024-12-01 00:36:09.574078: W tensorflow/tsl/framework/bfc_allocator.cc:296] Allocator (GPU_0_bfc) ran out of memory trying to allocate 4.60GiB with freed_by_count=0. The caller indicates that this is not a failure, but this may mean that there could be performance gains if more memory were available.
13/13 [==============================] - 39s 763ms/step - loss: 5.3775 - accuracy: 0.0247
Epoch 2/25
13/13 [==============================] - 5s 385ms/step - loss: 4.2209 - accuracy: 0.0707
Epoch 3/25
13/13 [==============================] - 5s 370ms/step - loss: 3.8432 - accuracy: 0.1215
Epoch 4/25
13/13 [==============================] - 5s 368ms/step - loss: 3.5030 - accuracy: 0.1739
Epoch 5/25
13/13 [==============================] - 5s 376ms/step - loss: 3.3365 - accuracy: 0.2107
Epoch 6/25
13/13 [==============================] - 5s 371ms/step - loss: 3.0246 - accuracy: 0.2688
Epoch 7/25
13/13 [==============================] - 5s 373ms/step - loss: 2.6487 - accuracy: 0.3396
Epoch 8/25
13/13 [==============================] - 5s 374ms/step - loss: 2.2471 - accuracy: 0.4246
Epoch 9/25
13/13 [==============================] - 5s 375ms/step - loss: 1.9108 - accuracy: 0.4988
Epoch 10/25
13/13 [==============================] - 5s 374ms/step - loss: 1.6222 - accuracy: 0.5708
Epoch 11/25
13/13 [==============================] - 5s 376ms/step - loss: 1.4085 - accuracy: 0.6119
Epoch 12/25
13/13 [==============================] - 5s 385ms/step - loss: 1.1184 - accuracy: 0.6859
Epoch 13/25
13/13 [==============================] - 5s 396ms/step - loss: 1.0108 - accuracy: 0.7160
Epoch 14/25
13/13 [==============================] - 5s 397ms/step - loss: 0.8975 - accuracy: 0.7460
Epoch 15/25
13/13 [==============================] - 5s 417ms/step - loss: 0.6930 - accuracy: 0.8035
Epoch 16/25
13/13 [==============================] - 5s 415ms/step - loss: 0.5735 - accuracy: 0.8375
Epoch 17/25
13/13 [==============================] - 5s 414ms/step - loss: 0.4665 - accuracy: 0.8621
Epoch 18/25
13/13 [==============================] - 5s 389ms/step - loss: 0.3995 - accuracy: 0.8819
Epoch 19/25
13/13 [==============================] - 5s 375ms/step - loss: 0.3647 - accuracy: 0.8939
Epoch 20/25
13/13 [==============================] - 5s 384ms/step - loss: 0.3451 - accuracy: 0.8980
Epoch 21/25
13/13 [==============================] - 5s 381ms/step - loss: 0.3194 - accuracy: 0.9040
Epoch 22/25
13/13 [==============================] - 5s 411ms/step - loss: 0.2781 - accuracy: 0.9201
Epoch 23/25
13/13 [==============================] - 5s 401ms/step - loss: 0.4304 - accuracy: 0.8811
Epoch 24/25
13/13 [==============================] - 5s 382ms/step - loss: 0.2810 - accuracy: 0.9201
Epoch 25/25
13/13 [==============================] - 5s 382ms/step - loss: 0.1870 - accuracy: 0.9457