Open tlb1galaxy opened 1 year ago
So this issue seems to stem from setting resource limits below what the pod/container actually works. Figured this out with trial and error. Once I set the limits higher, the pod was able to function. I have noticed the same issue with other components within the stack (Reaper, Cassandra, Medusa)
Suggestion: Document the minimum resources required to run each component (pod/container). I know every situation can be covered but at least a baseline for people starting with the stack. The JVM heap size has a documented default, which is helpful.
Thanks @tlb1galaxy, I've updated the issue title to match your findings.
What happened? New deployment of K8ssandracluster via Operator to include Stargate fails to startup.
How to reproduce it (as minimally and precisely as possible): brand new deployment of k8ssandra-operator
Environment
K8ssandra Operator version:
docker.io/k8ssandra/k8ssandra-operator:v1.2.1
* Kubernetes version information:```yaml clientVersion: buildDate: "2022-04-13T19:07:00Z" compiler: gc gitCommit: 696a9fdd2a58340e61e0d815c5769d266fca0802 gitTreeState: clean gitVersion: v1.21.12 goVersion: go1.16.15 major: "1" minor: "21" platform: linux/amd64 serverVersion: buildDate: "2022-04-13T19:01:10Z" compiler: gc gitCommit: 696a9fdd2a58340e61e0d815c5769d266fca0802 gitTreeState: clean gitVersion: v1.21.12 goVersion: go1.16.15 major: "1" minor: "21" platform: linux/amd64Kubernetes cluster kind:
kubeadm
* Manifests:30-k8ssandra-k8ssandracluster.yaml``` yaml
apiVersion: k8ssandra.io/v1alpha1 kind: K8ssandraCluster metadata: name: demo spec:
auth: true
cassandra: serverVersion: "4.0.7" softPodAntiAffinity: true datacenters:
racks:
- name: rack1
stargate: size: 1 resources: limits: cpu: "250m" memory: 512Mi requests: cpu: "250m" memory: 512Mi heapSize: 256Mi allowStargateOnDataNodes: true
affinity:
podAffinity:
preferredDuringSchedulingIgnoredDuringExecution:
...
NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES demo-dc1-default-stargate-deployment-7d788488c4-w69dh 0/1 Running 10 25m 192.168.205.5 tlb1-node-w2
demo-dc1-default-sts-0 2/2 Running 0 33m 192.168.205.13 tlb1-node-w2
demo-dc1-default-sts-1 2/2 Running 0 33m 192.168.57.17 tlb1-node-w1
demo-dc1-default-sts-2 2/2 Running 0 33m 192.168.221.25 tlb1-node-w3
k8ssandra-operator-56fb7bf949-6nx4d 1/1 Running 0 3h32m 192.168.57.30 tlb1-node-w1
k8ssandra-operator-cass-operator-87d58f57d-g594s 1/1 Running 4 3h32m 192.168.57.62 tlb1-node-w1
1.669311384718185e+09 INFO controller.k8ssandracluster Preparing to update replication for system keyspaces {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "K8SContext": "", "replication": {"dc1":3}} 1.6693113847182887e+09 INFO controller.k8ssandracluster Ensuring that keyspace system_traces exists in cluster demo... {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "K8SContext": ""} 1.669311384718559e+09 INFO controller.k8ssandracluster client::callNodeMgmtEndpoint {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "K8SContext": ""} 1.6693113847225592e+09 INFO controller.k8ssandracluster keyspace system_traces already exists in cluster demo {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "K8SContext": ""} 1.6693113847227492e+09 INFO controller.k8ssandracluster client::callNodeMgmtEndpoint {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "K8SContext": ""} 1.6693113847271338e+09 INFO controller.k8ssandracluster Successfully got keyspace system_traces replication {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "K8SContext": ""} 1.669311384727168e+09 INFO controller.k8ssandracluster Keyspace system_traces has desired replication {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "K8SContext": ""} 1.6693113847271755e+09 INFO controller.k8ssandracluster Ensuring that keyspace system_distributed exists in cluster demo... {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "K8SContext": ""} 1.6693113847273576e+09 INFO controller.k8ssandracluster client::callNodeMgmtEndpoint {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "K8SContext": ""} 1.6693113847314203e+09 INFO controller.k8ssandracluster keyspace system_distributed already exists in cluster demo {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "K8SContext": ""} 1.66931138473166e+09 INFO controller.k8ssandracluster client::callNodeMgmtEndpoint {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "K8SContext": ""} 1.6693113847365508e+09 INFO controller.k8ssandracluster Successfully got keyspace system_distributed replication {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "K8SContext": ""} 1.6693113847366087e+09 INFO controller.k8ssandracluster Keyspace system_distributed has desired replication {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "K8SContext": ""} 1.6693113847366242e+09 INFO controller.k8ssandracluster Ensuring that keyspace system_auth exists in cluster demo... {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "K8SContext": ""} 1.669311384736776e+09 INFO controller.k8ssandracluster client::callNodeMgmtEndpoint {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "K8SContext": ""} 1.6693113847400286e+09 INFO controller.k8ssandracluster keyspace system_auth already exists in cluster demo {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "K8SContext": ""} 1.6693113847402036e+09 INFO controller.k8ssandracluster client::callNodeMgmtEndpoint {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "K8SContext": ""} 1.669311384744554e+09 INFO controller.k8ssandracluster Successfully got keyspace system_auth replication {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "K8SContext": ""} 1.6693113847445831e+09 INFO controller.k8ssandracluster Keyspace system_auth has desired replication {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "K8SContext": ""} 1.669311384753976e+09 DEBUG controller-runtime.webhook.webhooks received request {"webhook": "/validate-k8ssandra-io-v1alpha1-k8ssandracluster", "UID": "3c471ae8-70ea-4080-8ab4-23a4072e0316", "kind": "k8ssandra.io/v1alpha1, Kind=K8ssandraCluster", "resource": {"group":"k8ssandra.io","version":"v1alpha1","resource":"k8ssandraclusters"}} 1.669311384754744e+09 INFO k8ssandracluster-webhook validate K8ssandraCluster update {"K8ssandraCluster": "demo"} 1.6693113847548587e+09 DEBUG controller-runtime.webhook.webhooks wrote response {"webhook": "/validate-k8ssandra-io-v1alpha1-k8ssandracluster", "code": 200, "reason": "", "UID": "3c471ae8-70ea-4080-8ab4-23a4072e0316", "allowed": true} 1.6693113852773392e+09 INFO controller.k8ssandracluster Reconciling Stargate auth keyspace data_endpoint_auth {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "K8SContext": ""} 1.6693113852773778e+09 INFO controller.k8ssandracluster Ensuring that keyspace data_endpoint_auth exists in cluster demo... {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "K8SContext": ""} 1.6693113852775557e+09 INFO controller.k8ssandracluster client::callNodeMgmtEndpoint {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "K8SContext": ""} 1.669311385281156e+09 INFO controller.k8ssandracluster keyspace data_endpoint_auth already exists in cluster demo {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "K8SContext": ""} 1.6693113852814054e+09 INFO controller.k8ssandracluster client::callNodeMgmtEndpoint {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "K8SContext": ""} 1.6693113852877092e+09 INFO controller.k8ssandracluster Successfully got keyspace data_endpoint_auth replication {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "K8SContext": ""} 1.669311385287738e+09 INFO controller.k8ssandracluster Keyspace data_endpoint_auth has desired replication {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "K8SContext": ""} 1.669311385287745e+09 INFO controller.k8ssandracluster Reconciling Stargate auth table data_endpoint_auth.token {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "K8SContext": ""} 1.6693113852879176e+09 INFO controller.k8ssandracluster client::callNodeMgmtEndpoint {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "K8SContext": ""} 1.6693113852932274e+09 INFO controller.k8ssandracluster Successfully got keyspace data_endpoint_auth tables {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "K8SContext": ""} 1.6693113852932549e+09 INFO controller.k8ssandracluster Table data_endpoint_auth.token already exists {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "K8SContext": ""} 1.6693113852932658e+09 INFO controller.k8ssandracluster All DCs reconciled {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo"} 1.6693113852932825e+09 INFO controller.k8ssandracluster Reconciling Stargate and Reaper for dc dc1 {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1"} 1.669311385293289e+09 INFO controller.k8ssandracluster reconciling telemetry {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1"} 1.66931138529332e+09 INFO controller.k8ssandracluster merged TelemetrySpec constructed {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "mergedSpec": null, "cluster": "demo"} I1124 17:36:27.044352 1 request.go:665] Waited for 1.748218688s due to client-side throttling, not priority and fairness, request: GET:https://10.96.0.1:443/apis/storage.k8s.io/v1beta1?timeout=32s 1.6693113870963464e+09 INFO controller.k8ssandracluster Reconcile Stargate {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "Stargate": "k8ssandra-operator/demo-dc1-stargate"} 1.6693113870965798e+09 INFO controller.k8ssandracluster Waiting for Stargate to become ready {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo", "CassandraDatacenter": "k8ssandra-operator/dc1", "Stargate": "k8ssandra-operator/demo-dc1-stargate"} 1.6693113871966133e+09 INFO controller.k8ssandracluster updated k8ssandracluster status {"reconciler group": "k8ssandra.io", "reconciler kind": "K8ssandraCluster", "name": "demo", "namespace": "k8ssandra-operator", "K8ssandraCluster": "k8ssandra-operator/demo"} 1.6693113927903256e+09 INFO controller.stargate Fetching Stargate resource {"reconciler group": "stargate.k8ssandra.io", "reconciler kind": "Stargate", "name": "demo-dc1-stargate", "namespace": "k8ssandra-operator", "Stargate": {"namespace": "k8ssandra-operator", "name": "demo-dc1-stargate"}} 1.6693113927904809e+09 INFO controller.stargate Fetching CassandraDatacenter resource {"reconciler group": "stargate.k8ssandra.io", "reconciler kind": "Stargate", "name": "demo-dc1-stargate", "namespace": "k8ssandra-operator", "CassandraDatacenter": {"namespace": "k8ssandra-operator", "name": "dc1"}} 1.669311392790532e+09 INFO controller.stargate Reconciling Stargate configmap {"reconciler group": "stargate.k8ssandra.io", "reconciler kind": "Stargate", "name": "demo-dc1-stargate", "namespace": "k8ssandra-operator"} 1.6693113927905777e+09 INFO controller.stargate Reconciling Stargate Cassandra yaml configMap on namespace k8ssandra-operator for cluster demo and dc dc1 {"reconciler group": "stargate.k8ssandra.io", "reconciler kind": "Stargate", "name": "demo-dc1-stargate", "namespace": "k8ssandra-operator"} 1.6693113927906632e+09 INFO controller.stargate Stargate ConfigMap successfully reconciled {"reconciler group": "stargate.k8ssandra.io", "reconciler kind": "Stargate", "name": "demo-dc1-stargate", "namespace": "k8ssandra-operator", "StargateConfigMap": "k8ssandra-operator/demo-dc1-cassandra-config"} 1.6693113927911608e+09 INFO controller.stargate reconciling telemetry {"reconciler group": "stargate.k8ssandra.io", "reconciler kind": "Stargate", "name": "demo-dc1-stargate", "namespace": "k8ssandra-operator", "stargate": "demo-dc1-stargate"} 1.6693113945947533e+09 INFO controller.stargate Waiting for deployments to be rolled out {"reconciler group": "stargate.k8ssandra.io", "reconciler kind": "Stargate", "name": "demo-dc1-stargate", "namespace": "k8ssandra-operator", "Stargate": {"namespace": "k8ssandra-operator", "name": "demo-dc1-stargate"}}
INFO [epollEventLoopGroup-7-3] 2022-11-24 17:37:28,147 Keyspace.java:386 - Creating replication strategy none params KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.NetworkTopologyStrategy, dc1=1}} INFO [epollEventLoopGroup-7-3] 2022-11-24 17:37:28,147 NetworkTopologyStrategy.java:89 - Configured datacenter replicas are dc1:rf(1) INFO [Messaging-EventLoop-3-1] 2022-11-24 17:39:26,839 InboundConnectionInitiator.java:464 - /192.168.205.5:7000(/192.168.205.5:38640)->/192.168.205.13:7000-URGENT_MESSAGES-4347a7c8 messaging connection established, version = 12, framing = CRC, encryption = unencrypted INFO [Messaging-EventLoop-3-1] 2022-11-24 17:39:27,243 OutboundConnection.java:1150 - /192.168.205.13:7000(/192.168.205.13:55068)->/192.168.205.5:7000-URGENT_MESSAGES-cbaee9ed successfully connected, version = 12, framing = LZ4, encryption = unencrypted INFO [GossipStage:1] 2022-11-24 17:39:37,155 Gossiper.java:1364 - Node /192.168.205.5:7000 is now part of the cluster INFO [GossipStage:1] 2022-11-24 17:39:37,157 Gossiper.java:1312 - InetAddress /192.168.205.5:7000 is now UP INFO [Messaging-EventLoop-3-1] 2022-11-24 17:39:44,442 InboundConnectionInitiator.java:464 - /192.168.205.5:7000(/192.168.205.5:38770)->/192.168.205.13:7000-SMALL_MESSAGES-f7f76a0e messaging connection established, version = 12, framing = CRC, encryption = unencrypted INFO [Messaging-EventLoop-3-1] 2022-11-24 17:39:44,538 OutboundConnection.java:1150 - /192.168.205.13:7000(/192.168.205.13:55194)->/192.168.205.5:7000-SMALL_MESSAGES-88064e56 successfully connected, version = 12, framing = LZ4, encryption = unencrypted INFO [Messaging-EventLoop-3-1] 2022-11-24 17:39:44,933 InboundConnectionInitiator.java:464 - /192.168.205.5:7000(/192.168.205.5:38780)->/192.168.205.13:7000-LARGE_MESSAGES-b43f054b messaging connection established, version = 12, framing = CRC, encryption = unencrypted INFO [Messaging-EventLoop-3-1] 2022-11-24 17:39:47,751 NoSpamLogger.java:92 - /192.168.205.13:7000->/192.168.205.5:7000-URGENT_MESSAGES-[no-channel] failed to connect io.netty.channel.AbstractChannel$AnnotatedConnectException: finishConnect(..) failed: Connection refused: /192.168.205.5:7000 Caused by: java.net.ConnectException: finishConnect(..) failed: Connection refused at io.netty.channel.unix.Errors.throwConnectException(Errors.java:124) at io.netty.channel.unix.Socket.finishConnect(Socket.java:251) at io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.doFinishConnect(AbstractEpollChannel.java:673) at io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.finishConnect(AbstractEpollChannel.java:650) at io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.epollOutReady(AbstractEpollChannel.java:530) at io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:470) at io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:378) at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.base/java.lang.Thread.run(Unknown Source) INFO [GossipStage:1] 2022-11-24 17:40:11,846 Gossiper.java:1328 - InetAddress /192.168.205.5:7000 is now DOWN INFO [GossipTasks:1] 2022-11-24 17:40:14,849 Gossiper.java:1057 - FatClient /192.168.205.5:7000 has been silent for 30000ms, removing from gossip INFO [Messaging-EventLoop-3-1] 2022-11-24 17:40:17,820 NoSpamLogger.java:92 - /192.168.205.13:7000->/192.168.205.5:7000-URGENT_MESSAGES-[no-channel] failed to connect io.netty.channel.AbstractChannel$AnnotatedConnectException: finishConnect(..) failed: Connection refused: /192.168.205.5:7000 Caused by: java.net.ConnectException: finishConnect(..) failed: Connection refused at io.netty.channel.unix.Errors.throwConnectException(Errors.java:124) at io.netty.channel.unix.Socket.finishConnect(Socket.java:251) at io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.doFinishConnect(AbstractEpollChannel.java:673) at io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.finishConnect(AbstractEpollChannel.java:650) at io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.epollOutReady(AbstractEpollChannel.java:530) at io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:470) at io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:378) at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.base/java.lang.Thread.run(Unknown Source) WARN [epollEventLoopGroup-7-3] 2022-11-24 17:40:46,726 K8SeedProvider4x.java:58 - Seed provider couldn't lookup host demo-dc1-additional-seed-service INFO [epollEventLoopGroup-7-3] 2022-11-24 17:40:46,826 Keyspace.java:386 - Creating replication strategy none params KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.NetworkTopologyStrategy, dc1=1}} INFO [epollEventLoopGroup-7-3] 2022-11-24 17:40:46,826 NetworkTopologyStrategy.java:89 - Configured datacenter replicas are dc1:rf(1) INFO [Messaging-EventLoop-3-1] 2022-11-24 17:40:47,841 NoSpamLogger.java:92 - /192.168.205.13:7000->/192.168.205.5:7000-URGENT_MESSAGES-[no-channel] failed to connect io.netty.channel.AbstractChannel$AnnotatedConnectException: finishConnect(..) failed: Connection refused: /192.168.205.5:7000 Caused by: java.net.ConnectException: finishConnect(..) failed: Connection refused at io.netty.channel.unix.Errors.throwConnectException(Errors.java:124) at io.netty.channel.unix.Socket.finishConnect(Socket.java:251) at io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.doFinishConnect(AbstractEpollChannel.java:673) at io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.finishConnect(AbstractEpollChannel.java:650) at io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.epollOutReady(AbstractEpollChannel.java:530) at io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:470) at io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:378) at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.base/java.lang.Thread.run(Unknown Source) WARN [epollEventLoopGroup-7-3] 2022-11-24 17:40:52,735 K8SeedProvider4x.java:58 - Seed provider couldn't lookup host demo-dc1-additional-seed-service INFO [epollEventLoopGroup-7-3] 2022-11-24 17:40:52,764 Keyspace.java:386 - Creating replication strategy none params KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.NetworkTopologyStrategy, dc1=1}} INFO [epollEventLoopGroup-7-3] 2022-11-24 17:40:52,765 NetworkTopologyStrategy.java:89 - Configured datacenter replicas are dc1:rf(1) INFO [Messaging-EventLoop-3-1] 2022-11-24 17:41:14,241 InboundConnectionInitiator.java:464 - /192.168.205.5:7000(/192.168.205.5:39592)->/192.168.205.13:7000-URGENT_MESSAGES-d3ff6833 messaging connection established, version = 12, framing = CRC, encryption = unencrypted INFO [Messaging-EventLoop-3-1] 2022-11-24 17:41:14,439 OutboundConnection.java:1150 - /192.168.205.13:7000(/192.168.205.13:56006)->/192.168.205.5:7000-URGENT_MESSAGES-27f98657 successfully connected, version = 12, framing = LZ4, encryption = unencrypted INFO [GossipStage:1] 2022-11-24 17:41:16,232 Gossiper.java:1364 - Node /192.168.205.5:7000 is now part of the cluster INFO [GossipStage:1] 2022-11-24 17:41:16,234 Gossiper.java:1312 - InetAddress /192.168.205.5:7000 is now UP INFO [Messaging-EventLoop-3-1] 2022-11-24 17:41:21,938 InboundConnectionInitiator.java:464 - /192.168.205.5:7000(/192.168.205.5:39668)->/192.168.205.13:7000-SMALL_MESSAGES-0c600d1b messaging connection established, version = 12, framing = CRC, encryption = unencrypted INFO [Messaging-EventLoop-3-1] 2022-11-24 17:41:22,240 OutboundConnection.java:1150 - /192.168.205.13:7000(/192.168.205.13:56096)->/192.168.205.5:7000-SMALL_MESSAGES-96fd42cc successfully connected, version = 12, framing = LZ4, encryption = unencrypted INFO [Messaging-EventLoop-3-1] 2022-11-24 17:41:22,735 InboundConnectionInitiator.java:464 - /192.168.205.5:7000(/192.168.205.5:39684)->/192.168.205.13:7000-LARGE_MESSAGES-82bb03b4 messaging connection established, version = 12, framing = CRC, encryption = unencrypted INFO [Messaging-EventLoop-3-1] 2022-11-24 17:41:25,136 NoSpamLogger.java:92 - /192.168.205.13:7000->/192.168.205.5:7000-URGENT_MESSAGES-[no-channel] failed to connect io.netty.channel.AbstractChannel$AnnotatedConnectException: finishConnect(..) failed: Connection refused: /192.168.205.5:7000 Caused by: java.net.ConnectException: finishConnect(..) failed: Connection refused at io.netty.channel.unix.Errors.throwConnectException(Errors.java:124) at io.netty.channel.unix.Socket.finishConnect(Socket.java:251) at io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.doFinishConnect(AbstractEpollChannel.java:673) at io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.finishConnect(AbstractEpollChannel.java:650) at io.netty.channel.epoll.AbstractEpollChannel$AbstractEpollUnsafe.epollOutReady(AbstractEpollChannel.java:530) at io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:470) at io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:378) at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.base/java.lang.Thread.run(Unknown Source) INFO [GossipStage:1] 2022-11-24 17:41:37,143 Gossiper.java:1328 - InetAddress /192.168.205.5:7000 is now DOWN
INFO [MigrationStage:1] 2022-11-24 17:47:31,037 NetworkTopologyStrategy.java:88 - Configured datacenter replicas are dc1:rf(3) INFO [MigrationStage:1] 2022-11-24 17:47:31,038 ColumnFamilyStore.java:385 - Initializing system_traces.events INFO [MigrationStage:1] 2022-11-24 17:47:31,434 ColumnFamilyStore.java:385 - Initializing system_traces.sessions INFO [MigrationStage:1] 2022-11-24 17:47:31,741 Keyspace.java:386 - Creating replication strategy system_auth params KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.NetworkTopologyStrategy, dc1=3}} INFO [MigrationStage:1] 2022-11-24 17:47:31,741 NetworkTopologyStrategy.java:88 - Configured datacenter replicas are dc1:rf(3) INFO [main] 2022-11-24 17:47:34,342 Gossiper.java:2206 - No gossip backlog; proceeding INFO [main] 2022-11-24 17:47:34,354 StartupClusterConnectivityChecker.java:111 - Blocking coordination until only a single peer is DOWN in the local datacenter, timeout=10s INFO [Messaging-EventLoop-3-1] 2022-11-24 17:47:34,536 OutboundConnection.java:1150 - /192.168.205.5:7000(/192.168.205.5:42720)->demo-seed-service.k8ssandra-operator.svc.cluster.local/192.168.205.13:7000-LARGE_MESSAGES-7b8c5bd9 successfully connected, version = 12, framing = CRC, encryption = unencrypted INFO [Messaging-EventLoop-3-1] 2022-11-24 17:47:34,539 OutboundConnection.java:1150 - /192.168.205.5:7000(/192.168.205.5:42276)->demo-seed-service.k8ssandra-operator.svc.cluster.local/192.168.57.17:7000-LARGE_MESSAGES-8b6c39f7 successfully connected, version = 12, framing = CRC, encryption = unencrypted INFO [Messaging-EventLoop-3-1] 2022-11-24 17:47:34,633 OutboundConnection.java:1150 - /192.168.205.5:7000(/192.168.205.5:37600)->demo-seed-service.k8ssandra-operator.svc.cluster.local/192.168.221.25:7000-LARGE_MESSAGES-f45d43ee successfully connected, version = 12, framing = CRC, encryption = unencrypted INFO [Messaging-EventLoop-3-1] 2022-11-24 17:47:34,734 InboundConnectionInitiator.java:464 - /192.168.221.25:7000(/192.168.221.25:45236)->/192.168.205.5:7000-LARGE_MESSAGES-10fb42be messaging connection established, version = 12, framing = LZ4, encryption = unencrypted INFO [Messaging-EventLoop-3-1] 2022-11-24 17:47:34,735 InboundConnectionInitiator.java:464 - /192.168.205.13:7000(/192.168.205.13:59148)->/192.168.205.5:7000-LARGE_MESSAGES-34d65030 messaging connection established, version = 12, framing = LZ4, encryption = unencrypted INFO [Messaging-EventLoop-3-1] 2022-11-24 17:47:34,735 InboundConnectionInitiator.java:464 - /192.168.57.17:7000(/192.168.57.17:38906)->/192.168.205.5:7000-LARGE_MESSAGES-e2883410 messaging connection established, version = 12, framing = LZ4, encryption = unencrypted INFO [main] 2022-11-24 17:47:34,833 StartupClusterConnectivityChecker.java:166 - Ensured sufficient healthy connections with [dc1] after 475 milliseconds INFO [main] 2022-11-24 17:47:34,833 CassandraDaemon.java:695 - Not starting native transport as requested. Use JMX (StorageService->startNativeTransport()) or nodetool (enablebinary) to start it INFO [main] 2022-11-24 17:47:35,038 Keyspace.java:386 - Creating replication strategy stargate_system params KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.LocalStrategy}} INFO [main] 2022-11-24 17:47:35,039 ColumnFamilyStore.java:385 - Initializing stargate_system.local INFO [main] 2022-11-24 17:47:35,538 ColumnFamilyStore.java:385 - Initializing stargate_system.peers INFO [main] 2022-11-24 17:47:36,136 ColumnFamilyStore.java:385 - Initializing stargate_system.peers_v2 INFO [main] 2022-11-24 17:47:37,634 BaseActivator.java:178 - Registering persistence-cassandra-4.0 as io.stargate.db.Persistence INFO [main] 2022-11-24 17:47:37,635 BaseActivator.java:185 - Started persistence-cassandra-4.0 INFO [main] 2022-11-24 17:47:37,636 BaseActivator.java:178 - Registering core services as io.stargate.core.metrics.api.MetricsScraper INFO [main] 2022-11-24 17:47:37,636 BaseActivator.java:178 - Registering core services as com.codahale.metrics.health.HealthCheckRegistry INFO [main] 2022-11-24 17:47:37,637 BaseActivator.java:178 - Registering core services as io.stargate.core.metrics.api.HttpMetricsTagProvider INFO [main] 2022-11-24 17:47:37,637 BaseActivator.java:185 - Started core services Starting bundle io.stargate.cql INFO [main] 2022-11-24 17:47:39,437 BaseActivator.java:97 - Starting CQL ... Starting bundle io.stargate.graphql INFO [main] 2022-11-24 17:47:40,833 BaseActivator.java:97 - Starting GraphQL ... Starting bundle io.stargate.grpc INFO [main] 2022-11-24 17:47:41,537 BaseActivator.java:97 - Starting gRPC ... Starting bundle io.stargate.health INFO [main] 2022-11-24 17:47:42,446 BaseActivator.java:97 - Starting healthchecker ... Starting bundle org.apache.felix.scr Starting bundle null Starting bundle null Starting bundle osgi.cmpn Starting bundle org.osgi.util.function Starting bundle org.osgi.util.promise Starting bundle io.stargate.db INFO [main] 2022-11-24 17:47:42,638 BaseActivator.java:97 - Starting DB services ... INFO [main] 2022-11-24 17:47:42,641 BaseActivator.java:178 - Registering DB services as io.stargate.db.Persistence INFO [main] 2022-11-24 17:47:42,642 BaseActivator.java:178 - Registering DB services as io.stargate.db.datastore.DataStoreFactory INFO [main] 2022-11-24 17:47:42,643 HealthCheckerActivator.java:50 - Starting healthchecker.... INFO [Service Thread] 2022-11-24 17:47:44,454 GCInspector.java:299 - MarkSweepCompact GC in 416ms. Eden Space: 1024208 -> 0; Survivor Space: 8912888 -> 0; Tenured Gen: 81992056 -> 39566024 INFO [main] 2022-11-24 17:47:44,843 Version.java:21 - HV000001: Hibernate Validator null INFO [main] 2022-11-24 17:47:46,034 FileUtil.java:249 - No oshi.properties file found from ClassLoader sun.misc.Launcher$AppClassLoader70dea4e INFO [main] 2022-11-24 17:47:46,034 FileUtil.java:249 - No oshi.properties file found from ClassLoader sun.misc.Launcher$AppClassLoader70dea4e INFO [main] 2022-11-24 17:47:51,542 Log.java:170 - Logging initialized 70908ms to org.eclipse.jetty.util.log.Slf4jLog INFO [main] 2022-11-24 17:47:52,136 SimpleServerFactory.java:142 - Registering jersey handler with root path prefix: / INFO [main] 2022-11-24 17:47:52,138 SimpleServerFactory.java:145 - Registering admin handler with root path prefix: /admin INFO [main] 2022-11-24 17:47:52,150 SimpleServerFactory.java:142 - Registering jersey handler with root path prefix: / INFO [main] 2022-11-24 17:47:52,150 SimpleServerFactory.java:145 - Registering admin handler with root path prefix: /admin INFO [main] 2022-11-24 17:47:52,150 AbstractServerFactory.java:717 - Starting Server INFO [main] 2022-11-24 17:47:52,939 SetUIDListener.java:217 - Opened Server@5ba603a3{HTTP/1.1, (http/1.1)}{0.0.0.0:8084} INFO [main] 2022-11-24 17:47:52,941 Server.java:375 - jetty-9.4.41.v20210516; built: 2021-05-16T23:56:28.993Z; git: 98607f93c7833e7dc59489b13f3cb0a114fb9f4c; jvm 1.8.0_292-b10 INFO [main] 2022-11-24 17:47:53,132 AdminEnvironment.java:74 - tasks =
INFO [main] 2022-11-24 17:47:53,138 ContextHandler.java:916 - Started i.d.j.MutableServletContextHandler@5dd44130{/admin,null,AVAILABLE} INFO [main] 2022-11-24 17:47:57,136 DropwizardResourceConfig.java:287 - The following paths were found for the configured resources:
INFO [main] 2022-11-24 17:47:57,138 ContextHandler.java:916 - Started i.d.j.MutableServletContextHandler@7b555abb{/,null,AVAILABLE} INFO [main] 2022-11-24 17:47:57,238 AbstractConnector.java:331 - Started Server@5ba603a3{HTTP/1.1, (http/1.1)}{0.0.0.0:8084} INFO [main] 2022-11-24 17:47:57,239 Server.java:415 - Started 76606ms INFO [main] 2022-11-24 17:47:57,240 HealthCheckerActivator.java:69 - Started healthchecker.... INFO [main] 2022-11-24 17:47:57,240 BaseActivator.java:185 - Started healthchecker INFO [main] 2022-11-24 17:47:57,340 AuthnTableBasedService.java:78 - Initializing keyspace data_endpoint_auth and table token for table based auth INFO [main] 2022-11-24 17:47:58,233 BaseActivator.java:178 - Registering authnTableBasedService and authzTableBasedServie as io.stargate.auth.AuthenticationService WARN [dw-99 - GET /checker/liveness] 2022-11-24 17:47:59,435 BundleStateChecker.java:48 - Inactive bundles: [io.stargate.db, io.stargate.db.limiter.global, io.stargate.web] INFO [StorageServiceShutdownHook] 2022-11-24 17:48:00,132 HintsService.java:220 - Paused hints dispatch WARN [StorageServiceShutdownHook] 2022-11-24 17:48:00,137 Gossiper.java:1993 - No local state, state is in silent shutdown, or node hasn't joined, not announcing shutdown INFO [StorageServiceShutdownHook] 2022-11-24 17:48:00,138 MessagingService.java:441 - Waiting for messaging service to quiesce WARN [dw-100 - GET /checker/readiness] 2022-11-24 17:48:00,138 BundleStateChecker.java:48 - Inactive bundles: [io.stargate.db, io.stargate.db.limiter.global, io.stargate.web] WARN [dw-100 - GET /checker/readiness] 2022-11-24 17:48:00,233 CheckerResource.java:88 - Failed health check: bundles WARN [dw-100 - GET /checker/readiness] 2022-11-24 17:48:00,233 CheckerResource.java:88 - Failed health check: grpc WARN [dw-100 - GET /checker/readiness] 2022-11-24 17:48:00,233 CheckerResource.java:94 - Missing required health checks: [graphql, restapi, cql] INFO [JettyShutdownThread] 2022-11-24 17:48:00,538 AbstractConnector.java:381 - Stopped Server@5ba603a3{HTTP/1.1, (http/1.1)}{0.0.0.0:8084} INFO [dw-99] 2022-11-24 17:48:00,634 Slf4jRequestLogWriter.java:62 - 10.253.50.16 - - [24/Nov/2022:17:47:58 +0000] "GET /checker/liveness HTTP/1.1" 503 4 "-" "kube-probe/1.21" 2296 INFO [dw-100] 2022-11-24 17:48:00,634 Slf4jRequestLogWriter.java:62 - 10.253.50.16 - - [24/Nov/2022:17:47:58 +0000] "GET /checker/readiness HTTP/1.1" 503 9 "-" "kube-probe/1.21" 2296 INFO [JettyShutdownThread] 2022-11-24 17:48:00,840 ContextHandler.java:1154 - Stopped i.d.j.MutableServletContextHandler@7b555abb{/,null,STOPPED} INFO [JettyShutdownThread] 2022-11-24 17:48:00,933 ContextHandler.java:1154 - Stopped i.d.j.MutableServletContextHandler@5dd44130{/admin,null,STOPPED} INFO [main] 2022-11-24 17:48:01,456 BaseActivator.java:185 - Started gRPC INFO [main] 2022-11-24 17:48:01,532 BaseActivator.java:164 - The persistence-cassandra-4.0 is already started. Ignoring the start request. INFO [StorageServiceShutdownHook] 2022-11-24 17:48:02,546 ColumnFamilyStore.java:878 - Enqueuing flush of local: 40.037KiB (0%) on-heap, 0.000KiB (0%) off-heap INFO [PerDiskMemtableFlushWriter_0:1] 2022-11-24 17:48:02,635 Memtable.java:469 - Writing Memtable-local@12264324(9.268KiB serialized bytes, 1 ops, 0%/0% of on/off-heap limit), flushed range = (min(-9223372036854775808), max(9223372036854775807)] INFO [PerDiskMemtableFlushWriter_0:1] 2022-11-24 17:48:02,732 Memtable.java:498 - Completed flushing /tmp/stargate-persistence-cassandra-4.01716269871846557906/data/stargate_system/local-bf0dc51a28c73c8d9b9a5b70e206e04d/nb-1-big-Data.db (5.532KiB) for commitlog position CommitLogPosition(segmentId=1669312020041, position=61409) INFO [StorageServiceShutdownHook] 2022-11-24 17:48:02,934 ColumnFamilyStore.java:878 - Enqueuing flush of peers_v2: 12.523KiB (0%) on-heap, 0.000KiB (0%) off-heap INFO [StorageServiceShutdownHook] 2022-11-24 17:48:02,935 ColumnFamilyStore.java:878 - Enqueuing flush of peers: 11.824KiB (0%) on-heap, 0.000KiB (0%) off-heap INFO [StorageServiceShutdownHook] 2022-11-24 17:48:02,935 ColumnFamilyStore.java:878 - Enqueuing flush of local: 0.501KiB (0%) on-heap, 0.000KiB (0%) off-heap INFO [PerDiskMemtableFlushWriter_0:2] 2022-11-24 17:48:02,940 Memtable.java:469 - Writing Memtable-peers_v2@1568217655(2.280KiB serialized bytes, 40 ops, 0%/0% of on/off-heap limit), flushed range = (min(-9223372036854775808), max(9223372036854775807)] INFO [PerDiskMemtableFlushWriter_0:2] 2022-11-24 17:48:02,941 Memtable.java:498 - Completed flushing /tmp/stargate-persistence-cassandra-4.01716269871846557906/data/system/peers_v2-c4325fbb8e5e3bafbd070f9250ed818e/nb-1-big-Data.db (1.486KiB) for commitlog position CommitLogPosition(segmentId=1669312020041, position=61409) INFO [PerDiskMemtableFlushWriter_0:1] 2022-11-24 17:48:02,942 Memtable.java:469 - Writing Memtable-peers@819013565(2.210KiB serialized bytes, 40 ops, 0%/0% of on/off-heap limit), flushed range = (min(-9223372036854775808), max(9223372036854775807)] INFO [PerDiskMemtableFlushWriter_0:1] 2022-11-24 17:48:03,032 Memtable.java:498 - Completed flushing /tmp/stargate-persistence-cassandra-4.01716269871846557906/data/system/peers-37f71aca7dc2383ba70672528af04d4f/nb-1-big-Data.db (1.457KiB) for commitlog position CommitLogPosition(segmentId=1669312020041, position=61409) INFO [PerDiskMemtableFlushWriter_0:2] 2022-11-24 17:48:03,336 Memtable.java:469 - Writing Memtable-local@710275964(0.060KiB serialized bytes, 3 ops, 0%/0% of on/off-heap limit), flushed range = (min(-9223372036854775808), max(9223372036854775807)] INFO [PerDiskMemtableFlushWriter_0:2] 2022-11-24 17:48:03,336 Memtable.java:498 - Completed flushing /tmp/stargate-persistence-cassandra-4.01716269871846557906/data/system/local-7ad54392bcdd35a684174e047860b377/nb-4-big-Data.db (0.043KiB) for commitlog position CommitLogPosition(segmentId=1669312020041, position=61409) INFO [StorageServiceShutdownHook] 2022-11-24 17:48:03,438 HintsService.java:220 - Paused hints dispatch INFO [main] 2022-11-24 17:48:03,537 Version.java:21 - HV000001: Hibernate Validator null