Closed xueyuan1990 closed 1 month ago
hi, could you please share some details regarding your:
spark-nlp_2.12-5.3.3.jar
? it doesn't look like a Fat JAR we provide in our release notes that comes all the dependenciesCentOS Linux 7 CPU spark-nlp_2.12-5.3.3.jar is provided in this link : https://sparknlp.org/docs/en/install#scala-and-java
<!-- https://mvnrepository.com/artifact/com.johnsnowlabs.nlp/spark-nlp -->
<dependency>
<groupId>com.johnsnowlabs.nlp</groupId>
<artifactId>spark-nlp_2.12</artifactId>
<version>5.3.3</version>
</dependency>
I don't know where is the Fat JAR.
I suspect you are either missing some dependencies or your OS is too old.
Please use the Fat JAR instead and try BertEmbeddings.pretrained()
for this test.
I use the Fat JAR downloaded from https://s3.amazonaws.com/auxdata.johnsnowlabs.com/public/jars/spark-nlp-assembly-5.3.3.jar. And try BertEmbeddings.pretrained("bert_embeddings_chinese_roberta_wwm_ext","zh"). But the error still comes. So you think my OS is too old. What's your suggest version?
Thanks, but not to rule out the model's issue, I need you to just use BertEmbeddings.pretrained()
no model inside. I just need to see if it's the model that has the problem or the OS. (from my experience, CentOS 7 is just way too old to support new TensorFlow/ONNX requirements, but let's see if the default model works)
please just run this code:
bert = BertEmbeddings.pretrained()
The default model is ONNX, so I want to see what happens when you use ONNX instead of TF.
So this is what I was worried about, your Operating System is old. Both TensorFlow and ONNX use a much newer libstdc++
. So you either have to upgrade the machines to a newer release that comes with newer libstdc++
or somehow manually update them.
I upgrade my OS to CentOS Linux release 7.9.2009 (Core) and gcc to gcc (GCC) 8.3.1 20190311 (Red Hat 8.3.1-3). BertEmbeddings.pretrained() can load model successfully. But BertEmbeddings.pretrained("bert_embeddings_chinese_roberta_wwm_ext","zh") still can not. The error still comes:
#
# A fatal error has been detected by the Java Runtime Environment:
#
# SIGILL (0x4) at pc=0x00007f5cd2387260, pid=6533, tid=0x00007f5e08838700
#
# JRE version: Java(TM) SE Runtime Environment (8.0_144-b01) (build 1.8.0_144-b01)
# Java VM: Java HotSpot(TM) 64-Bit Server VM (25.144-b01 mixed mode linux-amd64 compressed oops)
# Problematic frame:
# C [libtensorflow_framework.so.2+0x16bc260] nsync::nsync_mu_init(nsync::nsync_mu_s_*)+0x0
#
# Core dump written. Default location: /home/xueyuan.zhang/article_recom/core or core.6533
#
# If you would like to submit a bug report, please visit:
# http://bugreport.java.com/bugreport/crash.jsp
# The crash happened outside the Java Virtual Machine in native code.
# See problematic frame for where to report the bug.
#
--------------- T H R E A D ---------------
Current thread (0x00007f5e00017800): JavaThread "main" [_thread_in_native, id=6613, stack(0x00007f5e08739000,0x00007f5e08839000)]
siginfo: si_signo: 4 (SIGILL), si_code: 2 (ILL_ILLOPN), si_addr: 0x00007f5cd2387260
Registers:
RAX=0x00007f5ccc0d0a80, RBX=0x00007f5cd2b89aa0, RCX=0x00007f5e00000020, RDX=0x00007f5ccc0d0a80
RSP=0x00007f5e08832c58, RBP=0x00007f5e08832d60, RSI=0x00007f5ccc0d0b00, RDI=0x00007f5ccc0d0a80
R8 =0x00000000000001b1, R9 =0x00007f5e00000740, R10=0x00007f5e000008d0, R11=0x0000000000000002
R12=0x00007f5ccc0d0a80, R13=0x00007f5e08832cd8, R14=0x00007f5e08832ce0, R15=0x00007f5e08832ca0
RIP=0x00007f5cd2387260, EFLAGS=0x0000000000010202, CSGSFS=0x0000000000000033, ERR=0x0000000000000000
TRAPNO=0x0000000000000006
Top of Stack: (sp=0x00007f5e08832c58)
0x00007f5e08832c58: 00007f5cd14357a7 d86fca2c22619800
0x00007f5e08832c68: 00007f5e08832cb0 00007f5ccc0d0c38
0x00007f5e08832c78: 00007f5ccc0d0c50 00007f5e08832c9f
0x00007f5e08832c88: 00007f5ccc0d0c40 61636f6c3674534e
0x00007f5e08832c98: 746563616635656c 00007f5e01785c08
0x00007f5e08832ca8: 00007f5e01785c38 00007f5ccc0d0c30
0x00007f5e08832cb8: 00007f5ccc0d0c40 00007f5ccc0d0c40
0x00007f5e08832cc8: 00007f5cd2b7b440 0000000000000001
0x00007f5e08832cd8: 00007f5e017f2f28 00007f5ccc042ab8
0x00007f5e08832ce8: 00007f5ccc0d0c50 00007f5ccc0d0c60
0x00007f5e08832cf8: 00007f5ccc0d0c60 00007f5cd2b87c80
0x00007f5e08832d08: 00007f5cd2ac0f98 00007f5cd2b886e0
0x00007f5e08832d18: 00007f5e08832d28 0000000000ff0000
0x00007f5e08832d28: 4744c89227432000 00007f5cd2acad00
0x00007f5e08832d38: 00007f5cd28759b0 00007fffd55e7958
0x00007f5e08832d48: 00007f5e0009c720 00007f5cd2876ab0
0x00007f5e08832d58: 0000000000000000 00007f5e08832e30
0x00007f5e08832d68: 00007f5cd140039a 00007f5cd2b874a8
0x00007f5e08832d78: 00007f5cd2afb9ed 00007f5cd2b87db8
0x00007f5e08832d88: 00007f5cd2b874a8 00007f5cd2b86f60
0x00007f5e08832d98: 00007f5cd2afbc25 00007f5cd2b809d8
0x00007f5e08832da8: 00007f5cd2b874a0 00007f5cd2b875c0
0x00007f5e08832db8: 00007f5cd2b0b09d 00007f5e07fd96a0
0x00007f5e08832dc8: 00007f5cd2b86f60 00007f5cd2b89654
0x00007f5e08832dd8: 00007f5cd2b875c0 00007f5cd2b87b60
0x00007f5e08832de8: 00007f5cd2acaccd 00007f5e0009c720
0x00007f5e08832df8: 4744c89227432000 00007f5cd2acad00
0x00007f5e08832e08: 00007f5cd28759b0 00007fffd55e7958
0x00007f5e08832e18: 00007f5e0009c720 00007f5cd2876ab0
0x00007f5e08832e28: 0000000000000000 0000000000000021
0x00007f5e08832e38: 00007f5e086260fa 00000000ffffffff
0x00007f5e08832e48: 0000000000000021 00007fffd55e7958
Instructions: (pc=0x00007f5cd2387260)
0x00007f5cd2387240: 31 c0 c3 66 66 66 66 2e 0f 1f 84 00 00 00 00 00
0x00007f5cd2387250: 8b 07 89 c2 83 e2 fd f0 0f b1 17 75 f3 c3 66 90
0x00007f5cd2387260: c5 f9 ef c0 c5 f8 11 07 c3 0f 1f 80 00 00 00 00
0x00007f5cd2387270: 55 48 89 e5 41 57 4c 8d 7e 08 41 56 41 55 49 89
Register to memory mapping:
RAX=0x00007f5ccc0d0a80 is an unknown value
RBX=0x00007f5cd2b89aa0: _ZNSs4_Rep20_S_empty_rep_storageE+0 in /lib64/libstdc++.so.6 at 0x00007f5cd2a16000
RCX=0x00007f5e00000020 is an unknown value
RDX=0x00007f5ccc0d0a80 is an unknown value
RSP=0x00007f5e08832c58 is pointing into the stack for thread: 0x00007f5e00017800
RBP=0x00007f5e08832d60 is pointing into the stack for thread: 0x00007f5e00017800
RSI=0x00007f5ccc0d0b00 is an unknown value
RDI=0x00007f5ccc0d0a80 is an unknown value
R8 =0x00000000000001b1 is an unknown value
R9 =0x00007f5e00000740 is an unknown value
R10=0x00007f5e000008d0 is an unknown value
R11=0x0000000000000002 is an unknown value
R12=0x00007f5ccc0d0a80 is an unknown value
R13=0x00007f5e08832cd8 is pointing into the stack for thread: 0x00007f5e00017800
R14=0x00007f5e08832ce0 is pointing into the stack for thread: 0x00007f5e00017800
R15=0x00007f5e08832ca0 is pointing into the stack for thread: 0x00007f5e00017800
Stack: [0x00007f5e08739000,0x00007f5e08839000], sp=0x00007f5e08832c58, free space=999k
Native frames: (J=compiled Java code, j=interpreted, Vv=VM code, C=native code)
C [libtensorflow_framework.so.2+0x16bc260] nsync::nsync_mu_init(nsync::nsync_mu_s_*)+0x0
C [libtensorflow_framework.so.2+0x73539a] _GLOBAL__sub_I_loader.cc+0x6a
C [ld-linux-x86-64.so.2+0xf0fa] call_init.part.0+0x9a
Java frames: (J=compiled Java code, j=interpreted, Vv=VM code)
j java.lang.ClassLoader$NativeLibrary.load(Ljava/lang/String;Z)V+0
j java.lang.ClassLoader.loadLibrary0(Ljava/lang/Class;Ljava/io/File;)Z+328
j java.lang.ClassLoader.loadLibrary(Ljava/lang/Class;Ljava/lang/String;Z)V+48
j java.lang.Runtime.load0(Ljava/lang/Class;Ljava/lang/String;)V+57
j java.lang.System.load(Ljava/lang/String;)V+7
j org.bytedeco.javacpp.Loader.loadLibrary(Ljava/lang/Class;[Ljava/net/URL;Ljava/lang/String;[Ljava/lang/String;)Ljava/lang/String;+953
j org.bytedeco.javacpp.Loader.load(Ljava/lang/Class;Ljava/util/Properties;ZLjava/lang/String;)Ljava/lang/String;+597
j org.bytedeco.javacpp.Loader.load(Ljava/lang/Class;Ljava/util/Properties;Z)Ljava/lang/String;+4
j org.bytedeco.javacpp.Loader.load()Ljava/lang/String;+10
j org.tensorflow.internal.c_api.global.tensorflow.<clinit>()V+0
v ~StubRoutines::call_stub
J 1053 java.lang.Class.forName0(Ljava/lang/String;ZLjava/lang/ClassLoader;Ljava/lang/Class;)Ljava/lang/Class; (0 bytes) @ 0x00007f5df13cc2ba [0x00007f5df13cc240+0x7a]
J 5231 C2 java.lang.Class.forName(Ljava/lang/String;ZLjava/lang/ClassLoader;)Ljava/lang/Class; (53 bytes) @ 0x00007f5df1e476c8 [0x00007f5df1e476a0+0x28]
j org.bytedeco.javacpp.Loader.load(Ljava/lang/Class;Ljava/util/Properties;ZLjava/lang/String;)Ljava/lang/String;+286
j org.bytedeco.javacpp.Loader.load(Ljava/lang/Class;Ljava/util/Properties;Z)Ljava/lang/String;+4
j org.bytedeco.javacpp.Loader.load(Ljava/lang/Class;)Ljava/lang/String;+7
j org.tensorflow.NativeLibrary.load()V+2
j org.tensorflow.TensorFlow.<clinit>()V+0
v ~StubRoutines::call_stub
J 1053 java.lang.Class.forName0(Ljava/lang/String;ZLjava/lang/ClassLoader;Ljava/lang/Class;)Ljava/lang/Class; (0 bytes) @ 0x00007f5df13cc2ba [0x00007f5df13cc240+0x7a]
J 2238 C1 java.lang.Class.forName(Ljava/lang/String;)Ljava/lang/Class; (15 bytes) @ 0x00007f5df172774c [0x00007f5df17275a0+0x1ac]
j org.tensorflow.Graph.<clinit>()V+19
v ~StubRoutines::call_stub
j com.johnsnowlabs.ml.tensorflow.TensorflowWrapper$.readGraph(Ljava/lang/String;)Lorg/tensorflow/Graph;+12
j com.johnsnowlabs.ml.tensorflow.TensorflowWrapper$.unpackWithoutBundle(Ljava/lang/String;)Lscala/Tuple4;+19
j com.johnsnowlabs.ml.tensorflow.TensorflowWrapper$.read(Ljava/lang/String;ZZ[Ljava/lang/String;ZLscala/Option;)Lscala/Tuple2;+329
j com.johnsnowlabs.ml.tensorflow.ReadTensorflowModel.readTensorflowModel(Ljava/lang/String;Lorg/apache/spark/sql/SparkSession;Ljava/lang/String;ZZ[Ljava/lang/String;ZLscala/Option;)Lcom/johnsnowlabs/ml/tensorflow/TensorflowWrapper;+197
j com.johnsnowlabs.ml.tensorflow.ReadTensorflowModel.readTensorflowModel$(Lcom/johnsnowlabs/ml/tensorflow/ReadTensorflowModel;Ljava/lang/String;Lorg/apache/spark/sql/SparkSession;Ljava/lang/String;ZZ[Ljava/lang/String;ZLscala/Option;)Lcom/johnsnowlabs/ml/tensorflow/TensorflowWrapper;+14
j com.johnsnowlabs.nlp.embeddings.BertEmbeddings$.readTensorflowModel(Ljava/lang/String;Lorg/apache/spark/sql/SparkSession;Ljava/lang/String;ZZ[Ljava/lang/String;ZLscala/Option;)Lcom/johnsnowlabs/ml/tensorflow/TensorflowWrapper;+14
j com.johnsnowlabs.nlp.embeddings.ReadBertDLModel.readModel(Lcom/johnsnowlabs/nlp/embeddings/BertEmbeddings;Ljava/lang/String;Lorg/apache/spark/sql/SparkSession;)V+99
j com.johnsnowlabs.nlp.embeddings.ReadBertDLModel.readModel$(Lcom/johnsnowlabs/nlp/embeddings/ReadBertDLModel;Lcom/johnsnowlabs/nlp/embeddings/BertEmbeddings;Ljava/lang/String;Lorg/apache/spark/sql/SparkSession;)V+4
j com.johnsnowlabs.nlp.embeddings.BertEmbeddings$.readModel(Lcom/johnsnowlabs/nlp/embeddings/BertEmbeddings;Ljava/lang/String;Lorg/apache/spark/sql/SparkSession;)V+4
j com.johnsnowlabs.nlp.embeddings.ReadBertDLModel.$anonfun$$init$$1(Lcom/johnsnowlabs/nlp/embeddings/ReadBertDLModel;Lcom/johnsnowlabs/nlp/embeddings/BertEmbeddings;Ljava/lang/String;Lorg/apache/spark/sql/SparkSession;)V+4
j com.johnsnowlabs.nlp.embeddings.ReadBertDLModel.$anonfun$$init$$1$adapted(Lcom/johnsnowlabs/nlp/embeddings/ReadBertDLModel;Lcom/johnsnowlabs/nlp/embeddings/BertEmbeddings;Ljava/lang/String;Lorg/apache/spark/sql/SparkSession;)Ljava/lang/Object;+4
j com.johnsnowlabs.nlp.embeddings.ReadBertDLModel$$Lambda$276.apply(Ljava/lang/Object;Ljava/lang/Object;Ljava/lang/Object;)Ljava/lang/Object;+16
j com.johnsnowlabs.nlp.ParamsAndFeaturesReadable.$anonfun$onRead$1(Lcom/johnsnowlabs/nlp/HasFeatures;Ljava/lang/String;Lorg/apache/spark/sql/SparkSession;Lscala/Function3;)V+4
j com.johnsnowlabs.nlp.ParamsAndFeaturesReadable.$anonfun$onRead$1$adapted(Lcom/johnsnowlabs/nlp/HasFeatures;Ljava/lang/String;Lorg/apache/spark/sql/SparkSession;Lscala/Function3;)Ljava/lang/Object;+4
j com.johnsnowlabs.nlp.ParamsAndFeaturesReadable$$Lambda$2416.apply(Ljava/lang/Object;)Ljava/lang/Object;+16
J 8508 C2 scala.collection.mutable.ResizableArray.foreach(Lscala/Function1;)V (37 bytes) @ 0x00007f5df1f72a78 [0x00007f5df1f729e0+0x98]
J 6935 C2 scala.collection.mutable.ArrayBuffer.foreach(Lscala/Function1;)V (6 bytes) @ 0x00007f5df1ea3794 [0x00007f5df1ea3760+0x34]
j com.johnsnowlabs.nlp.ParamsAndFeaturesReadable.onRead(Lcom/johnsnowlabs/nlp/HasFeatures;Ljava/lang/String;Lorg/apache/spark/sql/SparkSession;)V+14
j com.johnsnowlabs.nlp.ParamsAndFeaturesReadable.$anonfun$read$1(Lcom/johnsnowlabs/nlp/ParamsAndFeaturesReadable;Lcom/johnsnowlabs/nlp/HasFeatures;Ljava/lang/String;Lorg/apache/spark/sql/SparkSession;)V+4
j com.johnsnowlabs.nlp.ParamsAndFeaturesReadable.$anonfun$read$1$adapted(Lcom/johnsnowlabs/nlp/ParamsAndFeaturesReadable;Lcom/johnsnowlabs/nlp/HasFeatures;Ljava/lang/String;Lorg/apache/spark/sql/SparkSession;)Ljava/lang/Object;+4
j com.johnsnowlabs.nlp.ParamsAndFeaturesReadable$$Lambda$1633.apply(Ljava/lang/Object;Ljava/lang/Object;Ljava/lang/Object;)Ljava/lang/Object;+16
j com.johnsnowlabs.nlp.FeaturesReader.load(Ljava/lang/String;)Lcom/johnsnowlabs/nlp/HasFeatures;+38
j com.johnsnowlabs.nlp.FeaturesReader.load(Ljava/lang/String;)Ljava/lang/Object;+2
j com.johnsnowlabs.nlp.pretrained.ResourceDownloader$.downloadModel(Lorg/apache/spark/ml/util/DefaultParamsReadable;Lcom/johnsnowlabs/nlp/pretrained/ResourceRequest;)Lorg/apache/spark/ml/PipelineStage;+26
j com.johnsnowlabs.nlp.pretrained.ResourceDownloader$.downloadModel(Lorg/apache/spark/ml/util/DefaultParamsReadable;Ljava/lang/String;Lscala/Option;Ljava/lang/String;)Lorg/apache/spark/ml/PipelineStage;+25
j com.johnsnowlabs.nlp.HasPretrained.pretrained(Ljava/lang/String;Ljava/lang/String;Ljava/lang/String;)Lorg/apache/spark/ml/PipelineStage;+45
j com.johnsnowlabs.nlp.HasPretrained.pretrained$(Lcom/johnsnowlabs/nlp/HasPretrained;Ljava/lang/String;Ljava/lang/String;Ljava/lang/String;)Lorg/apache/spark/ml/PipelineStage;+4
j com.johnsnowlabs.nlp.embeddings.BertEmbeddings$.com$johnsnowlabs$nlp$embeddings$ReadablePretrainedBertModel$$super$pretrained(Ljava/lang/String;Ljava/lang/String;Ljava/lang/String;)Lcom/johnsnowlabs/nlp/embeddings/BertEmbeddings;+4
j com.johnsnowlabs.nlp.embeddings.ReadablePretrainedBertModel.pretrained(Ljava/lang/String;Ljava/lang/String;Ljava/lang/String;)Lcom/johnsnowlabs/nlp/embeddings/BertEmbeddings;+4
j com.johnsnowlabs.nlp.embeddings.ReadablePretrainedBertModel.pretrained$(Lcom/johnsnowlabs/nlp/embeddings/ReadablePretrainedBertModel;Ljava/lang/String;Ljava/lang/String;Ljava/lang/String;)Lcom/johnsnowlabs/nlp/embeddings/BertEmbeddings;+4
j com.johnsnowlabs.nlp.embeddings.BertEmbeddings$.pretrained(Ljava/lang/String;Ljava/lang/String;Ljava/lang/String;)Lcom/johnsnowlabs/nlp/embeddings/BertEmbeddings;+4
j com.johnsnowlabs.nlp.embeddings.BertEmbeddings$.pretrained(Ljava/lang/String;Ljava/lang/String;Ljava/lang/String;)Lorg/apache/spark/ml/PipelineStage;+4
j com.johnsnowlabs.nlp.HasPretrained.pretrained(Ljava/lang/String;Ljava/lang/String;)Lorg/apache/spark/ml/PipelineStage;+9
j com.johnsnowlabs.nlp.HasPretrained.pretrained$(Lcom/johnsnowlabs/nlp/HasPretrained;Ljava/lang/String;Ljava/lang/String;)Lorg/apache/spark/ml/PipelineStage;+3
j com.johnsnowlabs.nlp.embeddings.BertEmbeddings$.com$johnsnowlabs$nlp$embeddings$ReadablePretrainedBertModel$$super$pretrained(Ljava/lang/String;Ljava/lang/String;)Lcom/johnsnowlabs/nlp/embeddings/BertEmbeddings;+3
j com.johnsnowlabs.nlp.embeddings.ReadablePretrainedBertModel.pretrained(Ljava/lang/String;Ljava/lang/String;)Lcom/johnsnowlabs/nlp/embeddings/BertEmbeddings;+3
j com.johnsnowlabs.nlp.embeddings.ReadablePretrainedBertModel.pretrained$(Lcom/johnsnowlabs/nlp/embeddings/ReadablePretrainedBertModel;Ljava/lang/String;Ljava/lang/String;)Lcom/johnsnowlabs/nlp/embeddings/BertEmbeddings;+3
j com.johnsnowlabs.nlp.embeddings.BertEmbeddings$.pretrained(Ljava/lang/String;Ljava/lang/String;)Lcom/johnsnowlabs/nlp/embeddings/BertEmbeddings;+3
j com.algo.recom.article_recommender.v20240511.test_spark_nlp$.main([Ljava/lang/String;)V+7
j com.algo.recom.article_recommender.v20240511.test_spark_nlp.main([Ljava/lang/String;)V+4
v ~StubRoutines::call_stub
j sun.reflect.NativeMethodAccessorImpl.invoke0(Ljava/lang/reflect/Method;Ljava/lang/Object;[Ljava/lang/Object;)Ljava/lang/Object;+0
j sun.reflect.NativeMethodAccessorImpl.invoke(Ljava/lang/Object;[Ljava/lang/Object;)Ljava/lang/Object;+100
J 2140 C1 sun.reflect.DelegatingMethodAccessorImpl.invoke(Ljava/lang/Object;[Ljava/lang/Object;)Ljava/lang/Object; (10 bytes) @ 0x00007f5df16bb94c [0x00007f5df16bb840+0x10c]
J 2139 C1 java.lang.reflect.Method.invoke(Ljava/lang/Object;[Ljava/lang/Object;)Ljava/lang/Object; (62 bytes) @ 0x00007f5df16bb35c [0x00007f5df16baf60+0x3fc]
j org.apache.spark.deploy.JavaMainApplication.start([Ljava/lang/String;Lorg/apache/spark/SparkConf;)V+97
j org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(Lorg/apache/spark/deploy/SparkSubmitArguments;Z)V+449
j org.apache.spark.deploy.SparkSubmit.doRunMain$1(Lorg/apache/spark/deploy/SparkSubmitArguments;Z)V+111
j org.apache.spark.deploy.SparkSubmit.submit(Lorg/apache/spark/deploy/SparkSubmitArguments;Z)V+65
j org.apache.spark.deploy.SparkSubmit.doSubmit([Ljava/lang/String;)V+78
j org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit([Ljava/lang/String;)V+2
j org.apache.spark.deploy.SparkSubmit$.main([Ljava/lang/String;)V+10
j org.apache.spark.deploy.SparkSubmit.main([Ljava/lang/String;)V+4
v ~StubRoutines::call_stub
Great, there is something wrong with the model. suggest you either found another model with Chinese support, try to import the same model yourself with ONNX, or import another model yourself:
Import new model(s):
Is there an existing issue for this?
I find it here(I'm not sure it's the same error) : https://github.com/JohnSnowLabs/spark-nlp/issues/5918
Current Behavior
I got error below:
hs_err_pid23030.log
Steps To Reproduce
I load model from https://sparknlp.org/2022/04/11/bert_embeddings_chinese_roberta_wwm_ext_zh_3_0.html and put it onto hdfs_path. But when I load it, the error comes.
spark-submit :
Spark NLP version and Apache Spark
spark version 3.3.2 Scala version 2.12.15 Java HotSpot(TM) 64-Bit Server VM, 1.8.0_144