Closed tuoluzhe8521 closed 5 months ago
Is there any impact on usage?
Is there any impact on usage? when i submit sparksql task,it always give error,like: :: problems summary :: :::: WARNINGS ::::::::::::::::::::::::::::::::::::::::::::::
:: UNRESOLVED DEPENDENCIES ::
::::::::::::::::::::::::::::::::::::::::::::::
:: org.apache.doris#spark-doris-connector-3.2_2.12;1.3.0: java.text.ParseException: inconsistent module descriptor file found in 'https://repo1.maven.org/maven2/org/apache/doris/spark-doris-connector-3.2_2.12/1.3.0/spark-doris-connector-3.2_2.12-1.3.0.pom': bad revision: expected='1.3.0' found='1.3.0-SNAPSHOT';
::::::::::::::::::::::::::::::::::::::::::::::
:::: ERRORS central: bad revision found in https://repo1.maven.org/maven2/org/apache/doris/spark-doris-connector-3.2_2.12/1.3.0/spark-doris-connector-3.2_2.12-1.3.0.pom: expected='1.3.0 found='1.3.0-SNAPSHOT'
:: USE VERBOSE OR DEBUG MESSAGE LEVEL FOR MORE DETAILS Exception in thread "main" java.lang.RuntimeException: [unresolved dependency: org.apache.doris#spark-doris-connector-3.2_2.12;1.3.0: java.text.ParseException: inconsistent module descriptor file found in 'https://repo1.maven.org/maven2/org/apache/doris/spark-doris-connector-3.2_2.12/1.3.0/spark-doris-connector-3.2_2.12-1.3.0.pom': bad revision: expected='1.3.0' found='1.3.0-SNAPSHOT'; ] at org.apache.spark.deploy.SparkSubmitUtils$.resolveMavenCoordinates(SparkSubmit.scala:1458) at org.apache.spark.util.DependencyUtils$.resolveMavenDependencies(DependencyUtils.scala:185) at org.apache.spark.deploy.SparkSubmit.prepareSubmitEnvironment(SparkSubmit.scala:319) at org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:909) at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:191) at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:214) at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:90) at org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:1054) at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:1063) at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
Is there any impact on usage? when i submit sparksql task,it always give error,like: :: problems summary :: :::: WARNINGS ::::::::::::::::::::::::::::::::::::::::::::::
:: UNRESOLVED DEPENDENCIES :: :::::::::::::::::::::::::::::::::::::::::::::: :: org.apache.doris#spark-doris-connector-3.2_2.12;1.3.0: java.text.ParseException: inconsistent module descriptor file found in 'https://repo1.maven.org/maven2/org/apache/doris/spark-doris-connector-3.2_2.12/1.3.0/spark-doris-connector-3.2_2.12-1.3.0.pom': bad revision: expected='1.3.0' found='1.3.0-SNAPSHOT'; ::::::::::::::::::::::::::::::::::::::::::::::
:::: ERRORS central: bad revision found in https://repo1.maven.org/maven2/org/apache/doris/spark-doris-connector-3.2_2.12/1.3.0/spark-doris-connector-3.2_2.12-1.3.0.pom: expected='1.3.0 found='1.3.0-SNAPSHOT'
:: USE VERBOSE OR DEBUG MESSAGE LEVEL FOR MORE DETAILS Exception in thread "main" java.lang.RuntimeException: [unresolved dependency: org.apache.doris#spark-doris-connector-3.2_2.12;1.3.0: java.text.ParseException: inconsistent module descriptor file found in 'https://repo1.maven.org/maven2/org/apache/doris/spark-doris-connector-3.2_2.12/1.3.0/spark-doris-connector-3.2_2.12-1.3.0.pom': bad revision: expected='1.3.0' found='1.3.0-SNAPSHOT'; ] at org.apache.spark.deploy.SparkSubmitUtils$.resolveMavenCoordinates(SparkSubmit.scala:1458) at org.apache.spark.util.DependencyUtils$.resolveMavenDependencies(DependencyUtils.scala:185) at org.apache.spark.deploy.SparkSubmit.prepareSubmitEnvironment(SparkSubmit.scala:319) at org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:909) at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:191) at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:214) at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:90) at org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:1054) at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:1063) at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
i submit sparksql task in k8s
1.3.1 fixed this problem, can you help test it?
i meet this problem when i submit sparksql task in k8s,because i can't sovle it. I give up k8s, so i can't test it now
Search before asking
Version
v1.3.0和v1.2.0都有这个问题
What's Wrong?
中央仓库的pom文件没有更新