问题描述
我正在使用 AWS glue 运行一些 pyspark python 代码,有时成功但有时失败并出现依赖错误:Resource Setup Error: Exception in thread "main" java.lang.RuntimeException: [unresolved dependency: JohnSNowLabs#spark-nlp;2.5.4: not found]
,这是错误日志:
:: problems summary ::
:::: WARNINGS
module not found: JohnSNowLabs#spark-nlp;2.5.4
==== local-m2-cache: tried
file:/root/.m2/repository/JohnSNowLabs/spark-nlp/2.5.4/spark-nlp-2.5.4.pom
-- artifact JohnSNowLabs#spark-nlp;2.5.4!spark-nlp.jar:
file:/root/.m2/repository/JohnSNowLabs/spark-nlp/2.5.4/spark-nlp-2.5.4.jar
==== local-ivy-cache: tried
/root/.ivy2/local/JohnSNowLabs/spark-nlp/2.5.4/ivys/ivy.xml
-- artifact JohnSNowLabs#spark-nlp;2.5.4!spark-nlp.jar:
/root/.ivy2/local/JohnSNowLabs/spark-nlp/2.5.4/jars/spark-nlp.jar
==== central: tried
https://repo1.maven.org/maven2/JohnSNowLabs/spark-nlp/2.5.4/spark-nlp-2.5.4.pom
-- artifact JohnSNowLabs#spark-nlp;2.5.4!spark-nlp.jar:
https://repo1.maven.org/maven2/JohnSNowLabs/spark-nlp/2.5.4/spark-nlp-2.5.4.jar
==== spark-packages: tried
https://dl.bintray.com/spark-packages/maven/JohnSNowLabs/spark-nlp/2.5.4/spark-nlp-2.5.4.pom
-- artifact JohnSNowLabs#spark-nlp;2.5.4!spark-nlp.jar:
https://dl.bintray.com/spark-packages/maven/JohnSNowLabs/spark-nlp/2.5.4/spark-nlp-2.5.4.jar
::::::::::::::::::::::::::::::::::::::::::::::
:: UNRESOLVED DEPENDENCIES ::
::::::::::::::::::::::::::::::::::::::::::::::
:: JohnSNowLabs#spark-nlp;2.5.4: not found
::::::::::::::::::::::::::::::::::::::::::::::
:: USE VERBOSE OR DEBUG MESSAGE LEVEL FOR MORE DETAILS
Exception in thread "main" java.lang.RuntimeException: [unresolved dependency: JohnSNowLabs#spark-nlp;2.5.4: not found]
从成功运行的日志中,我可以看到胶水能够从 https://dl.bintray.com/spark-packages/maven/JohnSNowLabs/spark-nlp/2.5.4/spark-nlp-2.5.4.pom
下载依赖项,失败的作业也尝试从那里下载,但失败了。
这个问题似乎在上周自行解决了,但在最近几天它又出现了,到目前为止还没有自行解决。有没有人见过这个奇怪的问题?谢谢。
解决方法
spark-packages 于 2021 年 5 月 1 日移动。在我的 Scala 项目中,我不得不像这样添加一个不同的解析器。在java中它必须是相似的。
resolvers in ThisBuild ++= Seq(
"SparkPackages" at "https://repos.spark-packages.org"
## remove -> "MVNRepository" at "https://dl.bintray.com/spark-packages/maven"
)
去看看你自己,那个包不在你要找的那个解析器上。我的也不是。