pyspark --packages com.microsoft.ml.spark:mmlspark_2.11:1.0.0-rc1
See original GitHub issueCommand
pyspark --packages com.microsoft.ml.spark:mmlspark_2.11:1.0.0-rc1
returns an error.
unresolved dependencies
com.microsoft.ml.spark:mmlspark_2.11;1.0.0-rc1 not found
I’m not sure if this is relevant but there’s a semicolon instead of a colon in that error. This is with Python 3.7.3, Apache Maven 3.6.1, jdk1.8.0_221, spark-2.4.3-bin-hadoop2.7
Issue Analytics
- State:
- Created 4 years ago
- Comments:7
Top Results From Across the Web
packages com.microsoft.ml.spark:mmlspark_2.11:1.0.0-rc1 #715
Command pyspark --packages com.microsoft.ml.spark:mmlspark_2.11:1.0.0-rc1 returns an error. unresolved dependencies ...
Read more >Microsoft Machine Learning for Apache Spark - NET
MMLSpark's API spans Scala, Python, Java, and R so you can integrate with any ecosystem. Try our PySpark Examples. Install. Spark Packages. Databricks....
Read more >how do I install parckage(such as mmlspark) to CDH cluster ...
pyspark --packages com.microsoft.ml.spark:mmlspark_2.11:1.0.0-rc1 ... I downloaded all packages and copy to local CDH cluster ...
Read more >PySpark 3.3.1 documentation - Apache Spark
PySpark supports most of Spark's features such as Spark SQL, DataFrame, Streaming, MLlib (Machine Learning) and Spark Core. PySpark Components. Spark SQL and ......
Read more >Spark Library Management - SQL Server Big Data Clusters
Scala Spark (Scala 2.12) and Hadoop base packages. PySpark (Python 3.8). Pandas, Sklearn, Numpy, and other data processing and machine learning ...
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
Hey @pairwiserr and @candalfigomoro sorry for this, looks like sbt thought it was a snapshot because of the -rc. You can get around this for now by using our maven repo:
by putting this in your build.sbt
resolvers += "MMLSpark Repo" at "https://mmlspark.azureedge.net/maven"
or adding that url to your
spark.jars.repositories https://mmlspark.azureedge.net/maven
to your spark settings