Error creating DStream on Azure Databricks
See original GitHub issueHi,
Following this example I’m trying to create EventHub DStream
but running into the exception below. Here is my code;
val ssc = new StreamingContext(sc, Seconds(batchIntervalSeconds))
ssc.checkpoint(checkpointDir)
val inputDStream = EventHubsUtils.createDirectStreams(
ssc,
eventhubsNamespace,
progressDir,
Map(eventhubsName -> eventhubsParameters))
- Actual behavior
java.lang.NoSuchMethodError: org.apache.spark.streaming.eventhubs.EventHubDirectDStream$.$lessinit$greater$default$5()Lscala/Function5;
at org.apache.spark.streaming.eventhubs.EventHubsUtils$.createDirectStreams(EventHubsUtils.scala:52)
-
Expected behavior
DStream
is successfully created -
Spark version Databricks Runtime 3.5 Apache Spark 2.2.1 Scala 2.11
-
spark-eventhubs artifactId and version com.microsoft.azure:azure-eventhubs-spark_2.11:jar:2.1.6
-
Dependencies automatically pulled by Databricks proton-j-0.22.0.jar slf4j-api-1.7.25.jar
Issue Analytics
- State:
- Created 6 years ago
- Comments:5 (3 by maintainers)
Top Results From Across the Web
Apache Spark DStream is not supported - Databricks
Problem. You are attempting to use a Spark Discretized Stream (DStream) in a Databricks streaming job, but the job is failing.
Read more >Apache Spark DStream is not supported - Azure Databricks
Problem. You are attempting to use a Spark Discretized Stream (DStream) in a Azure Databricks streaming job, but the job is failing. Cause....
Read more >Azure databricks: KafkaUtils createDirectStream causes ...
The problem arises only, if i try to create Kafka direct stream using KafkaUtils.createDirectStream() in azure databricks python notebook.
Read more >Stream processing with Azure Databricks - GitHub
Stream processing with Azure Databricks. Contribute to mspnp/azure-databricks-streaming-analytics development by creating an account on GitHub.
Read more >Build Streaming Data Pipelines with Confluent, Databricks ...
Leveraging Confluent Cloud and Azure Databricks as fully managed services ... Step 1: Create a Kafka cluster; Step 2: Enable Schema Registry ...
Read more >
Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free
Top Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Ah yes, you’re right. Glad it’s working! I’ll be updating documentation over the next couple weeks so this process is a little clearer.
Hi @sabeegrewal I’ve peeked into the eventhubs-databricks.jar and the correct namespace there is
org.apache.spark.eventhubs.common.EventHubsUtils
. Thanks for the help!