The 0.16.0 version fails with "Could not initialize class com.google.cloud.spark.bigquery.SparkBigQueryConnectorUserAgentProvider"
See original GitHub issueThe current latest version fails to fetch data from BigQuery with
java.lang.NoClassDefFoundError: Could not initialize class com.google.cloud.spark.bigquery.SparkBigQueryConnectorUserAgentProvider
It can get basic metadata, so the data frames get created, but once we need to materialize or compute it fails.
The exact fat jar used: https://storage.googleapis.com/spark-lib/bigquery/spark-bigquery-latest_2.11.jar and it passed to the spark-submit with the --jars
The code is a simple spark load:
spark.read.format("bigquery").option("table", "project.dataset.tablename").load()
Issue Analytics
- State:
- Created 3 years ago
- Comments:6 (2 by maintainers)
Top Results From Across the Web
Could not initialize class com.google.cloud.spark.bigquery ...
fit(), it throws the below error. from pyspark.sql import SparkSession from pyspark.sql.functions import col, split from pyspark ...
Read more >Use the BigQuery connector with Spark - Google Cloud
Install the spark-bigquery-connector in the Spark jars directory of every node by using the Dataproc connectors initialization action when you create your ...
Read more >com.google.cloud.spark spark-bigquery_2.11 - Javadoc.io
SparkBigQueryUtil · SparkFilterUtils · SupportedCustomDataType · ToIterator. focushidecom.google.cloud.spark.bigquery.direct.
Read more >pandas-gbq Documentation - Read the Docs
You can install pandas-gbq with conda, pip, or by installing from source ... Note: The dependency on google-cloud-bigquery is new in version ......
Read more >spark-bigquery-connector - Scaladex
Create a Google Cloud Dataproc cluster (Optional). If you do not have an Apache Spark environment you can create a Cloud Dataproc cluster...
Read more >
Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free
Top Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
You can list all available Spark BQ connector jars using
gsutil
:You can use any of these jars.
instead of -latest.jar can you provide gs bucket path with specific version?