question-mark
Stuck on an issue?

Lightrun Answers was designed to reduce the constant googling that comes with debugging 3rd party libraries. It collects links to all the places you might be looking at while hunting down a tough bug.

And, if you’re still stuck at the end, we’re happy to hop on a call to see how we can help out.

Error using some pretrained pipelines in Spark/PySpark 3.x

See original GitHub issue

Description

When I try to use match_chunks and match_datetime pretrained pipelines in sparknlp_version 3.0.1, get errors while downloading these pipelines.

Current Behavior

1 2

Steps to Reproduce

  1. pipeline = PretrainedPipeline('match_chunks', lang='en')
  2. pipeline = PretrainedPipeline('match_datetime', lang='en')

Context

Your Environment

  • Spark NLP version sparknlp.version(): 3.0.1
  • Apache NLP version spark.version: 3.1.1
  • Java version java -version: openjdk version “1.8.0_282”

Issue Analytics

  • State:open
  • Created 2 years ago
  • Reactions:1
  • Comments:7 (6 by maintainers)

github_iconTop GitHub Comments

1reaction
Digaaricommented, Apr 12, 2021

Faced similar issue with check_spelling_dl in the same environment. image

1reaction
maziyarpanahicommented, Apr 9, 2021

@Digaari Only report anything related to public and open-source. We are not responsible for anything else.

Read more comments on GitHub >

github_iconTop Results From Across the Web

Introduction to Spark NLP: Foundations and Basic Components
This is the first article in a series of blog posts to help Data Scientists and NLP practitioners learn the basics of Spark...
Read more >
spark-nlp : DocumentAssembler initializing failing with 'java ...
your Apache Spark/PySpark in order to use any release of Spark NLP above 3.0.x. The key here is to find the correct Maven...
Read more >
Installation - Spark NLP - John Snow Labs
Spark NLP supports Python 3.6.x and above depending on your major ... and sentiment analysis by using Spark NLP pretrained pipelines.
Read more >
spark-nlp - PyPI
Spark NLP comes with 11000+ pretrained pipelines and models in more than 200+ ... Configure Zeppelin properly, use cells with %spark.pyspark or any...
Read more >
Configure Dataproc Python environment - Google Cloud
PySpark jobs on Dataproc are run by a Python interpreter on the cluster. Job code must be compatible at runtime with the Python...
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found