question-mark
Stuck on an issue?

Lightrun Answers was designed to reduce the constant googling that comes with debugging 3rd party libraries. It collects links to all the places you might be looking at while hunting down a tough bug.

And, if you’re still stuck at the end, we’re happy to hop on a call to see how we can help out.

Error using s3 protocol for marathon application resource: hadoop command not found

See original GitHub issue

When deploying a marathon application with a s3 uri the mesos task fails to staging with the following error in the sandbox:

E0605 14:18:44.685948 28351 shell.hpp:93] Command 'hadoop version 2>&1' failed; this is the output:
sh: hadoop: command not found
Failed to fetch 's3://bucket/object.tar.gz': Failed to create HDFS client: Failed to execute 'hadoop version 2>&1'; the command was either not found or exited with a non-zero exit status: 127
  • Ansible version (ansible --version): ansible 1.9.6
  • Python version (python --version): Python 2.7.10
  • Git commit hash or branch: 1.1
  • Cloud Environment: AWS
  • Terraform version (terraform version): Terraform v0.6.15

Issue Analytics

  • State:open
  • Created 7 years ago
  • Comments:6 (2 by maintainers)

github_iconTop GitHub Comments

2reactions
henningecommented, Oct 11, 2017

From the Mesos Fetcher documentation:

The Mesos fetcher can copy files from a local filesytem and it also natively supports the HTTP, HTTPS, FTP and FTPS protocols. If the requested URI is based on some other protocol, then the fetcher tries to utilise a local Hadoop client and hence supports any protocol supported by the Hadoop client, e.g., HDFS, S3. See the agent configuration documentation for how to configure the agent with a path to the Hadoop client.

1reaction
matthewdfullercommented, Sep 27, 2016

Also running into the same error. Task has nothing to do with Hadoop (it’s just a “hello world” container) and I’m getting the output:

I0927 16:01:57.900351 18469 logging.cpp:194] INFO level logging started!
I0927 16:01:57.900606 18469 fetcher.cpp:498] Fetcher Info: {"cache_directory":"\/tmp\/mesos\/fetch\/slaves\/0e5b7193-20c5-4335-a61b-6c7b6af6eb89-S4","items":[{"action":"BYPASS_CACHE","uri":{"cache":false,"executable":false,"extract":true,"value":"s3:\/\/my-custom-bucket\/docker.tar.gz"}}],"sandbox_directory":"\/var\/lib\/mesos\/slave\/slaves\/0e5b7193-20c5-4335-a61b-6c7b6af6eb89-S4\/frameworks\/0e5b7193-20c5-4335-a61b-6c7b6af6eb89-0000\/executors\/my-app.b7d51023-84cb-11e6-a58e-eeb994d4f386\/runs\/28aaaff1-7fdd-43a9-a708-1620932df51f"}
I0927 16:01:57.902078 18469 fetcher.cpp:409] Fetching URI 's3://my-custom-bucket/docker.tar.gz'
I0927 16:01:57.902096 18469 fetcher.cpp:250] Fetching directly into the sandbox directory
I0927 16:01:57.902112 18469 fetcher.cpp:187] Fetching URI 's3://my-custom-bucket/docker.tar.gz'
E0927 16:01:57.903787 18469 shell.hpp:106] Command 'hadoop version 2>&1' failed; this is the output:
sh: hadoop: command not found
Failed to fetch 's3://my-custom-bucket/docker.tar.gz': Failed to create HDFS client: Failed to execute 'hadoop version 2>&1'; the command was either not found or exited with a non-zero exit status: 127
Read more comments on GitHub >

github_iconTop Results From Across the Web

Troubleshooting - Apache Hadoop
hadoop.fs.s3a.S3AFileSystem. These are Hadoop filesystem client classes, found in the hadoop-aws JAR. An exception reporting this class as missing means ...
Read more >
Spark + s3 - error - java.lang.ClassNotFoundException ...
S3AFileSystem These are Hadoop filesystem client classes, found in the `hadoop-aws` JAR. An exception reporting this class as missing means that ...
Read more >
Resolve HTTP 403 "Access Denied" AmazonS3Exception ...
First, check the credentials or role specified in your application code. Run the following command on the EMR cluster's master node. Replace s3...
Read more >
Solved: I run a Hadoop job, but it got stucked and nothing...
Solved: I run this command - hadoop jar ... Job: Running job: job_1479831488614_0002 ... log4j:ERROR Could not find value for key log4j.appender.
Read more >
May 18, 2022•Knowledge 000153675 - Search
ClassNotFoundException:Class com.amazon.ws.emr.hadoop.fs.EmrFileSystem not found]" while running Amazon S3 mapping on Spark engine with EMR ...
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found