data_phase2 ray actor dies
See original GitHub issueHey Dian,
Trying to run data_phase2
and I get the following Ray error (seems to have issue with RemoteMainDataset
constructor?). I did some debugging by replacing all the @ray.remote
stuff and .remote()
commands with the non-ray versions, and the code runs with no issue (although the progress bar didn’t progress past 0 frames after a minute or two, not quite sure if it’s supposed to take that long or not).
Did you ever see anything like this/know what I should do?
(wor) aaron@Aarons-Machine:~/workspace/carla/WorldOnRails$ RAY_PDB=1 python -m rails.data_phase2 --num-workers=12
Traceback (most recent call last):
File "/home/aaron/anaconda3/envs/wor/lib/python3.7/runpy.py", line 193, in _run_module_as_main
2021-05-29 14:45:49,862 WARNING worker.py:1034 -- Traceback (most recent call last):
File "/home/aaron/anaconda3/envs/wor/lib/python3.7/site-packages/ray/function_manager.py", line 251, in get_execution_info
info = self._function_execution_info[job_id][function_id]
KeyError: FunctionID(41f68a98bcf1c9ebc84e01b0819040089631493c)
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "python/ray/_raylet.pyx", line 550, in ray._raylet.task_execution_handler
File "python/ray/_raylet.pyx", line 364, in ray._raylet.execute_task
File "/home/aaron/anaconda3/envs/wor/lib/python3.7/site-packages/ray/function_manager.py", line 256, in get_execution_info
raise KeyError(message)
KeyError: 'Error occurs in get_execution_info: job_id: JobID(01000000), function_descriptor: {type=PythonFunctionDescriptor, module_name=rails.datasets.main_dataset, class_name=RemoteMainDataset, function_name=__init__, function_hash=084f10af-7af1-46d7-8dda-ada171c2aad9}. Message: FunctionID(41f68a98bcf1c9ebc84e01b0819040089631493c)'
An unexpected internal error occurred while the worker was executing a task.
"__main__", mod_spec)
File "/home/aaron/anaconda3/envs/wor/lib/python3.7/runpy.py", line 85, in _run_code
2021-05-29 14:45:49,862 WARNING worker.py:1034 -- A worker died or was killed while executing task ffffffffffffffffcb230a5701000000.
exec(code, run_globals)
File "/home/aaron/workspace/carla/WorldOnRails/rails/data_phase2.py", line 67, in <module>
main(args)
File "/home/aaron/workspace/carla/WorldOnRails/rails/data_phase2.py", line 13, in main
total_frames = ray.get(dataset.num_frames.remote())
File "/home/aaron/anaconda3/envs/wor/lib/python3.7/site-packages/ray/worker.py", line 1381, in get
raise value
ray.exceptions.RayActorError: The actor died unexpectedly before finishing this task.
(wor) aaron@Aarons-Machine:~/workspace/carla/WorldOnRails$
Issue Analytics
- State:
- Created 2 years ago
- Comments:14 (8 by maintainers)
Top Results From Across the Web
Ray Liotta, 'Goodfellas' and 'Field of Dreams' star, dies at 67
Ray Liotta, the actor best known for portraying mobster Henry Hill "Goodfellas" died at 67. ... 'Goodfellas' star Ray Liotta dies at 67....
Read more >Ray Liotta, star of 'Good Fellas' and 'Field of Dreams,' dies at 67
Actor Ray Liotta in starred in the mob drama “Goodfellas” and in the ... “Field of Dreams,” was found dead May 26 in...
Read more >Ray Liotta Dead: 'Goodfellas' Star & 'Field Of Dreams' Actor ...
Ray Liotta has died. The 'Goodfellas' star and 'Field Of Dreams' actor was 67.
Read more >Ray Liotta Dead: 'Goodfellas' Actor Dies at Age 67 - Us Weekly
Ray Liotta earned his breakout role in 'Goodfellas' and recently starred on the series 'Shades of Blue' — read more.
Read more >'Goodfellas' star Ray Liotta dead at 67 - NPR
Actor Ray Liotta has died at 67. He starred in 'Goodfellas' According to his publicist, the actor died in his sleep in the...
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
That worked! Thanks a ton 😃
@dotchen messed around with it today and got some strange behavior. I’ve described them briefly below. I’m running
RAY_PDB=1 python -m rails.data_phase2 --num-workers=1
(btw theRAILS.md
tells users to use a--num-runners
argument rather than the correct--num-workers
argument for this phase)1. with ray
local_mode=False
The actor dies as described in the original post.2. with ray
local_mode=True
Produces the following error3. with ray
local_mode=True
The only thing that changed here is I printed outtorch.cuda.is_available()
right at the beginning ofrails.data_phase2
’s__main__
function (obviously to debug the above). For some reason, this makes it work and I successfully ran the script on a toy dataset of like 1000 frames in 2-3 minutes. See here - https://wandb.ai/aaronhuang/carla_data_phase2/runs/5flpwvwk?workspace=user-aaronhuang