Not enough memory error when parsing pbf dataset
See original GitHub issueHello.
I was trying to parse a pbf file (approx. 65 mb) and I get an “out of memorry” error. I also have noticed that others have this issue as well. My machine specs should be sufficient for parsing this data (16gb of ram, i7 CPU),… I suppose.
I tried to investigate why this memory error occurs on small datasets and in my case it looks like it comes from node elements that are returned from the parse_osm_data
(cython method) in pyrosm.py. Each node’s property that is returned is an array of size x (in my case 8000). I couldn’t manage to debug Cython code so I don’t quite understand why this behavior happens. Is there any reason why all these elements are returned only for nodes. ?
Thank you.
Issue Analytics
- State:
- Created 2 years ago
- Reactions:1
- Comments:10 (6 by maintainers)
Top Results From Across the Web
Fix numpy deprecation warning · Issue #50 · HTenkanen/pyrosm
Not enough memory error when parsing pbf dataset #150. Open. Sign up for free to join this conversation on GitHub. Already have an...
Read more >Not Enough Memory Error - Microsoft Power BI Community
Solved: Hi, I am using Power BI Desktop 64 Bit connecting to an Analysis Services Multi Dimensional Model via a Power Query query...
Read more >Error when importing merged osm.pbf files into nominatim's ...
I already imported succesfully 2 regions (not together, my PostgreSQL database contained one region at a time).
Read more >Benchmarks - Pyrosm - Read the Docs
Pyrosm aims to be an easy-to-use and fast Python tool for parsing OpenStreetMap data from Protocolbuffer Binary Format (PBF) files into geopandas which...
Read more >Osm2pgsql Manual
As a rule of thumb you need at least as much main memory as the PBF file with the ... Osm2pgsql does not...
Read more >
Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free
Top Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Can confirm this issue, i tried also to read the
mittelfranken
andniederbayern
pbf files, in both cases the Kernel just died without any error message.Hi @HTenkanen I think this issue has more to it than just the Memory issues. I am currently using this library in a GCP VM where i am facing ipython notebook kernel dying for the smallest of PBF files(Comores 3.3MB) even after raising the RAM size to 500 GB. As per your suggestion i used the 0.5.3 version of the library. However i can get results if i call certain functions like get_network() but when i try to do a custom filter it kills the kernel.