question-mark
Stuck on an issue?

Lightrun Answers was designed to reduce the constant googling that comes with debugging 3rd party libraries. It collects links to all the places you might be looking at while hunting down a tough bug.

And, if you’re still stuck at the end, we’re happy to hop on a call to see how we can help out.

Inference: own data

See original GitHub issue

Hi,

I just had a few questions regarding using our own data and running inference using PENet pretrained weights.

  1. How sparse can the depth map be? Currently, my inference image is from the Kitti360 dataset which is quite similar to the previous kitti that the network was trained on. But there is no GT depth to sample the depth from. So my sparse depth map is quite sparse. When I run inference on this image, the prediction is also sparse i.e I have prediction only in the regions covered by the sparse depth map. Is this an expected behaviour?

  2. What should my input be for ‘positions’ (i.e the cropped image), I don’t want to crop the images for running inference, so should I just set input['positions'] = input['rgb']?

It would be great if you can answer these questions when time permits 😃

Regards, Shrisha

Issue Analytics

  • State:open
  • Created 2 years ago
  • Comments:9 (5 by maintainers)

github_iconTop GitHub Comments

1reaction
JUGGHMcommented, Mar 4, 2022

Okay I understand regarding the sparsity for training.

But I am still unsure about what the positional encodings are because I am preparing my own data, so is there a flag for evaluation? I understand --co to std is for training right?

I can see in model.py that u & v coordinates of input[‘position’] are used, but I am still not sure what the positional encodings are or how they can be created for my own data.

You could refer to [An intriguing failing of convolutional neural networks and the CoordConv solution] by Liu for more details about positional encoding. In our default settings, we use the geometric encoding (ie. 3d coordianates) described in our paper. And the evaluation and training process should share consistent settings.

0reactions
JUGGHMcommented, Mar 5, 2022

But we dont need GT maps in this case where we just evaluate. The result that I’ve shared here is not based on fine-tuning but merely test_completion

I think two points could be taken into consideration:

  1. The sparse depth maps in KITTI360 seem denser than the ones we use in KITTI depth. This means that there exists domain gaps between those two datasets, leading to the failure of predicted results. We suggest that you could (i) Construct denser GT maps in KITTI360 for further training or finetuning. This step is necessary for transfer learning. Or (ii) Consider depth completion methods with “Sparsity Invariance”, which aims at countering the instability brought by unknown and varying density. You could refer to [Sparsity Invariant CNNs] by Uhrig or [A Normalized Convolutional Neural Network for Guided Sparse Depth Upsampling] from our group. Recently, this topic has been discussed in [Boosting Monocular Depth Estimation with Lightweight 3D Point Fusion] as well.

  2. A secondary reason is that: Not all pixels in the predicted depth map are reliable. You could refer to a previous issue for this.

Read more comments on GitHub >

github_iconTop Results From Across the Web

Use Your Own Inference Code - Amazon SageMaker
You can use Amazon SageMaker to interact with Docker containers and run your own inference code in one of two ways: To use...
Read more >
Inference on my own data? · Issue #1 · j-min/VL-T5
Would it be possible for you to provide an example code for extracting image features and text features for a custom dataset? (not...
Read more >
What is Machine Learning Inference?
Machine learning inference is the process of running live data into a machine learning algorithm to calculate output such as a single numerical...
Read more >
Understanding Machine Learning Inference
Machine learning inference—involves putting the model to work on live data to produce an actionable output. During this phase, the inference system accepts ......
Read more >
Dataset Inference: Ownership Resolution in Machine ...
We develop an approach for dataset inference that combines statistical testing with the ability to estimate the distance of multiple data points to...
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found