question-mark
Stuck on an issue?

Lightrun Answers was designed to reduce the constant googling that comes with debugging 3rd party libraries. It collects links to all the places you might be looking at while hunting down a tough bug.

And, if you’re still stuck at the end, we’re happy to hop on a call to see how we can help out.

Refinement stage query

See original GitHub issue

Hi, Awesome work! Thanks for sharing it. I had a couple of questions regarding the refinement stage:

  1. In the paper you mention that you use the output from the 1st fconvnet and feed it (after some adjustments) into a 2nd fconvnet. Do I understand this correctly? So if we were to use the refined network it would in fact be 2 fconvnets joined end-to-end?
  2. When training the refined network, does the model_best.pth obtained contain BOTH these fconvnets? or does it only contain the 2nd fconvnet that uses the results of the 1st one?
  3. In the case of training the model for car, pedestrian and cyclists, the number of classes for fconvnet would still remain 2 right (object, background)? Is the classification then taken from the 2D object detection model?

Cheers!

Issue Analytics

  • State:closed
  • Created 4 years ago
  • Comments:5

github_iconTop GitHub Comments

2reactions
zhixinwangcommented, Feb 18, 2020

All you said is right.

  1. Yes. We use prediction of the 1st f-convnet as the input of our 2nd f-convnet during evaluation. The command of python kitti/prepare_data_refine.py --car_only --gen_val_rgb_detection save the prediction of the first stage. The whole pipeline includes two f-convnet networks.
  2. Only contain parameters of 2nd f-convnet.
  3. Yes. We encode class prediction from 2D detection as one-hot label and use it as the intermediate features in our network. So, we only need to distinguish foreground and background. Actually, we found in our experiments the RGB features are more distinctive than point clouds.
1reaction
sharma-ncommented, May 13, 2020

From what i understand, FConvNet is responsible for the 3D box regression. The two classification classes of background/foreground allows us to get some kind of “confidence score” from FConvNet.

You could also potentially train 3 separate FConvNets for each class (slightly higher accuracy?). This would require you to pipe them accordingly during inference time based on the 2d detection. But you’ll be using 3 times the GPU memory to keep the 3 networks loaded.

And yes, to train the FConvNet you’ll need 2D detection boxes and classifications. Once you have that, the code provided by @zhixinwang automatically calculates the frustums. Also, @zhixinwang also already provides those rgb detections for the KITTI dataset (thanks!), stored in frustum-convnet/kitti/rgb_detections/rgb_detections_train.txt

Read more comments on GitHub >

github_iconTop Results From Across the Web

Query Refinement: What Is Query Refinement? - WordStream
Query refinement refers to the process of refining (changing or narrowing down) a search query. Search suggestions can reduce the need for query...
Read more >
Interactive Query Refinement - OpenProceedings.org
Shrink (SnS) framework for Interactive Query Refinement. SnS refines a query in two phases, with each phase utilizing the SSE procedure described in...
Read more >
Understanding Conversational Search Refinement Queries in ...
Understanding Conversational Search Refinement Queries in Walmart Shopping ... multi-stage conversation while refining the search results is ...
Read more >
Entity-Centric Query Refinement - AKBC 2022
We introduce the task of entity-centric query refinement. ... Table 2: Stage 2 evaluation assesses the overall quality of refinement sets. The notation....
Read more >
Two-step spatial query processing: filter and refinement step.
Support for nontraditional data, including spatial objects, in an efficient manner is of ongoing interest in database research. Toward this goal, access methods ......
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found