What if the training pairs are without bbox?
See original GitHub issue area[i, 0] = obj['bbox'][2]*obj['bbox'][3]
This method requires bbox during training. What if the training pairs are without bbox?
How can we calculate the area
and loss of offset
?
Issue Analytics
- State:
- Created 3 years ago
- Comments:12
Top Results From Across the Web
Why is my Trained SSD Object Detector not giving a bounding ...
Training on single GPU. Initializing input data normalization. I have tried training both with just the license plate bounding boxes(mydata. ...
Read more >Intersection over Union (IoU) for object detection
Your goal is to take the training images + bounding boxes, construct an object detector, and then evaluate its performance on the testing...
Read more >Building an object detector in TensorFlow using bounding-box ...
However, the downside of using only positive image for training the localisation branch is that it gives false positives for negative images.
Read more >LayoutLM - Hugging Face
We're on a journey to advance and democratize artificial intelligence through open source and open science.
Read more >14.4. Anchor Boxes — Dive into Deep Learning 1.0.0-beta0 ...
At this time, the IoU of any pair of predicted bounding boxes in L is below the threshold ϵ ; thus, no pair...
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
For your case, I think the heatmap is more suitable because you just need to detect the head keypoints and the heatmap can detect all the same kind of the keypoints precisely and efficiently.
Hi, we use the area of one person’s bounding box to normalize this person’s offset during calculating the loss of the offset map. Alternatively, you can use the maximum and minimum coordinates of the key points to estimate the size of the people. The selection of this normalization tool doesn’t influence the performance.
By the way, we use the average of all keypoints belong to the same person as the center of this person. In the code, we can choose the center of the person’s bounding box as the center of this person, which may cause confusion. The APs of the two choices are the same.