question-mark
Stuck on an issue?

Lightrun Answers was designed to reduce the constant googling that comes with debugging 3rd party libraries. It collects links to all the places you might be looking at while hunting down a tough bug.

And, if you’re still stuck at the end, we’re happy to hop on a call to see how we can help out.

[question] Report of evaluating AAE with ground truth bounding boxes is weird

See original GitHub issue

Describe the bug I installed some libraries about latex and try to generate report.pdf of evaluating AAE with ground truth bounding boxes, but the result is weird. I retrained an Augmented Autoencoder, some of their results are similar, while others are a little different. Their translation error is far large than your report in doc directory. @MartinSmeyer Do you know where is the problem and what should I do?

System Info Describe the characteristic of your environment:

  • GPU GTX 2070
  • Python python2.7
  • PyOpenGL 3.1.3rc1

The configuration of the two reports is the same. Train config:

[Paths]
MODEL_PATH: /home/zhiyong/server/data/t-less/t-less_v2/models/models_reconst/obj_05.ply
BACKGROUND_IMAGES_GLOB: /home/zhiyong/server/data/VOCdevkit/VOC2012/JPEGImages/*.jpg

[Dataset]
MODEL: reconst
H: 128
W: 128
C: 3
RADIUS: 700
RENDER_DIMS: (720, 540)
K: [1075.65, 0, 720/2, 0, 1073.90, 540/2, 0, 0, 1]
# Scale vertices to mm
VERTEX_SCALE: 1
ANTIALIASING: 1
PAD_FACTOR: 1.2
CLIP_NEAR: 10
CLIP_FAR: 10000
NOOF_TRAINING_IMGS: 20000
NOOF_BG_IMGS: 15000

[Augmentation]
REALISTIC_OCCLUSION: False
SQUARE_OCCLUSION: False
MAX_REL_OFFSET: 0.20
CODE: Sequential([
	#Sometimes(0.5, PerspectiveTransform(0.05)),
	#Sometimes(0.5, CropAndPad(percent=(-0.05, 0.1))),
	Sometimes(0.5, Affine(scale=(1.0, 1.2))),
	Sometimes(0.5, CoarseDropout( p=0.2, size_percent=0.05) ),
	Sometimes(0.5, GaussianBlur(1.2*np.random.rand())),
    Sometimes(0.5, Add((-25, 25), per_channel=0.3)),
    Sometimes(0.3, Invert(0.2, per_channel=True)),
    Sometimes(0.5, Multiply((0.6, 1.4), per_channel=0.5)),
    Sometimes(0.5, Multiply((0.6, 1.4))),
    Sometimes(0.5, ContrastNormalization((0.5, 2.2), per_channel=0.3))
	], random_order=False)

[Embedding]
EMBED_BB: True
MIN_N_VIEWS: 2562
NUM_CYCLO: 36

[Network]
BATCH_NORMALIZATION: False
AUXILIARY_MASK: False
VARIATIONAL: 0
LOSS: L2
BOOTSTRAP_RATIO: 4
NORM_REGULARIZE: 0
LATENT_SPACE_SIZE: 128
NUM_FILTER: [128, 256, 512, 512]
STRIDES: [2, 2, 2, 2]
KERNEL_SIZE_ENCODER: 5
KERNEL_SIZE_DECODER: 5


[Training]
OPTIMIZER: Adam
NUM_ITER: 30000
BATCH_SIZE: 64
LEARNING_RATE: 2e-4
SAVE_INTERVAL: 10000

[Queue]
# OPENGL_RENDER_QUEUE_SIZE: 500
NUM_THREADS: 10
QUEUE_SIZE: 50

Evaluate config:

[METHOD]
method = ae

[DATA]
dataset = tless
cam_type = primesense
scenes = [] #empty means all scenes
obj_id = 5

[BBOXES]
estimate_bbs = False
ckpt = /path/to/detector/checkpoints/freezed_graph.pb
external = /home/zhiyong/server/data/t-less/t-less_v2/test_primesense/test_predicts
pad_factor = 1.2
single_instance = True

[EVALUATION]
icp = False
compute_errors = False
evaluate_errors = False
top_n_eval = 0

[METRIC]
error_thres = {'vsd':0.3,'cou':0.5,'te':5.0,'re':5.0}
error_thresh_fact = {'add':0.1,'adi':0.1}
error_type = ['vsd','re','te']
top_n = 1
vsd_delta = 15
vsd_tau = 20
vsd_cost = step

[PLOT]
nearest_neighbors = False
scene_with_estimate = False
reconstruction = False
cum_t_error_hist = True
cum_r_error_hist = True
cum_vsd_error_hist = True
vsd_occlusion = True
r_error_occlusion = True
embedding_pca = True
animate_embedding_pca = False
viewsphere = True
reconstruction_test_batch = True

Additional context The translation error of the first report: image

The rotation error of the first report: image

The PCA embedding view sphere of the first report is: image

The Recall vs translation error of the two reports are similar: image

The translation error of the second report: image

The rotation error of the second report: image

The PCA embedding view sphere of the second report is: image

Issue Analytics

  • State:closed
  • Created 4 years ago
  • Comments:5 (1 by maintainers)

github_iconTop GitHub Comments

1reaction
qyz55commented, Dec 16, 2019

I trained on obj_05 twice, with no changes to train_template.cfg except the model file, and got 60/66 vsd with gt bbox. Maybe you need to delete all the files in $AE_WORKSPACE_PATH/tmp_datasets to generate new images for training. Otherwise no differences can be seen by repetitive training.

0reactions
DateBrocommented, Dec 16, 2019

Sorry for not reading the paper of the arxiv version, which added some results such as translation error and rotation error. Besides, I forgot that the report in the docs directory is the result of ICP. The improvement of translation error which is generated by ICP is impressive. Thanks for your help!

Read more comments on GitHub >

github_iconTop Results From Across the Web

CS 2770: Homework 3
We will look at how well a predicted positive bounding box matches a ground-truth bounding box, and will then compute mean average precision ......
Read more >
Bounding Box - Amazon SageMaker - AWS Documentation
To classify and localize one or more objects within images, use the Amazon SageMaker Ground Truth bounding box labeling job task type.
Read more >
Handling data without ground truth bounding boxes in SSD ...
@Sycorax As I understand the question, it does not regard unlabeled data, rather data which are correctly labeled as having no objects in...
Read more >
Towards Monocular 6D Object Pose Estimation - mediaTUM
studying the problem of 6D pose estimation from monocular input using deep learning. ... in the image, based on the encapsulating ground truth...
Read more >
Evaluation Methodology for Object Detection and Tracking in ...
the evaluation process of software used to detect vehicles on ... bounding boxes taken from ground truth (GT) are more trustworthy than ...
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found