question-mark
Stuck on an issue?

Lightrun Answers was designed to reduce the constant googling that comes with debugging 3rd party libraries. It collects links to all the places you might be looking at while hunting down a tough bug.

And, if you’re still stuck at the end, we’re happy to hop on a call to see how we can help out.

Document `ObjectDetectionData` transforms

See original GitHub issue

📚 Documentation

It is possible (although not trivial) to get bounding box augmentations / transforms working with ObjectDetectionData (see #414). We should document this somewhere, ideally providing some examples.

cc @hzitoun

Issue Analytics

  • State:closed
  • Created 2 years ago
  • Comments:5 (4 by maintainers)

github_iconTop GitHub Comments

1reaction
ethanwharriscommented, Sep 6, 2021

Hi @gianscarpe, some documentation here would definitely be welcome 😃 Our object detection now relies on IceVision underneath so our support for albumentations is much better than it used to be. You should now be able to wrap albumentations augmentations in the IceVisionTransformAdapter (maybe the name AlbumentationsAdapter or something would be better?) here: https://github.com/PyTorchLightning/lightning-flash/blob/cf86275d629092e4c43f2d996061bff7abe17784/flash/core/integrations/icevision/transforms.py#L198

There’s a couple of examples further down that page where we use the default transforms from IceVision. It would be very welcome for you to add a documentation section about this similar to what we have for image classification here: https://lightning-flash.readthedocs.io/en/latest/reference/image_classification.html#custom-transformations

Thanks for your help 😃

0reactions
ethanwharriscommented, Sep 13, 2021

Hi @gianscarpe in theory we should now support this, but it would be great for you to try i out and maybe update the documentation there if you’re interested. The way it should work is that you can pass a metric from Icevision to the ObjectDetector task. Here’s some code that should be working:

from icevision.metrics import COCOMetric, COCOMetricType

model = ObjectDetector(..., metrics=[COCOMetric(COCOMetricType.bbox)])

Let me know if that works and how the experience there could be improved! 😃

Read more comments on GitHub >

github_iconTop Results From Across the Web

arXiv:2008.10843v1 [cs.CV] 25 Aug 2020
In this paper, we present a novel end-to-end trainable deep learning based framework to localize graphical objects in the document images called ...
Read more >
Bounding boxes augmentation for object detection
You pass an image and bounding boxes to the augmentation pipeline and receive augmented images and boxes. Note. Some transforms in Albumentation don't...
Read more >
Perspective Transformation Data Augmentation for Object ...
Perspective Transformation Data Augmentation for Object Detection. Abstract: One major reason for the success of convolutional neural ...
Read more >
Finding Objects In Document Images | by Cinnamon AI - Medium
As the input to Document Object Recognition (DOR) is an image, CNN is employed to automatically transform this image into a set of...
Read more >
Prepare custom datasets for object detection
data.transforms to support almost all types of data augmentations. More specifically, the label of object detection task is described as follows:.
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found