Document `ObjectDetectionData` transforms
See original GitHub issue
Top Results From Across the Web
arXiv:2008.10843v1 [cs.CV] 25 Aug 2020
In this paper, we present a novel end-to-end trainable deep learning based framework to localize graphical objects in the document images called ...
Read more >Bounding boxes augmentation for object detection
You pass an image and bounding boxes to the augmentation pipeline and receive augmented images and boxes. Note. Some transforms in Albumentation don't...
Read more >Perspective Transformation Data Augmentation for Object ...
Perspective Transformation Data Augmentation for Object Detection. Abstract: One major reason for the success of convolutional neural ...
Read more >Finding Objects In Document Images | by Cinnamon AI - Medium
As the input to Document Object Recognition (DOR) is an image, CNN is employed to automatically transform this image into a set of...
Read more >Prepare custom datasets for object detection
data.transforms to support almost all types of data augmentations. More specifically, the label of object detection task is described as follows:.
Read more >
Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free
Top Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found

Hi @gianscarpe, some documentation here would definitely be welcome 😃 Our object detection now relies on IceVision underneath so our support for albumentations is much better than it used to be. You should now be able to wrap albumentations augmentations in the
IceVisionTransformAdapter(maybe the nameAlbumentationsAdapteror something would be better?) here: https://github.com/PyTorchLightning/lightning-flash/blob/cf86275d629092e4c43f2d996061bff7abe17784/flash/core/integrations/icevision/transforms.py#L198There’s a couple of examples further down that page where we use the default transforms from IceVision. It would be very welcome for you to add a documentation section about this similar to what we have for image classification here: https://lightning-flash.readthedocs.io/en/latest/reference/image_classification.html#custom-transformations
Thanks for your help 😃
Hi @gianscarpe in theory we should now support this, but it would be great for you to try i out and maybe update the documentation there if you’re interested. The way it should work is that you can pass a metric from Icevision to the ObjectDetector task. Here’s some code that should be working:
Let me know if that works and how the experience there could be improved! 😃