-
Notifications
You must be signed in to change notification settings - Fork 7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[feature request] transforms for object detection #3286
Comments
Hi, Thanks for opening this issue. This has been in our radar for a while already, but we never really managed to find out the right balance between simplicity and generality about the API. For an earlier attempt for the APIs, see #1406 and the discussion within. I would love to hear your thoughts on this. |
FYI, It seems that the existing vision/torchvision/models/detection/transform.py Lines 199 to 217 in 51500c7
|
🚀 Feature
I would like to start adding/supporting transforms (both functional and class) for object detection, I know I can take some of them from
references
folder. But, it would nice to have OOTB. Here are a few basic transforms I would like to add first -RandomHorizontalFlipWithBBox
RandomVerticalFlipWithBBox
LetterBox
Pitch
All of the above transforms will accept 2 arguments when they are called. This breaks the purpose of
Compose
andnn.Sequential
, but currently aren't we writing customCompose
ornn.Sequential
? So I think it's ok to start introducing necessary transforms taking 2 arguments for detection, segmentation, etc and let users write customCompose
ornn.Sequential
the way they would to like to call the transforms.Additional context
Current code:
Thank you!
cc @vfdev-5, @fmassa
The text was updated successfully, but these errors were encountered: