Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Run inference on own picture with externally inputted object labels and bounding boxes #182

Open
aliencaocao opened this issue Jan 17, 2022 · 0 comments

Comments

@aliencaocao
Copy link

I have some images that contain a mixture of seen and unseen object classes. I have my own custom object detection model based on YOLOv5, and it is able to output bounding box and class label and confidence. Is it possible to input these YOLOv5 results into Oscar+ and thus only use the text generation part of Oscar+ to generate a caption of the image? Original image with bounding box can be inputted, but I do not want Oscar+ to do the object detection part as my own model take care of some unseen objects.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant