Skip to content
#

vizwiz-vqa

Here are 7 public repositories matching this topic...

Language: All
Filter by language

A self-evident application of the VQA task is to design systems that aid blind people with sight reliant queries. The VizWiz VQA dataset originates from images and questions compiled by members of the visually impaired community and as such, highlights some of the challenges presented by this particular use case.

  • Updated Dec 12, 2023
  • Python

A Light weight deep learning model with with a web application to answer image-based questions with a non-generative approach for the VizWiz grand challenge 2023 by carefully curating the answer vocabulary and adding linear layer on top of Open AI's CLIP model as image and text encoder

  • Updated Jun 27, 2023
  • Jupyter Notebook

Improve this page

Add a description, image, and links to the vizwiz-vqa topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the vizwiz-vqa topic, visit your repo's landing page and select "manage topics."

Learn more