PyTorch code for BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation
-
Updated
May 20, 2024 - Jupyter Notebook
PyTorch code for BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation
Deep Modular Co-Attention Networks for Visual Question Answering
FiLM: Visual Reasoning with a General Conditioning Layer
Recent Papers including Neural Symbolic Reasoning, Logical Reasoning, Visual Reasoning, planning and any other topics connecting deep learning and reasoning
RAVEN: A Dataset for Relational and Analogical Visual rEasoNing
Pytorch implementation of "Explainable and Explicit Visual Reasoning over Scene Graphs "
[ICML 2023] UPop: Unified and Progressive Pruning for Compressing Vision-Language Transformers.
Image captioning using python and BLIP
[CVPR 2022 (oral)] Bongard-HOI for benchmarking few-shot visual reasoning
Visual Question Reasoning on General Dependency Tree
Learning Perceptual Inference by Contrasting
NeuSyRE: A Neuro-Symbolic Visual Understanding and Reasoning Framework based on Scene Graph Enrichment
Source code for my honours thesis: "Graph Attention Networks for Compositional Visual Question Answering"
LaTeX files for my honours thesis: "Graph Attention Networks for Compositional Visual Question Answering"
[ICLR 2022] RelViT: Concept-guided Vision Transformer for Visual Relational Reasoning
Abstract Spatial-Temporal Reasoning via Probabilistic Abduction and Execution
ACRE: Abstract Causal REasoning Beyond Covariation
An alternative EQA paradigm and informative benchmark + models (BMVC 2019, ViGIL 2019 spotlight)
Add a description, image, and links to the visual-reasoning topic page so that developers can more easily learn about it.
To associate your repository with the visual-reasoning topic, visit your repo's landing page and select "manage topics."