Skip to content
/ DAST Public

[MM23] Code for paper "Depth-Aware Sparse Transformer for Video-Language Learning"

Notifications You must be signed in to change notification settings

zchoi/DAST

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

11 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Depth-Aware Sparse Transformer for Video-Language Learning

[Paper] | ACM MM23

This is the code implementation of the paper "Depth-Aware Sparse Transformer for Video-Language Learning", the checkpoint and feature will be released soon.

Overview

In Video-Language (VL) learning tasks, a massive amount of text annotations are describing geometrical relationships of instances (e.g., 19.6% to 45.0% in MSVD, MSR-VTT, MSVD-QA, and MSVRTTQA), which often become the bottleneck of the current VL tasks (e.g., 60.8% vs. 98.2% CIDEr in MSVD for geometrical and non-geometrical annotations). Considering the rich spatial information of depth map, an intuitive way is to enrich the conventional 2D visual representations with depth information through current SOTA models, i.e., transformer. However, it is cumbersome to compute the self-attention on a long-range sequence and heterogeneous video-level representations with regard to computation cost and flexibility on various frame scales. To tackle this, we propose a hierarchical transformer, termed Depth-Aware Sparse Transformer (DAST).


Figure 1. Overview of the DAST for Video-Language Learning.

About

[MM23] Code for paper "Depth-Aware Sparse Transformer for Video-Language Learning"

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages