You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
According to the ReadMe at https://github.com/OpenGVLab/InternVideo/tree/main/InternVideo1/Downstream/Video-Text-Retrieval, the zero-shot retrieval results will be obtained after running the command ./zeroshot_scripts/eval_msrvtt.sh. This command will execute the main_task_retrieval.py. But in "main_task_retrieval.py", I find that the model is CLIP4CLIP, instead of ViCLIP. I'd like to know how to conduct zero-shot video-text retrieval experiments with pretrained ViCLIP.
The text was updated successfully, but these errors were encountered:
According to the ReadMe at https://github.com/OpenGVLab/InternVideo/tree/main/InternVideo1/Downstream/Video-Text-Retrieval, the zero-shot retrieval results will be obtained after running the command
./zeroshot_scripts/eval_msrvtt.sh
. This command will execute themain_task_retrieval.py
. But in "main_task_retrieval.py", I find that the model is CLIP4CLIP, instead of ViCLIP. I'd like to know how to conduct zero-shot video-text retrieval experiments with pretrained ViCLIP.The text was updated successfully, but these errors were encountered: