Skip to content

Latest commit

History

History
51 lines (37 loc) 路 2.09 KB

offline_demo.md

File metadata and controls

51 lines (37 loc) 路 2.09 KB

Running Video-ChatGPT Demo Offline

Please follow the instructions below to run the Video-ChatGPT demo on your local GPU machine.

Note: Our demo requires approximately 18 GB of GPU memory.

Clone the Video-ChatGPT repository

git clone https://github.com/mbzuai-oryx/Video-ChatGPT.git
cd Video-ChatGPT
export PYTHONPATH="./:$PYTHONPATH"

Download Video-ChatGPT weights

wget https://huggingface.co/MBZUAI/Video-ChatGPT-7B/resolve/main/video_chatgpt-7B.bin

More details at: https://huggingface.co/MBZUAI/Video-ChatGPT-7B.

Prepare LLaVA weights

Video-ChatGPT is build using LLaVA. Please follow the following instructions to get LLaVA weights.

  • Get the original LLaMA weights in the huggingface format by following the instructions here.
  • Use the following scripts to get LLaVA weights by applying our delta.
python scripts/apply_delta.py \ 
        --base-model-path <path to LLaMA 7B weights> \
        --target-model-path LLaVA-Lightning-7B-v1-1 \
        --delta-path liuhaotian/LLaVA-Lightning-7B-delta-v1-1

The above command will download the LLaVA-Lightening-7B-v1-1 delta from HuggingFace, apply it to the provided LLaMA weights and save the LLaVA-Lightening-7B-v1-1 weights in the current directory.

Alternatively you can download the ready LLaVA-Lightening-7B weights from mmaaz60/LLaVA-Lightening-7B-v1-1.

Download sample videos

Download the sample videos to test the demo from this link and place them in video_chatgpt/demo_sample_videos directory.

Run the Demo

python video_chatgpt/demo/video_demo.py 
        --model-name <path to the LLaVA-Lightening-7B-v1-1 weights prepared in step 3> \
        --projection_path <path to the downloaded video-chatgpt weights>

Follow the instructions on the screen to open the demo dashboard.