Please follow the instructions below to run the Video-ChatGPT demo on your local GPU machine.
Note: Our demo requires approximately 18 GB of GPU memory.
git clone https://github.com/mbzuai-oryx/Video-ChatGPT.git
cd Video-ChatGPT
export PYTHONPATH="./:$PYTHONPATH"
wget https://huggingface.co/MBZUAI/Video-ChatGPT-7B/resolve/main/video_chatgpt-7B.bin
More details at: https://huggingface.co/MBZUAI/Video-ChatGPT-7B.
Video-ChatGPT is build using LLaVA. Please follow the following instructions to get LLaVA weights.
- Get the original LLaMA weights in the huggingface format by following the instructions here.
- Use the following scripts to get LLaVA weights by applying our delta.
python scripts/apply_delta.py \
--base-model-path <path to LLaMA 7B weights> \
--target-model-path LLaVA-Lightning-7B-v1-1 \
--delta-path liuhaotian/LLaVA-Lightning-7B-delta-v1-1
The above command will download the LLaVA-Lightening-7B-v1-1 delta from HuggingFace, apply it to the provided LLaMA weights and save the LLaVA-Lightening-7B-v1-1 weights in the current directory.
Alternatively you can download the ready LLaVA-Lightening-7B weights from mmaaz60/LLaVA-Lightening-7B-v1-1.
Download the sample videos to test the demo from this link
and place them in video_chatgpt/demo_sample_videos
directory.
python video_chatgpt/demo/video_demo.py
--model-name <path to the LLaVA-Lightening-7B-v1-1 weights prepared in step 3> \
--projection_path <path to the downloaded video-chatgpt weights>
Follow the instructions on the screen to open the demo dashboard.