🚀 Demo 1. Clone the repository git clone https://github.com/Vision-CAIR/MiniGPT4-video.git cd MiniGPT4-video 2. Set up the environment conda env create -f environment.yml 3. Download the checkpoints MiniGPT4-Video (Llama2 Chat 7B)MiniGPT4-Video (Mistral 7B) Download Download 4...
Github code :https://github.com/Vision-CAIR/MiniGPT4-video huggingface demo : https://huggingface.co/spaces/Vision-CAIR/MiniGPT4-video huggingface package : https://huggingface.co/Vision-CAIR/MiniGPT4-video-llama-hf example of using huggingface package from transformers import AutoModel video_pat...
MiniGPT4-Video demo # Llama2 python minigpt4_video_demo.py --ckpt path_to_video_checkpoint --cfg-path test_configs/llama2_test_config.yaml # Mistral python minigpt4_video_demo.py --ckpt path_to_video_checkpoint --cfg-path test_configs/mistral_test_config.yaml Inference Do the previous...
6 changes: 3 additions & 3 deletions6goldfish_demo.py Original file line numberDiff line numberDiff line change Expand Up@@ -13,8 +13,8 @@ fromindeximportMemoryIndex hf_token=os.environ.get('HF_TKN') login(token=hf_token) #hf_token = os.environ.get('HF_TKN') ...
2023/04/19: Code release & Online Demo VideoChat with ChatGPT: Explicit communication with ChatGPT. Sensitive with time. MiniGPT-4 for video: Implicit communication with Vicuna. Not sensitive with time. (Simple extension ofMiniGPT-4, which will be improved in the future.) ...
Code README License MiniGPT4-Video: Advancing Multimodal LLMs for Video Understanding with Interleaved Visual-Textual Tokens Online Demo Overview This paper introduces MiniGPT4-Video, a multimodal Large Language Model (LLM) designed specifically for video understanding. The model is capable of processin...