Skip to content

mattmazzola/LLaVA-Interactive-Demo

 
 

Repository files navigation

🌋 LLaVA-Interactive

An All-in-One Demo for Image Chat, Segmentation and Generation/Editing.

[Project Page] [Demo] [Paper]


Install

Installing this project requires CUDA 11.7 or above. Follow the steps below:

git clone https://github.com/LLaVA-VL/LLaVA-Interactive-Demo.git
conda create -n llava_int -c conda-forge -c pytorch python=3.10.8 pytorch=2.0.1 -y
conda activate llava_int
cd LLaVA-Interactive-Demo
pip install -r requirements.txt
source setup.sh

Run the demo

To run the demo, simply run the shell script.

./run_demo.sh


Citation

If you find LLaVA-Interactive useful for your research and applications, please cite using this BibTeX:

  @article{chen2023llava_interactive,
    author      = {Chen, Wei-Ge and Spiridonova, Irina and Yang, Jianwei and Gao, Jianfeng and Li, Chunyuan},
    title       = {LLaVA-Interactive: An All-in-One Demo for Image Chat, Segmentation, Generation and Editing},
    publisher   = {arXiv:2311.00571},
    year        = {2023}
  }

Related Projects

Acknowledgement

  • LaMa: A nice tool we use to fill the background holes in images.

License

This project including LLaVA and SEEM are licensed under the Apache License. See the LICENSE file for more details. The GLIGEN project is licensed under the MIT License.

About

LLaVA-Interactive-Demo

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Languages

  • Python 96.9%
  • Shell 2.5%
  • Dockerfile 0.6%