Edit model card

TEOChat: Large Language and Vision Assistant for Temporal Earth Observation Data

If you like our project, please give us a star โญ on GitHub for latest update.

๐Ÿ˜ฎ Highlights

TEOChat is the first language and vision assistant that can engage in conversation about sequences of temporal earth observation imagery, and exhibits impressive performance on multiple temporal instruction-following tasks.

๐Ÿ“š TEOChatlas: A new instruction-following dataset for temporal EO data

We introduce a new instruction-following dataset for temporal EO data called TEOChatlas which we use to train TEOChat. TEOChatlas contains 554,071 examples spanning dozens of temporal instruction-following tasks.

๐Ÿค– TEOChat: A new vision-language model for temporal EO data

We design TEOChat to use a LLaVA-style architecture, combining a temporally shared vision encoder with a LLaMA 2 LLM connected through an MLP vision-language projector

๐Ÿค— Demo

Gradio Web UI

We provide an online demo in Huggingface Spaces.

You can also run the demo locally by running the following command:

python videollava/serve/teochat_demo.py 

๐Ÿ› ๏ธ Requirements and Installation

  • Python >= 3.9
  • Pytorch == 2.2.1
  • CUDA Version >= 12.1
  • Install required packages:
git clone https://github.com/ermongroup/TEOChat.git
cd TEOChat
conda create -n teochat python=3.9 -y
conda activate teochat
pip install --upgrade pip  # enable PEP 660 support
pip install -r requirements.txt

๐Ÿ—๏ธ Training & Validating

The training & validating instructions are in TRAIN_AND_VALIDATE.md.

๐Ÿ‘ Acknowledgement

  • Video-LLaVA The codebase and model we built upon.
  • GeoChat The single image instruction-following dataset we included in TEOChatlas.

๐Ÿ”’ License

  • The majority of this project is released under the Apache 2.0 license as found in the LICENSE file.
  • The service is a research preview intended for non-commercial use only, subject to the model License of LLaMA, Terms of Use of the data generated by OpenAI, and Privacy Practices of ShareGPT. Please contact us if you find any potential violation.

โœ๏ธ Citation

If you find our paper and code useful in your research, please consider giving a star โญ and citation โœ๏ธ.

@article{irvin2024teochat,
  title={TEOChat: A Large Vision-Language Assistant for Temporal Earth Observation Data},
  author={Liu, Emily Ruoyu and Chen, Joyce Chuyi and Dormoy, Ines and Kim, Jinyoung and Khanna, Samar and Zheng, Zhuo and Ermon, Stefano},
  journal={arXiv preprint arXiv:2410.06234},
  year={2024}
}
Downloads last month
155
Inference API
Unable to determine this model's library. Check the docs .

Space using jirvin16/TEOChat 1