-
Tongji University
- Shanghai
Stars
Collection of Open Source Projects Related to GPT,GPT相关开源项目合集🚀、精选🔥🔥
自动驾驶散修一枚 -> 记录自己对自动驾驶的学习过程和相关学习链接
A Multi-Modal Large Language Model with Retrieval-augmented In-context Learning capacity designed for generalisable and explainable end-to-end driving
CVPR 2024 Papers Autonomous Driving
PyTorch implementation for the paper "Driving with LLMs: Fusing Object-Level Vector Modality for Explainable Autonomous Driving"
A curated list of awesome LLM for Autonomous Driving resources (continually updated)
LLaMA-VID: An Image is Worth 2 Tokens in Large Language Models (ECCV 2024)
Video-LLaVA: Learning United Visual Representation by Alignment Before Projection
[CVPR 2024 Highlight🔥] Chat-UniVi: Unified Visual Representation Empowers Large Language Models with Image and Video Understanding
Demo code of the paper "Deep Image Registration With Depth-Aware Homography Estimation"
Obtain bird's eye view of a scene from a single input image
[TMM 2023] Self-paced Curriculum Adapting of CLIP for Visual Grounding.
Transferable Decoding with Visual Entities for Zero-Shot Image Captioning, ICCV 2023
🦦 Otter, a multi-modal model based on OpenFlamingo (open-sourced version of DeepMind's Flamingo), trained on MIMIC-IT and showcasing improved instruction-following and in-context learning ability.
BuboGPT: Enabling Visual Grounding in Multi-Modal LLMs
[CVPR 2024] 🎬💭 chat with over 10K frames of video!
ltp1995 / KAGS
Forked from Tongji-MIC-Lab/KAGS[TPAMI'2023]Knowledge-enriched Attention Network with Group-wise Semantic for Visual Storytelling
[ICIP'23]Structure-aware Generative Adversarial Network for Text-to-image Generation
Codes From Top Teams in 2023 AIC challenge
VideoCC is a dataset containing (video-URL, caption) pairs for training video-text machine learning models. It is created using an automatic pipeline starting from the Conceptual Captions Image-Cap…
Codes and Models for VALOR: Vision-Audio-Language Omni-Perception Pretraining Model and Dataset
[CVPR2024 Highlight][VideoChatGPT] ChatGPT with video understanding! And many more supported LMs such as miniGPT4, StableLM, and MOSS.
Open-sourced codes for MiniGPT-4 and MiniGPT-v2 (https://minigpt-4.github.io, https://minigpt-v2.github.io/)