Stars
[CVPR 2024 Highlight] MIGC and [TPAMI 2024] MIGC++ (Official Implementation)
An efficient, flexible and full-featured toolkit for fine-tuning LLM (InternLM2, Llama3, Phi3, Qwen, Mistral, ...)
The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
Official implementation of our paper "Finetuned Multimodal Language Models are High-Quality Image-Text Data Filters".
🔥 [CVPR2024] Official implementation of "Self-correcting LLM-controlled Diffusion Models (SLD)
💡 同济大学计算机科学与技术、信息安全专业课程资源共享仓库。含部分课堂笔记、课程教材、推荐书籍、课后习题、习题解答、报告模板、实验工具等内容。期待更多课程加入……
[Neurips 2023] T2I-CompBench: A Comprehensive Benchmark for Open-world Compositional Text-to-image Generation
LLMs can generate feedback on their work, use it to improve the output, and repeat this process iteratively.
LLM-grounded Diffusion: Enhancing Prompt Understanding of Text-to-Image Diffusion Models with Large Language Models (LLM-grounded Diffusion: LMD, TMLR 2024)
[CVPR 2024] Code release for "InstanceDiffusion: Instance-level Control for Image Generation"
[ICLR 2024] Official code for the paper "LLM Blueprint: Enabling Text-to-Image Generation with Complex and Detailed Prompts"
[ECCV 2024] The official code of paper "Open-Vocabulary SAM".
OMG-LLaVA and OMG-Seg codebase [CVPR-24 and NeurIPS-24]
TinyGPT-V: Efficient Multimodal Large Language Model via Small Backbones
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
[CVPR 2024] Aligning and Prompting Everything All at Once for Universal Visual Perception
Official PyTorch implementation of "Multi-modal Queried Object Detection in the Wild" (accepted by NeurIPS 2023)
OpenMMLab Pose Estimation Toolbox and Benchmark.
OpenMMLab Detection Toolbox and Benchmark
[ECCV 2024] Official implementation of the paper "Grounding DINO: Marrying DINO with Grounded Pre-Training for Open-Set Object Detection"
[MICCAI 2024] On-the-Fly Guidance Training for Medical Image Registration. Pre-print available in link below.