-
Tsinghua University
Highlights
- Pro
Lists (3)
Sort Name ascending (A-Z)
Stars
DeepSeek-VL: Towards Real-World Vision-Language Understanding
Qwen2.5 is the large language model series developed by Qwen team, Alibaba Cloud.
Use PEFT or Full-parameter to finetune 450+ LLMs (Qwen2.5, InternLM3, GLM4, Llama3.3, Mistral, Yi1.5, Baichuan2, DeepSeek-R1, ...) and 150+ MLLMs (Qwen2.5-VL, Qwen2-Audio, Llama3.2-Vision, Llava, I…
Personal Project: MPP-Qwen14B & MPP-Qwen-Next(Multimodal Pipeline Parallel based on Qwen-LM). Support [video/image/multi-image] {sft/conversations}. Don't let the poverty limit your imagination! Tr…
[ICLR2024] FTIC: Frequency-aware Transformer for Learned Image Compression
A Flexible Framework for Experiencing Cutting-edge LLM Inference Optimizations
Standardized benchmark for computational pathology foundation models.
Toolkit for large-scale whole-slide image processing.
Multimodal Distillation-Driven Ensemble Learning for Long-Tailed Histopathology Whole Slide Images Analysis
CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image
Prompt Learning for Vision-Language Models (IJCV'22, CVPR'22)
Awesome lists about framework figures in papers
List of pathology feature extractors and foundation models
An easy-to-use PyTorch library for Pathology Image ANalysis Orchestrator (PIANO), including generating patches from whole slide images, using pathology foundation models to create features, and so on.
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
Janus-Series: Unified Multimodal Understanding and Generation Models
The official implementation of Tensor ProducT ATTenTion Transformer (T6)
[Preprint] ConvMLP: Hierarchical Convolutional MLPs for Vision, 2021
Stain normalization tools for histological analysis and computational pathology
Clean, minimal, accessible reproduction of DeepSeek R1-Zero
⚡ Flash Diffusion ⚡: Accelerating Any Conditional Diffusion Model for Few Steps Image Generation (AAAI 2025 Oral)
[ICLR 2025] LLaVA-MoD: Making LLaVA Tiny via MoE-Knowledge Distillation
Mixture-of-Experts for Large Vision-Language Models
[NeurIPS 2023 Oral] Quilt-1M: One Million Image-Text Pairs for Histopathology.