Highlights
- Pro
Stars
MoBA: Mixture of Block Attention for Long-Context LLMs
A collection of resources and papers on Diffusion Models
把萌萌哒的看板娘抱回家 (ノ≧∇≦)ノ | Live2D widget for web platform
Pytorch implementation for Semantic Segmentation/Scene Parsing on MIT ADE20K dataset
[NeurIPS 2024] Official Repository of The Mamba in the Llama: Distilling and Accelerating Hybrid Models
Code for paper [Neat: Nonlinear Parameter-efficient Adaptation of Pre-trained Models]
Awesome Knowledge Distillation
🚀 Efficient implementations of state-of-the-art linear attention models in Torch and Triton
✨✨Latest Advances on Multimodal Large Language Models
Code for paper [Low-Rank Interconnected Adaptation across Layers]
An efficient, flexible and full-featured toolkit for fine-tuning LLM (InternLM2, Llama3, Phi3, Qwen, Mistral, ...)
CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image
Lecture materials for Cornell CS5785 Applied Machine Learning (Fall 2024)
Speech To Speech: an effort for an open-sourced and modular GPT4-o
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
Collection of AWESOME vision-language models for vision tasks
Stanford NLP Python library for Representation Finetuning (ReFT)
A simple and efficient Mamba implementation in pure PyTorch and MLX.
A Survey on multimodal learning research.
Course to get into Large Language Models (LLMs) with roadmaps and Colab notebooks.
Building blocks for foundation models.
Fast and memory-efficient exact attention
Code for our EMNLP 2023 Paper: "LLM-Adapters: An Adapter Family for Parameter-Efficient Fine-Tuning of Large Language Models"
[ICML2024 (Oral)] Official PyTorch implementation of DoRA: Weight-Decomposed Low-Rank Adaptation