Lists (1)
Sort Name ascending (A-Z)
Stars
Fully open reproduction of DeepSeek-R1
This is the official repository for The Hundred-Page Language Models Book by Andriy Burkov
Efficient Triton Kernels for LLM Training
一个用于llava 7b模型微调的仓库,主要用于理解算法设计、deepspeed分布训练、模型量化等
一个快速学习deepseekV3模型以及r1强化学习grpo的仓库,侧重于理解技术报告与模型设计细节,理解原理
Train transformer language models with reinforcement learning.
verl: Volcano Engine Reinforcement Learning for LLMs
Large Language-and-Vision Assistant for Biomedicine, built towards multimodal GPT-4 level capabilities.
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
PyTorch code for BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation
A library for efficient similarity search and clustering of dense vectors.
PyTorch implementation of MAE https//arxiv.org/abs/2111.06377
Hiera: A fast, powerful, and simple hierarchical vision transformer.
This is an official implementation for "Swin Transformer: Hierarchical Vision Transformer using Shifted Windows".
一个用于快速入门transformer的仓库,梳理相关nlp和vit模型结构、原理,训练的基本步骤及微调方法, 配套能快速学习的代码实战项目
we want to create a repo to illustrate usage of transformers in chinese
Official repository of "SAMURAI: Adapting Segment Anything Model for Zero-Shot Visual Tracking with Motion-Aware Memory"
整理开源的中文大语言模型,以规模较小、可私有化部署、训练成本较低的模型为主,包括底座模型,垂直领域微调及应用,数据集与教程等。
Comprehensive toolkit for Reinforcement Learning from Human Feedback (RLHF) training, featuring instruction fine-tuning, reward model training, and support for PPO and DPO algorithms with various c…
MedicalGPT: Training Your Own Medical GPT Model with ChatGPT Training Pipeline. 训练医疗大模型,实现了包括增量预训练(PT)、有监督微调(SFT)、RLHF、DPO、ORPO、GRPO。
Reference implementation for DPO (Direct Preference Optimization)
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
手把手带你实战 Huggingface Transformers 课程视频同步更新在B站与YouTube
🚀🚀 「大模型」2小时完全从0训练26M的小参数GPT!🌏 Train a 26M-parameter GPT from scratch in just 2h!