Starred repositories
The official repo of Pai-Megatron-Patch for LLM & VLM large scale training developed by Alibaba Cloud.
A huggingface transformers implementation of "Transformer Memory as a Differentiable Search Index"
Code for the paper "A Structural Model for Contextual Code Changes"
✨ Light and Fast AI Assistant. Support: Web | iOS | MacOS | Android | Linux | Windows
A library that provides an embeddable, persistent key-value store for fast storage.
🚀 Collection of components for development, training, tuning, and inference of foundation models leveraging PyTorch native components.
Medusa: Simple Framework for Accelerating LLM Generation with Multiple Decoding Heads
🚀 Efficiently (pre)training foundation models with native PyTorch features, including FSDP for training and SDPA implementation of Flash attention v2.
General technology for enabling AI capabilities w/ LLMs and MLLMs
Train transformer language models with reinforcement learning.
CodeRAG-Bench: Can Retrieval Augment Code Generation?
SGLang is a fast serving framework for large language models and vision language models.
A library with extensible implementations of DPO, KTO, PPO, ORPO, and other human-aware loss functions (HALOs).
Code at the speed of thought – Zed is a high-performance, multiplayer code editor from the creators of Atom and Tree-sitter.
SWE-agent takes a GitHub issue and tries to automatically fix it, using GPT-4, or your LM of choice. It can also be employed for offensive cybersecurity or competitive coding challenges. [NeurIPS 2…
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
The official Python library for the OpenAI API
Dataflow-guided retrieval augmentation for repository-level code completion, ACL 2024 (main)
Fast and memory-efficient exact attention
Source codes for paper ”ReACC: A Retrieval-Augmented Code Completion Framework“
Retrieval and Retrieval-augmented LLMs
Fast, indexed regexp search over large file trees
official repository of aiXcoder-7B Code Large Language Model