Stars
DeepGEMM: clean and efficient FP8 GEMM kernels with fine-grained scaling
An extremely fast Python package and project manager, written in Rust.
Code for the paper "VinePPO: Unlocking RL Potential For LLM Reasoning Through Refined Credit Assignment"
My learning notes/codes for ML SYS.
Scalable RL solution for advanced reasoning of language models
(WIP) A small but powerful, homemade PyTorch from scratch.
Puzzles for exploring transformers
Demo of the unit_scaling library, showing how a model can be easily adapted to train in FP8.
Long context evaluation for large language models
A bibliography and survey of the papers surrounding o1
NISQA - Non-Intrusive Speech Quality and TTS Naturalness Assessment
For optimization algorithm research and development.
first base model for full-duplex conversational audio
Welcome to the Llama Cookbook! This is your go to guide for Building with Llama: Getting started with Inference, Fine-Tuning, RAG. We also show you how to solve end to end problems using Llama mode…
A fork of Anthropic Computer Use that you can run on Mac computers to give Claude and other AI models autonomous access to your computer.
prime is a framework for efficient, globally distributed training of AI models over the internet.
📃 A better UX for chat, writing content, and coding with LLMs.
Writing and Citation Assistant Tool
Translate LaTeX-formatted linguistic dictionaries to Toolbox dictionaries and .xlsx sheets.