- Stanford, CA
- https://kushalthaman.github.io/
Stars
(WIP) A small but powerful, homemade PyTorch from scratch.
Puzzles for exploring transformers
Demo of the unit_scaling library, showing how a model can be easily adapted to train in FP8.
Long context evaluation for large language models
A bibliography and survey of the papers surrounding o1
NISQA - Non-Intrusive Speech Quality and TTS Naturalness Assessment
For optimization algorithm research and development.
first base model for full-duplex conversational audio
Scripts for fine-tuning Meta Llama with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supporting…
A fork of Anthropic Computer Use that you can run on Mac computers to give Claude and other AI models autonomous access to your computer.
prime is a framework for efficient, globally distributed training of AI models over the internet.
📃 A better UX for chat, writing content, and coding with LLMs.
Translate LaTeX-formatted linguistic dictionaries to Toolbox dictionaries and .xlsx sheets.
Entropy Based Sampling and Parallel CoT Decoding
A curriculum for learning about foundation models, from scratch to the frontier
What would you do with 1000 H100s...
EleutherAI / nanoGPT-mup
Forked from karpathy/nanoGPTThe simplest, fastest repository for training/finetuning medium-sized GPTs.
A PyTorch native library for large model training