Stars
pre-train code
3 repositories
The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.
Finetune Llama 3.3, DeepSeek-R1, Mistral, Phi-4 & Gemma 2 LLMs 2-5x faster with 70% less memory