- Barcelona, Catalonia, Spain
- ggcr.github.io
Starred repositories
A curated list of awesome Machine Learning frameworks, libraries and software.
🧑🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), ga…
Deep Learning papers reading roadmap for anyone who are eager to learn this amazing tech!
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
A high-throughput and memory-efficient inference and serving engine for LLMs
The fundamental package for scientific computing with Python.
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
Finetune Llama 3.3, Mistral, Phi-4, Qwen 2.5 & Gemma LLMs 2-5x faster with 70% less memory
Ongoing research training transformer models at scale
The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.
SGLang is a fast serving framework for large language models and vision language models.
A framework for few-shot evaluation of language models.
Accessible large language models via k-bit quantization for PyTorch.
arXiv LaTeX Cleaner: Easily clean the LaTeX code of your paper to submit to arXiv
High-quality multi-lingual text-to-speech library by MyShell.ai. Support English, Spanish, French, Chinese, Japanese and Korean.
Qwen2-VL is the multimodal large language model series developed by Qwen team, Alibaba Cloud.
Qwen2.5-Coder is the code version of Qwen2.5, the large language model series developed by Qwen team, Alibaba Cloud.
Code and models for NExT-GPT: Any-to-Any Multimodal Large Language Model
🚀 Efficient implementations of state-of-the-art linear attention models in Pytorch and Triton
Learning Continuous Image Representation with Local Implicit Image Function, in CVPR 2021 (Oral)
Cool vision, learning, and graphics papers on Cats!
Official PyTorch implementation of Learning to (Learn at Test Time): RNNs with Expressive Hidden States