![:octocat: :octocat:](https://github.githubassets.com/images/icons/emoji/octocat.png)
-
Capital One AI Foundations
- New York
- https://gentawinata.com
- @gentaiscool
Highlights
- Pro
Stars
Learn how to design large-scale systems. Prep for the system design interview. Includes Anki flashcards.
π€ Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
Magnificent app which corrects your previous console command.
Tensors and Dynamic neural networks in Python with strong GPU acceleration
Models and examples built with TensorFlow
TensorFlow code and pre-trained models for BERT
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Facebook AI Research Sequence-to-Sequence Toolkit written in Python.
Pretrain, finetune ANY AI model of ANY size on multiple GPUs, TPUs with zero code changes.
Repository to track the progress in Natural Language Processing (NLP), including the datasets and the current state-of-the-art for the most common NLP tasks.
A set of examples around pytorch in Vision, Text, Reinforcement Learning, etc.
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
π€ The largest hub of ready-to-use datasets for ML models with fast, easy-to-use and efficient data manipulation tools
Fully open reproduction of DeepSeek-R1
Open source code for AlphaFold 2.
A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)
π Scalable embedding, reasoning, ranking for images and sentences with CLIP
An open-source NLP research library, built on PyTorch.
Ongoing research training transformer models at scale
A PyTorch Extension: Tools for easy mixed precision and distributed training in Pytorch
π A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support
An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library.
A Python implementation of global optimization with gaussian processes.
Chinese version of GPT2 training code, using BERT tokenizer.
An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries
Open Source Neural Machine Translation and (Large) Language Models in PyTorch
A natural language modeling framework based on PyTorch