Highlights
- Pro
Stars
The simplest, fastest repository for training/finetuning medium-sized GPTs.
Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.
Freeing data processing from scripting madness by providing a set of platform-agnostic customizable pipeline processing blocks.
Meditron is a suite of open-source medical Large Language Models (LLMs).
Large-scale linear classification, regression and ranking in Python
Reference implementations of MLPerf™ training benchmarks
PMLB: A large, curated repository of benchmark datasets for evaluating supervised machine learning algorithms.
Convolutional Neural Networks for Sentence Classification in Keras
Landmark Attention: Random-Access Infinite Context Length for Transformers
Example code and applications for machine learning on Graphcore IPUs
Stochastic Gradient Push for Distributed Deep Learning
Practical low-rank gradient compression for distributed optimization: https://arxiv.org/abs/1905.13727
Accelerating your LLM training to full speed! Made with ❤️ by ServiceNow Research
Language Identification with Support for More Than 2000 Labels -- EMNLP 2023
Code for NeurIPS 2024 Spotlight: "Scaling Laws and Compute-Optimal Training Beyond Fixed Training Durations"
Code for WWW 2017 conference paper "Leveraging large amounts of weakly supervised data for multi-language sentiment classification"
CoLa - Decentralized Linear Learning: https://arxiv.org/abs/1808.04883
Codebase for ICML submission "DOGE: Domain Reweighting with Generalization Estimation"
Open Challenge - Automatic Training for Deep Learning