Lists (1)
Sort Name ascending (A-Z)
Stars
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
The fundamental package for scientific computing with Python.
Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"
Perform data science on data that remains in someone else's server
An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All.
Official repo for consistency models.
Modeling, training, eval, and inference code for OLMo
H2O LLM Studio - a framework and no-code GUI for fine-tuning LLMs. Documentation: https://docs.h2o.ai/h2o-llmstudio/
[ICLR 2020] Once for All: Train One Network and Specialize it for Efficient Deployment
37 traditional FL (tFL) or personalized FL (pFL) algorithms, 3 scenarios, and 24 datasets. www.pfllib.com/
A curated list for Efficient Large Language Models
Implementation of 🦩 Flamingo, state-of-the-art few-shot visual question answering attention net out of Deepmind, in Pytorch
[ICML 2024] Break the Sequential Dependency of LLM Inference Using Lookahead Decoding
Research on Tabular Deep Learning: Papers & Packages
A Supervised and Semi-Supervised Object Detection Library for YOLO Series
This is the repository of HaluEval, a large-scale hallucination evaluation benchmark for Large Language Models.
[CVPR23] Official Implementation of MIC: Masked Image Consistency for Context-Enhanced Domain Adaptation
HAKE: Human Activity Knowledge Engine (CVPR'18/19/20, NeurIPS'20, TPAMI'21)
Shepherd: A foundational framework enabling federated instruction tuning for large language models
Spec-Bench: A Comprehensive Benchmark and Unified Evaluation Platform for Speculative Decoding (ACL 2024 Findings)
Cross-silo Federated Learning playground in Python. Discover 7 real-world federated datasets to test your new FL strategies and try to beat the leaderboard.
[ICLR 2024] Skeleton-of-Thought: Prompting LLMs for Efficient Parallel Generation
Block Transformer: Global-to-Local Language Modeling for Fast Inference (NeurIPS 2024)