Lists (1)
Sort Name ascending (A-Z)
Stars
[ICLR 2025] Dynamic Mixture of Experts: An Auto-Tuning Approach for Efficient Transformer Models
Inference Code for Paper "Harder Tasks Need More Experts: Dynamic Routing in MoE Models"
The official GitHub page for the survey paper "A Survey on Mixture of Experts in Large Language Models".
A curated list for Efficient Large Language Models
Chain of Experts (CoE) enables communication between experts within Mixture-of-Experts (MoE) models
Official inference repo for FLUX.1 models
Proxy to enable P2P only cameras to work with standard protocols.
Official PyTorch implementation of Learning to (Learn at Test Time): RNNs with Expressive Hidden States
Production-tested AI infrastructure tools for efficient AGI development and community-driven innovation
🌍 Repository for "AppWorld: A Controllable World of Apps and People for Benchmarking Interactive Coding Agent", ACL'24 Best Resource Paper.
TL-Training: A Task-Feature-Based Framework for Training Large Language Models in Tool Use
[NeurIPS 2024] Classification Done Right for Vision-Language Pre-Training
Synthetic question-answering dataset to formally analyze the chain-of-thought output of large language models on a reasoning task.
Official codebase for the paper "Beyond A* Better Planning with Transformers via Search Dynamics Bootstrapping".
💡 LeetCode in C++20/Java/Python/MySQL/TypeScript (respect coding conventions)
OCR, layout analysis, reading order, table recognition in 90+ languages
verl: Volcano Engine Reinforcement Learning for LLMs
星辰语义大模型TeleChat2是由中国电信人工智能研究院研发训练的大语言模型,是首个完全国产算力训练并开源的千亿参数模型
[NeurIPS 2023 Spotlight] LightZero: A Unified Benchmark for Monte Carlo Tree Search in General Sequential Decision Scenarios (awesome MCTS)
[ICML'24 Spotlight] "TravelPlanner: A Benchmark for Real-World Planning with Language Agents"
OK影视、tvbox配置文件,如果喜欢,请Fork自用。使用前请仔细阅读仓库说明,一旦使用将被视为你已了解。