Stars
📚200+ Tensor/CUDA Cores Kernels, ⚡️flash-attn-mma, ⚡️hgemm with WMMA, MMA and CuTe (98%~100% TFLOPS of cuBLAS/FA2 🎉🎉).
开个新坑,从无名小卒到大模型(LLM)大英雄~ 欢迎关注后续!!!
🚀 Awesome System for Machine Learning ⚡️ AI System Papers and Industry Practice. ⚡️ System for Machine Learning, LLM (Large Language Model), GenAI (Generative AI). 🍻 OSDI, NSDI, SIGCOMM, SoCC, MLSy…
Repository hosting the code associated with "Enhancing MAP-Elites with Multiple Parallel Evolution Strategies".
🛰️ 基于真实医疗对话数据在ChatGLM上进行LoRA、P-Tuning V2、Freeze、RLHF等微调,我们的眼光不止于医疗问答
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
手把手带你实战 Huggingface Transformers 课程视频同步更新在B站与YouTube
Course to get into Large Language Models (LLMs) with roadmaps and Colab notebooks.
Attention is all you need implementation
A collection of LLM papers, blogs, and projects, with a focus on OpenAI o1 🍓 and reasoning techniques.
中文nlp解决方案(大模型、数据、模型、训练、推理)
中文翻译的 Hands-On-Large-Language-Models (hands-on-llms),动手学习大模型
Official code repo for the O'Reilly Book - "Hands-On Large Language Models"
记录计算智能优化算法的学习笔记,通过阅读论文并复现的形式加深对相关的启发式智能优化的理解。
Deep reinforcement learning approaches for CHP system economic dispatch
Development repository for the Triton language and compiler
AIFoundation 主要是指AI系统遇到大模型,从底层到上层如何系统级地支持大模型训练和推理,全栈的核心技术。
Advanced evolutionary computation library built directly on top of PyTorch, created at NNAISENSE.
Evolutionary Neural Architecture Search on Transformers for RUL Prediction
This is a collection of our NAS and Vision Transformer work.
Flax is a neural network library for JAX that is designed for flexibility.
Massively parallel rigidbody physics simulation on accelerator hardware.