Stars
OpenRFT: Adapting Reasoning Foundation Model for Domain-specific Tasks with Reinforcement Fine-Tuning
✨✨Latest Papers and Benchmarks in Reasoning with Foundation Models
[NeurIPS 2024] SimPO: Simple Preference Optimization with a Reference-Free Reward
Reference implementation for DPO (Direct Preference Optimization)
Implementation for "Step-DPO: Step-wise Preference Optimization for Long-chain Reasoning of LLMs"
RUCAIBox / RLMEC
Forked from Timothy023/RLMECThe official repository of "Improving Large Language Models via Fine-grained Reinforcement Learning with Minimum Editing Constraint"
The Open Cookbook for Top-Tier Code Large Language Model
A large-scale, fine-grained, diverse preference dataset (and models).
Baselines for all tasks from Long Code Arena benchmarks 🏟️
This repository collects papers for "A Survey on Knowledge Distillation of Large Language Models". We break down KD into Knowledge Elicitation and Distillation Algorithms, and explore the Skill & V…
👨🎓 北京交通大学计算机科学与技术学院研究生课程资料、笔记、回忆和整理的期末考试卷及课程作业。希望对你们有所帮助❤️,如果喜欢记得给个star🌟
Recipes to train reward model for RLHF.
A recipe for online RLHF and online iterative DPO.
强化学习中文教程(蘑菇书🍄),在线阅读地址:https://datawhalechina.github.io/easy-rl/
[TMLR] A curated list of language modeling researches for code (and other software engineering activities), plus related datasets.
BeHonest: Benchmarking Honesty in Large Language Models
Awesome LLMs on Device: A Comprehensive Survey
TinyChatEngine: On-Device LLM Inference Library
Representation Engineering: A Top-Down Approach to AI Transparency
A curated list of safety-related papers, articles, and resources focused on Large Language Models (LLMs). This repository aims to provide researchers, practitioners, and enthusiasts with insights i…
[ACL 2024] User-friendly evaluation framework: Eval Suite & Benchmarks: UHGEval, HaluEval, HalluQA, etc.
The code of “Improving Weak-to-Strong Generalization with Scalable Oversight and Ensemble Learning”