Lists (11)
Sort Name ascending (A-Z)
Starred repositories
Supercharge Your LLM Application Evaluations 🚀
GLM-4 series: Open Multilingual Multimodal Chat LMs | 开源多语言多模态对话模型
整理开源的中文大语言模型,以规模较小、可私有化部署、训练成本较低的模型为主,包括底座模型,垂直领域微调及应用,数据集与教程等。
Qwen2.5 is the large language model series developed by Qwen team, Alibaba Cloud.
The official repo of Qwen (通义千问) chat & pretrained large language model proposed by Alibaba Cloud.
Chinese-Vicuna: A Chinese Instruction-following LLaMA-based Model —— 一个中文低资源的llama+lora方案,结构参考alpaca
Evaluating LLMs' multi-round chatting capability via assessing conversations generated by two LLM instances.
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
abditag2 / bert
Forked from google-research/bertTensorFlow code and pre-trained models for BERT
Retrieval and Retrieval-augmented LLMs
Chinese version of CLIP which achieves Chinese cross-modal retrieval and representation generation.
🎯 Task-oriented embedding tuning for BERT, CLIP, etc.
WordMultiSenseDisambiguation, chinese multi-wordsense disambiguation based on online bake knowledge base and semantic embedding similarity compute,基于百科知识库的中文词语多词义/义项获取与特定句子词语语义消歧.
Code repository for supporting the paper "Atlas Few-shot Learning with Retrieval Augmented Language Models",(https//arxiv.org/abs/2208.03299)
This is the homepage of a new book entitled "Mathematical Foundations of Reinforcement Learning."
Code for the paper Fine-Tuning Language Models from Human Preferences
Train transformer language models with reinforcement learning.
A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)
A Gradio web UI for Large Language Models with support for multiple inference backends.
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
ChatGLM2-6B: An Open Bilingual Chat LLM | 开源双语对话语言模型