Stars
OpenLLaMA, a permissively licensed open source reproduction of Meta AI’s LLaMA 7B trained on the RedPajama dataset
Official implementation of paper "Meta Prompting for AI Systems" (https://arxiv.org/abs/2311.11482)
WIP - Automated Question Answering for ArXiv Papers with Large Language Models (https://arxiv.taesiri.xyz/)
The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.
Resources of deep learning for mathematical reasoning (DL4MATH).
MNBVC(Massive Never-ending BT Vast Chinese corpus)超大规模中文语料集。对标chatGPT训练的40T数据。MNBVC数据集不但包括主流文化,也包括各个小众文化甚至火星文的数据。MNBVC数据集包括新闻、作文、小说、书籍、杂志、论文、台词、帖子、wiki、古诗、歌词、商品介绍、笑话、糗事、聊天记录等一切形式的纯文本中文数据。
DSPy: The framework for programming—not prompting—language models
Industry leading face manipulation platform
A family of open-sourced Mixture-of-Experts (MoE) Large Language Models
Task-based datasets, preprocessing, and evaluation for sequence models.
Multipack distributed sampler for fast padding-free training of LLMs
[ACL 2023] Gradient Ascent Post-training Enhances Language Model Generalization
Pipeline for pulling and processing online language model pretraining data from the web
The all-in-one Desktop & Docker AI application with built-in RAG, AI agents, and more.
🔥Highlighting the top ML papers every week.
MTEB: Massive Text Embedding Benchmark
pix2tex: Using a ViT to convert images of equations into LaTeX code.
Awasome Papers and Resources in Deep Neural Network Pruning with Source Code.
A list of papers, docs, codes about model quantization. This repo is aimed to provide the info for model quantization research, we are continuously improving the project. Welcome to PR the works (p…
中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)
AGiXT is a dynamic AI Agent Automation Platform that seamlessly orchestrates instruction management and complex task execution across diverse AI providers. Combining adaptive memory, smart features…
Using GPT to organize and access information, and generate questions. Long term goal is to make an agent-like research assistant.
Safe RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback
Benchmarking large language models' complex reasoning ability with chain-of-thought prompting