Stars
A natural language interface for computers
Platform to experiment with the AI Software Engineer. Terminal based. NOTE: Very different from https://gptengineer.app
LlamaIndex is the leading framework for building LLM-powered agents over your data.
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
RAGFlow is an open-source RAG (Retrieval-Augmented Generation) engine based on deep document understanding.
☁️ Build multimodal AI applications with cloud-native stack
Chat with your documents on your local device using GPT models. No data leaves your device and 100% private.
Python SDK, Proxy Server (LLM Gateway) to call 100+ LLM APIs in OpenAI format - [Bedrock, Azure, OpenAI, VertexAI, Cohere, Anthropic, Sagemaker, HuggingFace, Replicate, Groq]
Scalene: a high-performance, high-precision CPU, GPU, and memory profiler for Python with AI-powered optimization proposals
Large Language Model Text Generation Inference
Implementation of Nougat Neural Optical Understanding for Academic Documents
Unified framework for building enterprise RAG pipelines with small, specialized models
Supercharge Your LLM Application Evaluations 🚀
Unified embedding generation and search engine. Also available on cloud - cloud.marqo.ai
Adding guardrails to large language models.
NeMo Guardrails is an open-source toolkit for easily adding programmable guardrails to LLM-based conversational systems.
A fast inference library for running LLMs locally on modern consumer-class GPUs
Interact with your SQL database, Natural Language to SQL using LLMs
A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.
Open-source tools for prompt testing and experimentation, with support for both LLMs (e.g. OpenAI, LLaMA) and vector databases (e.g. Chroma, Weaviate, LanceDB).
mPLUG-Owl: The Powerful Multi-modal Large Language Model Family
An optimized deep prompt tuning strategy comparable to fine-tuning across scales and tasks
Running Llama 2 and other Open-Source LLMs on CPU Inference Locally for Document Q&A
Forward-Looking Active REtrieval-augmented generation (FLARE)