Stars
Most popular & widely deployed Open Source Container Native Storage platform for Stateful Persistent Applications on Kubernetes.
A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)
Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficie…
Prompt Generator is a flexible and user-friendly package that offers customizable scripts for generating meaningful and context-aware prompts. These prompts can be used to guide the writing, enhanc…
AI Agent for Twitter Personality Analysis
A custom extension for sd-webui that allow you to generate 3D model from txt or image, basing on OpenAI Shap-E.
Text-to-3D & Image-to-3D & Mesh Exportation with NeRF + Diffusion.
Ongoing research training transformer language models at scale, including: BERT & GPT-2
A collection of NVIDIA cuOpt samples and other resources
Install Kubernetes/K3s only, both Kubernetes/K3s and KubeSphere, and related cloud-native add-ons, it supports all-in-one, multi-node, and HA 🔥 ⎈ 🐳
CV-CUDA™ is an open-source, GPU accelerated library for cloud-scale image processing and computer vision.
DeepSeek Coder: Let the Code Write Itself
NVIDIA GPU Operator creates, configures, and manages GPUs in Kubernetes
User-friendly AI Interface (Supports Ollama, OpenAI API, ...)
基于Clash Core 制作的Clash For Linux备份仓库 A Clash For Linux Backup Warehouse Based on Clash Core
Reference implementations of MLPerf™ inference benchmarks
AutoDL平台服务器适配梯子, 使用 Clash 作为代理工具
Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with …
Making large AI models cheaper, faster and more accessible
real Transformer TeraFLOPS on various GPUs
Development repository for the Triton language and compiler
21 Lessons, Get Started Building with Generative AI 🔗 https://microsoft.github.io/generative-ai-for-beginners/
Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.
NVIDIA Data Center GPU Manager (DCGM) is a project for gathering telemetry and measuring the health of NVIDIA GPUs