-
ByteDance
- Shanghai, China
- https://de0ny6i65l.feishu.cn/wiki/FwECwhwBCiGdmJkjksJcUw3dnVe
- @flyer103
- All languages
- Arc
- Assembly
- Batchfile
- C
- C#
- C++
- CMake
- CSS
- Clojure
- CoffeeScript
- Common Lisp
- Cuda
- Dhall
- Dockerfile
- Elixir
- Emacs Lisp
- Gherkin
- Go
- Groff
- Groovy
- HTML
- Hack
- Handlebars
- Haskell
- Java
- JavaScript
- Jinja
- Jsonnet
- Jupyter Notebook
- Kotlin
- Lua
- MDX
- Makefile
- Markdown
- PHP
- PLpgSQL
- Perl
- Processing
- Puppet
- Python
- R
- Rich Text Format
- Roff
- Ruby
- Rust
- SCSS
- Scala
- Scheme
- Shell
- Smarty
- Starlark
- Swift
- TLA
- TSQL
- TypeScript
- WebAssembly
- Zig
Starred repositories
A flexible commandline tool for template rendering. Supports lots of local and remote datasources.
Official code repository for GATK versions 4 and up
An interactive NVIDIA-GPU process viewer and beyond, the one-stop solution for GPU process management.
GPU & Accelerator process monitoring for AMD, Apple, Huawei, Intel, NVIDIA and Qualcomm
LLM-powered multiagent persona simulation for imagination enhancement and business insights.
An AI search engine inspired by Perplexity
Database diagrams editor that allows you to visualize and design your DB with a single query.
Hyperlight is a lightweight Virtual Machine Manager (VMM) designed to be embedded within applications. It enables safe execution of untrusted code within micro virtual machines with very low latenc…
KV cache compression for high-throughput LLM inference
The fastest knowledge base for growing teams. Beautiful, realtime collaborative, feature packed, and markdown compatible.
Mooncake is the serving platform for Kimi, a leading LLM service provided by Moonshot AI.
A Flexible Framework for Experiencing Cutting-edge LLM Inference Optimizations
Python packaging and dependency management made easy
A collection of notebooks/recipes showcasing usecases of open-source models with Together AI.
Awesome-LLM-KV-Cache: A curated list of 📙Awesome LLM KV Cache Papers with Codes.
Automated resources sizing tool for containers in kubernetes
Finetune Llama 3.3, Mistral, Phi, Qwen 2.5 & Gemma LLMs 2-5x faster with 70% less memory
The AI-native database built for LLM applications, providing incredibly fast hybrid search of dense vector, sparse vector, tensor (multi-vector), and full-text
Meta Lingua: a lean, efficient, and easy-to-hack codebase to research LLMs.
Backend API for Comfy Registry + CI Dashboard
Official front-end implementation of ComfyUI
Gateway API Inference Extension
Lightweight and portable LLM sandbox runtime (code interpreter) Python library.
Educational framework exploring ergonomic, lightweight multi-agent orchestration. Managed by OpenAI Solution team.
Get up and running with Llama 3.3, Mistral, Gemma 2, and other large language models.