Stars
Implementation of E2-TTS, "Embarrassingly Easy Fully Non-Autoregressive Zero-Shot TTS", in Pytorch
Repo for paper "Unleashing Cognitive Synergy in Large Language Models: A Task-Solving Agent through Multi-Persona Self-Collaboration"
# Prompt Engineering Hub ⭐️ If you find this helpful, give it a star to show your support! This repository is a one-stop resource for prompt engineering. Also available on: https://promptengineerin…
Morphotensorial Field Theory (MFT): A mathematical framework unifying complex tensor operations, quantum fields, and scale-invariant transformations. Introduces the Morphing Infinity Spiral transfo…
SuperPrompt is an attempt to engineer prompts that might help us understand AI agents.
Supplementary material for the EMNLP 2024 paper "Conditional and Modal Reasoning in Large Language Models" by Wesley H. Holliday, Matthew Mandelkern, and Cedegao E. Zhang
JupyterLab for AI in Docker! Anaconda and PyTorch GPU supported.
AdaLoRA: Adaptive Budget Allocation for Parameter-Efficient Fine-Tuning (ICLR 2023).
Reasoning in LLMs: Papers and Resources, including Chain-of-Thought, OpenAI o1, and DeepSeek-R1 🍓
Papers and resources on Controllable Generation using Diffusion Models, including ControlNet, DreamBooth, IP-Adapter.
An implementation of local windowed attention for language modeling
[ICLR 2024] Efficient Streaming Language Models with Attention Sinks
Extend existing LLMs way beyond the original training length with constant memory usage, without retraining
Diffusion Reading Group at EleutherAI
Fast Hadamard transform in CUDA, with a PyTorch interface
Efficient GPU kernels for block-sparse matrix multiplication and convolution
🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support
Official repository of Agent Attention (ECCV2024)
Transformer based on a variant of attention that is linear complexity in respect to sequence length
Awesome list for LLM quantization
Code repo for the paper "SpinQuant LLM quantization with learned rotations"
Samples for CUDA Developers which demonstrates features in CUDA Toolkit
A list of papers, blogs, datasets and software in the field of lifelong/continual machine learning
A PyTorch implementation of the Transformer model in "Attention is All You Need".
Code for the article "What if Neural Networks had SVDs?", to be presented as a spotlight paper at NeurIPS 2020.