Skip to content
View dunchen's full-sized avatar
  • Rice Univerisity
  • Houston, TX

Block or report dunchen

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
325 results for source starred repositories
Clear filter

🏭 Chernobyl Nuclear Power Plant Simulator

NSIS 178 8 Updated Dec 5, 2023

Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.

Python 6,055 522 Updated Sep 6, 2024

BERT score for text generation

Jupyter Notebook 1,745 232 Updated Jul 30, 2024

An automatic evaluator for instruction-following language models. Human-validated, high-quality, cheap, and fast.

Jupyter Notebook 1,746 270 Updated Dec 27, 2024

🐙 Guides, papers, lecture, notebooks and resources for prompt engineering

MDX 56,168 5,551 Updated May 16, 2025

4 bits quantization of LLaMA using GPTQ

Python 3,050 459 Updated Jul 13, 2024

🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.

Python 18,491 1,873 Updated May 21, 2025

Code and documentation to train Stanford's Alpaca models, and generate the data.

Python 30,004 4,053 Updated Jul 17, 2024

LLaMA: Open and Efficient Foundation Language Models

Python 2,799 309 Updated Nov 8, 2023

Inference code for Llama models

Python 58,260 9,772 Updated Jan 26, 2025

Benchmark API for Multidomain Language Modeling

Python 24 3 Updated Aug 26, 2022

Progressive Prompts: Continual Learning for Language Models

Python 93 13 Updated Apr 24, 2023

Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch

Python 22,898 3,288 Updated Mar 5, 2025

Vision Transformer for 3D medical image registration (Pytorch)

Python 320 44 Updated Nov 6, 2022

To eventually become an unofficial Pytorch implementation / replication of Alphafold2, as details of the architecture get released

Python 1,601 264 Updated Oct 29, 2022

Pytorch reimplementation of the Vision Transformer (An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale)

Jupyter Notebook 2,033 398 Updated Jun 7, 2022

刷算法全靠套路,认准 labuladong 就够了!English version supported! Crack LeetCode, not only how, but also why.

Markdown 127,980 23,363 Updated Jan 31, 2025

Transformer seq2seq model, program that can build a language translator from parallel corpus

Python 1,397 351 Updated May 19, 2023

A small package to create visualizations of PyTorch execution graphs

Jupyter Notebook 3,360 285 Updated Dec 30, 2024

Leaf: A Benchmark for Federated Settings

Python 883 248 Updated Mar 24, 2023

Tutorials on getting started with PyTorch and TorchText for sentiment analysis.

Jupyter Notebook 4,510 1,182 Updated Mar 27, 2024

Python .mu reader/writer and blender import/export addon

Python 127 54 Updated Mar 7, 2025

Practice on cifar100(ResNet, DenseNet, VGG, GoogleNet, InceptionV3, InceptionV4, Inception-ResNetv2, Xception, Resnet In Resnet, ResNext,ShuffleNet, ShuffleNetv2, MobileNet, MobileNetv2, SqueezeNet…

Python 4,534 1,192 Updated Jul 15, 2024

A Pytorch implementation of Sparsely-Gated Mixture of Experts, for massively increasing the parameter count of language models

Python 752 60 Updated Sep 13, 2023

DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.

Python 38,510 4,383 Updated May 23, 2025

DCS World dynamic campaign.

Python 752 182 Updated May 4, 2025

PyTorch implementation of Mixer-nano (#parameters is 0.67M, originally Mixer-S/16 has 18M) with 90.83 % acc. on CIFAR-10. Training from scratch.

Python 32 6 Updated Nov 6, 2021

Implements MLP-Mixer (https://arxiv.org/abs/2105.01601) with the CIFAR-10 dataset.

Jupyter Notebook 56 7 Updated Jun 26, 2022

Let's train vision transformers (ViT) for cifar 10 / cifar 100!

Python 641 127 Updated Apr 28, 2025

Implementation of Memorizing Transformers (ICLR 2022), attention net augmented with indexing and retrieval of memories using approximate nearest neighbors, in Pytorch

Python 632 46 Updated Jul 17, 2023
Next