Skip to content
View dunchen's full-sized avatar
  • Rice Univerisity
  • Houston, TX

Block or report dunchen

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
327 results for source starred repositories
Clear filter

🏭 Chernobyl Nuclear Power Plant Simulator

NSIS 166 8 Updated Dec 5, 2023

Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.

Python 6,036 520 Updated Sep 6, 2024

BERT score for text generation

Jupyter Notebook 1,688 225 Updated Jul 30, 2024

An automatic evaluator for instruction-following language models. Human-validated, high-quality, cheap, and fast.

Jupyter Notebook 1,674 259 Updated Dec 27, 2024

🐙 Guides, papers, lecture, notebooks and resources for prompt engineering

MDX 53,913 5,261 Updated Jan 21, 2025

4 bits quantization of LLaMA using GPTQ

Python 3,042 461 Updated Jul 13, 2024

🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.

Python 17,649 1,770 Updated Mar 6, 2025

Code and documentation to train Stanford's Alpaca models, and generate the data.

Python 29,872 4,059 Updated Jul 17, 2024

LLaMA: Open and Efficient Foundation Language Models

Python 2,803 309 Updated Nov 8, 2023

Inference code for Llama models

Python 57,812 9,716 Updated Jan 26, 2025

Benchmark API for Multidomain Language Modeling

Python 24 3 Updated Aug 26, 2022

Progressive Prompts: Continual Learning for Language Models

Python 92 12 Updated Apr 24, 2023

Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch

Python 21,960 3,191 Updated Mar 5, 2025

Vision Transformer for 3D medical image registration (Pytorch)

Python 311 44 Updated Nov 6, 2022

To eventually become an unofficial Pytorch implementation / replication of Alphafold2, as details of the architecture get released

Python 1,596 263 Updated Oct 29, 2022

Pytorch reimplementation of the Vision Transformer (An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale)

Jupyter Notebook 2,002 390 Updated Jun 7, 2022

刷算法全靠套路,认准 labuladong 就够了!English version supported! Crack LeetCode, not only how, but also why.

Markdown 127,092 23,316 Updated Jan 31, 2025

Transformer seq2seq model, program that can build a language translator from parallel corpus

Python 1,381 349 Updated May 19, 2023

A small package to create visualizations of PyTorch execution graphs

Jupyter Notebook 3,309 282 Updated Dec 30, 2024

Leaf: A Benchmark for Federated Settings

Python 872 247 Updated Mar 24, 2023

Tutorials on getting started with PyTorch and TorchText for sentiment analysis.

Jupyter Notebook 4,463 1,179 Updated Mar 27, 2024

Python .mu reader/writer and blender import/export addon

Python 123 52 Updated Aug 9, 2024

Practice on cifar100(ResNet, DenseNet, VGG, GoogleNet, InceptionV3, InceptionV4, Inception-ResNetv2, Xception, Resnet In Resnet, ResNext,ShuffleNet, ShuffleNetv2, MobileNet, MobileNetv2, SqueezeNet…

Python 4,427 1,190 Updated Jul 15, 2024

A Pytorch implementation of Sparsely-Gated Mixture of Experts, for massively increasing the parameter count of language models

Python 699 54 Updated Sep 13, 2023

DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.

Python 37,239 4,281 Updated Mar 6, 2025

DCS World dynamic campaign.

Python 740 183 Updated Mar 6, 2025

PyTorch implementation of Mixer-nano (#parameters is 0.67M, originally Mixer-S/16 has 18M) with 90.83 % acc. on CIFAR-10. Training from scratch.

Python 30 6 Updated Nov 6, 2021

Implements MLP-Mixer (https://arxiv.org/abs/2105.01601) with the CIFAR-10 dataset.

Jupyter Notebook 55 7 Updated Jun 26, 2022

Let's train vision transformers (ViT) for cifar 10!

Python 594 116 Updated Jan 26, 2025

Implementation of Memorizing Transformers (ICLR 2022), attention net augmented with indexing and retrieval of memories using approximate nearest neighbors, in Pytorch

Python 631 46 Updated Jul 17, 2023
Next