Skip to content
View dunchen's full-sized avatar
  • Rice Univerisity
  • Houston, TX

Block or report dunchen

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
328 results for source starred repositories
Clear filter

🏭 Chernobyl Nuclear Power Plant Simulator

NSIS 161 7 Updated Dec 5, 2023

Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.

Python 6,021 518 Updated Sep 6, 2024

BERT score for text generation

Jupyter Notebook 1,652 223 Updated Jul 30, 2024

An automatic evaluator for instruction-following language models. Human-validated, high-quality, cheap, and fast.

Jupyter Notebook 1,609 248 Updated Dec 27, 2024

🐙 Guides, papers, lecture, notebooks and resources for prompt engineering

MDX 52,446 5,096 Updated Jan 9, 2025

4 bits quantization of LLaMA using GPTQ

Python 3,026 461 Updated Jul 13, 2024

🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.

Python 16,979 1,682 Updated Jan 15, 2025

Code and documentation to train Stanford's Alpaca models, and generate the data.

Python 29,741 4,056 Updated Jul 17, 2024

LLaMA: Open and Efficient Foundation Language Models

Python 2,806 309 Updated Nov 8, 2023

Inference code for Llama models

Python 57,216 9,656 Updated Aug 18, 2024

Benchmark API for Multidomain Language Modeling

Python 24 3 Updated Aug 26, 2022

Progressive Prompts: Continual Learning for Language Models

Python 91 12 Updated Apr 24, 2023

Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch

Python 21,460 3,147 Updated Jan 4, 2025

Vision Transformer for 3D medical image registration (Pytorch).

Python 308 43 Updated Nov 6, 2022

To eventually become an unofficial Pytorch implementation / replication of Alphafold2, as details of the architecture get released

Python 1,581 261 Updated Oct 29, 2022

Pytorch reimplementation of the Vision Transformer (An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale)

Jupyter Notebook 1,978 379 Updated Jun 7, 2022

刷算法全靠套路,认准 labuladong 就够了!English version supported! Crack LeetCode, not only how, but also why.

Markdown 126,513 23,287 Updated Sep 22, 2024

Transformer seq2seq model, program that can build a language translator from parallel corpus

Python 1,368 349 Updated May 19, 2023

A small package to create visualizations of PyTorch execution graphs

Jupyter Notebook 3,270 280 Updated Dec 30, 2024

Leaf: A Benchmark for Federated Settings

Python 862 247 Updated Mar 24, 2023

Tutorials on getting started with PyTorch and TorchText for sentiment analysis.

Jupyter Notebook 4,434 1,174 Updated Mar 27, 2024

Python .mu reader/writer and blender import/export addon

Python 122 52 Updated Aug 9, 2024

Practice on cifar100(ResNet, DenseNet, VGG, GoogleNet, InceptionV3, InceptionV4, Inception-ResNetv2, Xception, Resnet In Resnet, ResNext,ShuffleNet, ShuffleNetv2, MobileNet, MobileNetv2, SqueezeNet…

Python 4,368 1,187 Updated Jul 15, 2024

A Pytorch implementation of Sparsely-Gated Mixture of Experts, for massively increasing the parameter count of language models

Python 668 50 Updated Sep 13, 2023

DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.

Python 36,242 4,192 Updated Jan 15, 2025

DCS World dynamic campaign.

Python 732 184 Updated Jan 13, 2025

PyTorch implementation of Mixer-nano (#parameters is 0.67M, originally Mixer-S/16 has 18M) with 90.83 % acc. on CIFAR-10. Training from scratch.

Python 29 6 Updated Nov 6, 2021

Implements MLP-Mixer (https://arxiv.org/abs/2105.01601) with the CIFAR-10 dataset.

Jupyter Notebook 54 7 Updated Jun 26, 2022

Let's train vision transformers (ViT) for cifar 10!

Python 573 113 Updated Dec 13, 2024

Implementation of Memorizing Transformers (ICLR 2022), attention net augmented with indexing and retrieval of memories using approximate nearest neighbors, in Pytorch

Python 628 46 Updated Jul 17, 2023
Next