Skip to content
View nairouz's full-sized avatar
💭
I may be slow to respond.
💭
I may be slow to respond.

Block or report nairouz

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
Showing results

Code for the ICML 2023 paper "SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot".

Python 760 99 Updated Aug 20, 2024

source code for paper "Riemannian Preconditioned LoRA for Fine-Tuning Foundation Models"

Python 20 2 Updated Jun 20, 2024

Official code for our paper, "LoRA-Pro: Are Low-Rank Adapters Properly Optimized? "

Python 95 5 Updated Oct 24, 2024

[NeurIPS 2024] BAdam: A Memory Efficient Full Parameter Optimization Method for Large Language Models

Python 228 13 Updated Nov 30, 2024

A library for parameter-efficient and composable transfer learning for NLP with sparse fine-tunings.

Python 71 11 Updated Aug 9, 2024

Official implementation of the paper Connecting Large Language Models with Evolutionary Algorithms Yields Powerful Prompt Optimizers

Python 120 18 Updated May 29, 2024
Python 16 5 Updated Jul 5, 2024

Parameter Efficient Transfer Learning with Diff Pruning

Python 73 9 Updated Feb 3, 2021

Code for the paper: "SuS-X: Training-Free Name-Only Transfer of Vision-Language Models" [ICCV'23]

Python 96 5 Updated Aug 22, 2023

Code for "Training Neural Networks with Fixed Sparse Masks" (NeurIPS 2021).

Python 58 3 Updated Jan 14, 2022

[ICLR'24] "DeepZero: Scaling up Zeroth-Order Optimization for Deep Model Training" by Aochuan Chen*, Yimeng Zhang*, Jinghan Jia, James Diffenderfer, Jiancheng Liu, Konstantinos Parasyris, Yihua Zha…

Python 49 6 Updated Oct 9, 2024

LOMO: LOw-Memory Optimization

Python 979 69 Updated Jul 2, 2024

This repository contains code for the MicroAdam paper.

Python 15 4 Updated Dec 14, 2024

[NeurIPS 2023] MeZO: Fine-Tuning Language Models with Just Forward Passes. https://arxiv.org/abs/2305.17333

Python 1,075 67 Updated Jan 11, 2024
Python 188 53 Updated Jan 16, 2021

SLTrain: a sparse plus low-rank approach for parameter and memory efficient pretraining (NeurIPS 2024)

Python 27 1 Updated Nov 1, 2024

An Efficient LLM Fine-Tuning Factory Optimized for MoE PEFT

Python 61 7 Updated Dec 31, 2024

Code for Adam-mini: Use Fewer Learning Rates To Gain More https://arxiv.org/abs/2406.16793

Python 376 14 Updated Dec 5, 2024

A family of compressed models obtained via pruning and knowledge distillation

309 18 Updated Nov 13, 2024

[NeurIPS 2023] LLM-Pruner: On the Structural Pruning of Large Language Models. Support Llama-3/3.1, Llama-2, LLaMA, BLOOM, Vicuna, Baichuan, TinyLlama, etc.

Python 931 109 Updated Oct 7, 2024

Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.

Python 188 14 Updated Jul 17, 2024

GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection

Python 1,481 154 Updated Oct 28, 2024

Fira: Can We Achieve Full-rank Training of LLMs Under Low-rank Constraint?

Python 89 3 Updated Oct 21, 2024

torch-optimizer -- collection of optimizers for Pytorch

Python 3,068 299 Updated Mar 22, 2024

Code for NOLA, an implementation of "nola: Compressing LoRA using Linear Combination of Random Basis"

Python 50 3 Updated Aug 25, 2024
Python 49 6 Updated Oct 17, 2023
Shell 26 2 Updated Jun 11, 2024

CorDA: Context-Oriented Decomposition Adaptation of Large Language Models for task-aware parameter-efficient fine-tuning(NeurIPS 2024)

Python 41 1 Updated Jan 13, 2025
Next