Skip to content

In-context KV-Cache Eviction for LLMs via Attention-Gate

Notifications You must be signed in to change notification settings

CengZihao/Attention-Gate

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Code for Attention-Gate.

The current version is an early-stage implementation, and further improvements will be made in future updates.

The training script replaces modeling_llama_with_ag.py in the transformers library’s modeling_llama.py.

Continual pre-training uses the dataset redpajama_samples.jsonl.

About

In-context KV-Cache Eviction for LLMs via Attention-Gate

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages