Skip to content

Commit

Permalink
update README.md
Browse files Browse the repository at this point in the history
  • Loading branch information
shadowpa0327 committed Aug 1, 2024
1 parent 7509f40 commit dee2d8e
Showing 1 changed file with 2 additions and 2 deletions.
4 changes: 2 additions & 2 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -80,6 +80,8 @@ python run_ppl_eval.py \
--lt_hadamard
```

*Note*: `run_ppl_eval.py` hasn't suppoted multi-gpu evaluation yet. If your machine have multiple GPUs, please set `CUDA_VISIBLE_DEVICES` to the desired GPU id.

#### Zero-shot Evaluation
To run zero-shot evaluations of models with compressed KV-Cache, we can use the `run_lm_eval.py` script, which implement a wrapper around the [lm-evaluation-harness](https://github.com/EleutherAI/lm-evaluation-harness/tree/big-refactor) library.

Expand Down Expand Up @@ -131,8 +133,6 @@ CUDA_VISIBLE_DEVICES=0 python run_latency_kernel.py \
--total_rank 1024 --group_size 4
```



## Reference
If you find this work useful, please consider citing our paper:
```
Expand Down

0 comments on commit dee2d8e

Please sign in to comment.