Skip to content

moeiniamir/llm-foundry

 
 

Repository files navigation

Usage:

Having one of the following sample scripts in a method.py file:

from llmfoundry.eval.eval_ext import evaluate
from transformers import AutoModelForCausalLM

model = AutoModelForCausalLM.from_pretrained("ISTA-DASLab/Llama-3-8B-Instruct-GPTQ-4bit", device_map="auto")
evaluate(
    model, 
    "/path/to/hf_eval.yaml", 
    '/path/to/llm-foundry/scripts')
from llmfoundry.eval.eval_ext import evaluate
from hqq.engine.hf import HQQModelForCausalLM
import sys

model = HQQModelForCausalLM.from_quantized("PrunaAI/meta-llama-Meta-Llama-3-8B-Instruct-HQQ-4bit-smashed")
evaluate(
    model, 
    "/path/to/hf_eval.yaml", 
    '/path/to/llm-foundry/scripts',
    tokenizer_name='PrunaAI/meta-llama-Meta-Llama-3-8B-Instruct-HQQ-4bit-smashed', config_overrides=sys.argv[1:])

We then run the following command:

composer method.py

About

Quantization enabled llm-foundry

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 97.8%
  • Shell 2.1%
  • Other 0.1%