Skip to content

This model is finetune bangla LLAMA(LLM) model. We train this model using LORA approach

Notifications You must be signed in to change notification settings

saiful9379/Bangla_LLAMA

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Bangla_LLAMA

Llama is a family of open-source large language models released by Meta. we use for the Bangla Large language model to fine-tune. Llama models come in two flavors — pre-trained(From OdiaGenAI) and fine-tuned for our custom dataset. While the latter is typically used for Bangla general-purpose chat use cases, the former can be used as a foundation to be further fine-tuned for a specific use case.

Requirements

pip install -r requirements.txt

or

datasets==2.14.5
sentencepiece==0.1.99
transformers==4.34.0.dev0
bitsandbytes==0.41.1
loralib
peft==0.6.0.dev0

or [lates require module]

pip install -q bitsandbytes
pip install -q datasets loralib sentencepiece
pip install -q git+https://github.com/huggingface/transformers.git
pip install -q git+https://github.com/huggingface/peft.git

Data structure

Format the data structure into JSON Format, Following the below structure,

[
    {
        "instruction": "হাই! কেমন চলছে?",
        "input": "",
        "output": "আমি ভালো আছি. তোমার কি অবস্থা?"
    },
    .,
    .,
    .,
    {
        "instruction": "তুমি কোন স্কুলে যাও?",
        "input": "",
        "output": "আমি পিসিসিতে যাই।"
    }
]

Configuration

check the configuration for config.py and set your configuration.

Traning

Run the train.py file

python train.py

Interactive notebook file

./examples/Bangla_llama_Lora_finetune_final.ipynb

Evaluation

Not Yet Done

Infenence

Inference interactive notebook file,

./examples/Bangla_llama_lora_inference.ipynb

References

1. https://github.com/tatsu-lab/stanford_alpaca
2. https://colab.research.google.com/drive/1eWAmesrW99p7e1nah5bipn0zikMb8XYC
3. https://huggingface.co/OdiaGenAI
4. https://huggingface.co/docs/diffusers/training/lora

About

This model is finetune bangla LLAMA(LLM) model. We train this model using LORA approach

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages