|
|
||
|---|---|---|
| .. | ||
| README.md | ||
| test_hf_qlora_train_and_merge.py | ||
| test_unsloth_qlora_train_and_merge.py | ||
QLoRA Train and Merge Tests
Overview
Tests that performing QLoRA training and merging weights to 16-bits post-training maintains same behavior as trained model.
test_unsloth_qlora_train_and_merge.py: Test Unsloth QLoRA train and merge usingFastLanguageModel.from_pretrained,FastLanguageModel.get_peft_model, andFastLanguageModel.save_pretrained_mergedapistest_hf_qlora_train_and_merge.py: Test Hugging Face QLoRA train and merge usingfrom_pretrained,get_peft_model, andmerge_and_unloadapis.- Demonstrates that
peft'smerge_and_unloadresults in loss of accuracy as it requantizes the base layer after merging adapter weights so that the model still containsLinear4Bitlayers post merging. - I (@jeromeku) implemented a custom merge function that replaces all
LoraLayerswithLinearlayers whose weights are the dequantized base layer weights with adapter weights merged (compute done in fp32, cast to original dtype after merging), roughly equivalent toFastLanguageModel.save_pretrained_merged.
- Demonstrates that
Usage
Run unsloth test:
python tests/qlora/test_unsloth_qlora_train_and_merge.py
Run huggingface test:
python tests/qlora/test_hf_qlora_train_and_merge.py
Details
The tests train a QLoRA model on a single prompt dataset
QUESTION = "What day was I born?"
ANSWER = "January 1, 2058"
USER_MESSAGE = {"role": "user", "content": QUESTION}
ASSISTANT_MESSAGE = {"role": "assistant", "content": ANSWER}
Given that the answer is impossible to answer accurately without finetuning, we can only expect the model to answer the question correctly if the model has been trained on the question.
To check this behavior, we check the model's response to the question before and after training and after merging, checking that the model's response contains the answer after training and merging but not before training.
Results
For the unsloth test, the model's behavior is as expected:
- before training, the model's response does not contain the answer
- after training, the model's response contains the answer
- after merging, the model's response contains the answer
For the huggingface test, the model's behavior is as expected:
- before training, the model's response does not contain the answer
- after training, the model's response contains the answer
- after using peft's
merge_and_unload, the model's response does not contain the answer - after using my custom merge function, the model's response contains the answer
The scripts should output training params, training logs, as well as model responses before and after training and after merging (only prints model responses if answer is not contained in response).