unsloth/tests/qlora
2025-12-01 07:24:58 -08:00
..
README.md Fix typos (#2540) 2025-05-15 04:23:27 -07:00
test_hf_qlora_train_and_merge.py Revert "[pre-commit.ci] auto fixes from pre-commit.com hooks" 2025-12-01 07:24:58 -08:00
test_unsloth_qlora_train_and_merge.py Revert "[pre-commit.ci] auto fixes from pre-commit.com hooks" 2025-12-01 07:24:58 -08:00

QLoRA Train and Merge Tests

Overview

Tests that performing QLoRA training and merging weights to 16-bits post-training maintains same behavior as trained model.

  • test_unsloth_qlora_train_and_merge.py: Test Unsloth QLoRA train and merge using FastLanguageModel.from_pretrained, FastLanguageModel.get_peft_model, and FastLanguageModel.save_pretrained_merged apis
  • test_hf_qlora_train_and_merge.py: Test Hugging Face QLoRA train and merge using from_pretrained, get_peft_model, and merge_and_unload apis.
    • Demonstrates that peft's merge_and_unload results in loss of accuracy as it requantizes the base layer after merging adapter weights so that the model still contains Linear4Bit layers post merging.
    • I (@jeromeku) implemented a custom merge function that replaces all LoraLayers with Linear layers whose weights are the dequantized base layer weights with adapter weights merged (compute done in fp32, cast to original dtype after merging), roughly equivalent to FastLanguageModel.save_pretrained_merged.

Usage

Run unsloth test:

python tests/qlora/test_unsloth_qlora_train_and_merge.py

Run huggingface test:

python tests/qlora/test_hf_qlora_train_and_merge.py

Details

The tests train a QLoRA model on a single prompt dataset

QUESTION = "What day was I born?"
ANSWER = "January 1, 2058"
USER_MESSAGE = {"role": "user", "content": QUESTION}
ASSISTANT_MESSAGE = {"role": "assistant", "content": ANSWER}

Given that the answer is impossible to answer accurately without finetuning, we can only expect the model to answer the question correctly if the model has been trained on the question.

To check this behavior, we check the model's response to the question before and after training and after merging, checking that the model's response contains the answer after training and merging but not before training.

Results

For the unsloth test, the model's behavior is as expected:

  • before training, the model's response does not contain the answer
  • after training, the model's response contains the answer
  • after merging, the model's response contains the answer

For the huggingface test, the model's behavior is as expected:

  • before training, the model's response does not contain the answer
  • after training, the model's response contains the answer
  • after using peft's merge_and_unload, the model's response does not contain the answer
  • after using my custom merge function, the model's response contains the answer

The scripts should output training params, training logs, as well as model responses before and after training and after merging (only prints model responses if answer is not contained in response).