Michael Han
|
e2fd946fe1
|
Add files via upload
|
2026-04-02 03:00:10 -07:00 |
|
Michael Han
|
31d6aeb197
|
Unsloth new logo
|
2026-04-02 02:58:21 -07:00 |
|
Michael Han
|
f3b6e0e486
|
Add files via upload
|
2026-03-17 06:42:25 -07:00 |
|
Michael Han
|
9e2ef7c50c
|
Uploading HQ Unsloth Sticker
|
2025-05-04 05:31:57 -07:00 |
|
Michael Han
|
c4d0fd42be
|
Updating HQ logos
|
2025-05-04 05:25:06 -07:00 |
|
Daniel Han
|
6d34ab821b
|
Vision (#1318)
* Add files via upload
* Add files via upload
* Add files via upload
* Add files via upload
* Update README.md
* Update README.md
* Update README.md
* Update README.md
---------
Co-authored-by: Michael <107991372+shimmyshimmer@users.noreply.github.com>
|
2024-11-21 11:24:12 -08:00 |
|
Michael
|
778359ee9e
|
Add files via upload
|
2024-11-20 01:47:23 -08:00 |
|
Michael
|
26a3095d76
|
Add files via upload
|
2024-11-20 01:44:15 -08:00 |
|
Daniel Han
|
8ca886825c
|
Create Run.png
|
2024-07-23 13:14:21 -07:00 |
|
Daniel Han
|
f176cbd36a
|
Many bug fixes (#754)
* Update gemma2.py
* Update llama.py
* Update llama.py
* Update gemma2.py
* init
* Update gemma2.py
* Update gemma2.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* All RoPE Scaling support
* cleanup
* Update llama.py
* Update llama.py
* Update _utils.py
* Update _utils.py
* exec
* exec
* Attention_Module
* attention_module
* imports
* exec
* Update llama.py
* Update llama.py
* boolean mask
* revert masking
* Update llama.py
* Update save.py
* Update llama.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update utils.py
* retry
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update _utils.py
* Update _utils.py
* Update gemma2.py
* Update chat_templates.py
* Gemma 2 Ollama support
* Update llama.py
* Update llama.py
* error handling
* Update _utils.py
* Update _utils.py
* Stats for debugging
* Update _utils.py
* Update _utils.py
* Debugging
* Update tokenizer_utils.py
* Update _utils.py
* Update cross_entropy_loss.py
* Update cross_entropy_loss.py
* Update cross_entropy_loss.py
* Update rms_layernorm.py
* Update rms_layernorm.py
* Update rms_layernorm.py
* Update rms_layernorm.py
* Update rms_layernorm.py
* Update rms_layernorm.py
* Update rms_layernorm.py
* Check exec, eval
* Update _utils.py
* Update _utils.py
* Images
* Bug fixes
* Update pyproject.toml
* Bug fixes
* Update _utils.py
* Update _utils.py
|
2024-07-10 01:59:06 -07:00 |
|
Daniel Han
|
cc4c5d7785
|
Gemma2 (#709)
* Update mapper.py
* Update loader.py
* Update llama.py
* Update tokenizer_utils.py
* info
* edits
* Create chat template
* Fix tokenizer
* Update tokenizer_utils.py
* fix case where gguf saving fails due to first_conversion dtype (#630)
* Support revision parameter in FastLanguageModel.from_pretrained (#629)
* support `revision` parameter
* match unsloth formatting of named parameters
* clears any selected_adapters before calling internal_model.save_pretrained (#609)
* Update __init__.py (#602)
Check for incompatible modules before importing unsloth
* Fixed unsloth/tokenizer_utils.py for chat training (#604)
* Add GGML saving option to Unsloth for easier Ollama model creation and testing. (#345)
* Add save to llama.cpp GGML to save.py.
* Fix conversion command and path of convert to GGML function.
* Add autosaving lora to the GGML function
* Create lora save function for conversion to GGML
* Test fix #2 for saving lora
* Test fix #3 to save the lora adapters to convert to GGML
* Remove unwated tokenizer saving for conversion to ggml and added a few print statements.
* Needed tokenizer for saving, added it back, also made it more unslothy style by having positional arguments, and added a few messages.
* Positional arguments didn't work out, so reverted to older version of the code, and added a few comments.
* Test fix 1 for arch
* Test fix 2 new Mistral error.
* Test fix 3
* Revert to old version for testing.
* Upload issue test fix 1
* Fix 2 uploading ggml
* Positional ags added.
* Temporray remove positional args
* Fix upload again!!!
* Add print statements and fix link
* Make the calling name better
* Create local saving for GGML
* Add choosing directory to save local GGML.
* Fix lil variable error in the save_to_custom_dir func
* docs: Add LoraConfig parameters documentation (#619)
* llama.cpp failing (#371)
llama.cpp is failing to generate quantize versions for the trained models.
Error:
```bash
You might have to compile llama.cpp yourself, then run this again.
You do not need to close this Python program. Run the following commands in a new terminal:
You must run this in the same folder as you're saving your model.
git clone https://github.com/ggerganov/llama.cpp
cd llama.cpp && make clean && LLAMA_CUDA=1 make all -j
Once that's done, redo the quantization.
```
But when i do clone this with recursive it works.
Co-authored-by: Daniel Han <danielhanchen@gmail.com>
* fix libcuda_dirs import for triton 3.0 (#227)
* fix libcuda_dirs import for triton 3.0
* Update __init__.py
* Update __init__.py
---------
Co-authored-by: Daniel Han <danielhanchen@gmail.com>
* Update save.py
* Update __init__.py
* Update fast_lora.py
* Update save.py
* Update save.py
* Update save.py
* Update loader.py
* Update save.py
* Update save.py
* quantize now llama-quantize
* Update chat_templates.py
* Update loader.py
* Update mapper.py
* Update __init__.py
* embedding size
* Update qwen2.py
* docs
* Update README.md
* Update qwen2.py
* README: Fix minor typo. (#559)
* README: Fix minor typo.
One-character typo fix while reading.
* Update README.md
---------
Co-authored-by: Daniel Han <danielhanchen@gmail.com>
* Update mistral.py
* Update qwen2.py
* Update qwen2.py
* Update qwen2.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update README.md
* FastMistralModel
* Update mistral.py
* Update mistral.py
* Update mistral.py
* Update mistral.py
* Update mistral.py
* Auto check rope scaling
* Update llama.py
* Update llama.py
* Update llama.py
* GPU support
* Typo
* Update gemma.py
* gpu
* Multiple GGUF saving
* Update save.py
* Update save.py
* check PEFT and base
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update chat_templates.py
* Fix breaking bug in save.py with interpreting quantization_method as a string when saving to gguf (#651)
* Nightly (#649)
* Update llama.py
* offload
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* continued pretraining trainer
* Update trainer.py
* Update trainer.py
* Update trainer.py
* Update trainer.py
* is_bfloat16_supported
* Update __init__.py
* Update README.md
* Update llama.py
* is_bfloat16_supported
* Update __init__.py
* Mistral v3
* Phi 3 medium
* Update chat_templates.py
* Update chat_templates.py
* Phi-3
* Update save.py
* Update README.md
Mistral v3 to Mistral v0.3
* Untrained tokens
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* Update llama.py
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* Update save.py
* Update save.py
* Update save.py
* checkpoint
* Update _utils.py
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* Update llama.py
* accelerate
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update tokenizer_utils.py
* train_dataloader
* Update llama.py
* Update llama.py
* Update llama.py
* use_fast_convert
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* remove_special_tokens
* Ollama
* Update chat_templates.py
* Update chat_templates.py
* Update chat_templates.py
* Update llama.py
* Update chat_templates.py
* Support bfloat16 GGUF
* Update save.py
* Update llama.py
* fast_forward_inference
* Update mapper.py
* Update loader.py
* Update llama.py
* Update tokenizer_utils.py
* info
* edits
* Create chat template
* Fix tokenizer
* Update tokenizer_utils.py
* fix case where gguf saving fails due to first_conversion dtype (#630)
* Support revision parameter in FastLanguageModel.from_pretrained (#629)
* support `revision` parameter
* match unsloth formatting of named parameters
* clears any selected_adapters before calling internal_model.save_pretrained (#609)
* Update __init__.py (#602)
Check for incompatible modules before importing unsloth
* Fixed unsloth/tokenizer_utils.py for chat training (#604)
* Add GGML saving option to Unsloth for easier Ollama model creation and testing. (#345)
* Add save to llama.cpp GGML to save.py.
* Fix conversion command and path of convert to GGML function.
* Add autosaving lora to the GGML function
* Create lora save function for conversion to GGML
* Test fix #2 for saving lora
* Test fix #3 to save the lora adapters to convert to GGML
* Remove unwated tokenizer saving for conversion to ggml and added a few print statements.
* Needed tokenizer for saving, added it back, also made it more unslothy style by having positional arguments, and added a few messages.
* Positional arguments didn't work out, so reverted to older version of the code, and added a few comments.
* Test fix 1 for arch
* Test fix 2 new Mistral error.
* Test fix 3
* Revert to old version for testing.
* Upload issue test fix 1
* Fix 2 uploading ggml
* Positional ags added.
* Temporray remove positional args
* Fix upload again!!!
* Add print statements and fix link
* Make the calling name better
* Create local saving for GGML
* Add choosing directory to save local GGML.
* Fix lil variable error in the save_to_custom_dir func
* docs: Add LoraConfig parameters documentation (#619)
* llama.cpp failing (#371)
llama.cpp is failing to generate quantize versions for the trained models.
Error:
```bash
You might have to compile llama.cpp yourself, then run this again.
You do not need to close this Python program. Run the following commands in a new terminal:
You must run this in the same folder as you're saving your model.
git clone https://github.com/ggerganov/llama.cpp
cd llama.cpp && make clean && LLAMA_CUDA=1 make all -j
Once that's done, redo the quantization.
```
But when i do clone this with recursive it works.
Co-authored-by: Daniel Han <danielhanchen@gmail.com>
* fix libcuda_dirs import for triton 3.0 (#227)
* fix libcuda_dirs import for triton 3.0
* Update __init__.py
* Update __init__.py
---------
Co-authored-by: Daniel Han <danielhanchen@gmail.com>
* Update save.py
* Update __init__.py
* Update fast_lora.py
* Update save.py
* Update save.py
* Update save.py
* Update loader.py
* Update save.py
* Update save.py
* quantize now llama-quantize
* Update chat_templates.py
* Update loader.py
* Update mapper.py
* Update __init__.py
* embedding size
* Update qwen2.py
* docs
* Update README.md
* Update qwen2.py
* README: Fix minor typo. (#559)
* README: Fix minor typo.
One-character typo fix while reading.
* Update README.md
---------
Co-authored-by: Daniel Han <danielhanchen@gmail.com>
* Update mistral.py
* Update qwen2.py
* Update qwen2.py
* Update qwen2.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update README.md
* FastMistralModel
* Update mistral.py
* Update mistral.py
* Update mistral.py
* Update mistral.py
* Update mistral.py
* Auto check rope scaling
* Update llama.py
* Update llama.py
* Update llama.py
* GPU support
* Typo
* Update gemma.py
* gpu
* Multiple GGUF saving
* Update save.py
* Update save.py
* check PEFT and base
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update chat_templates.py
---------
Co-authored-by: Michael Han <107991372+shimmyshimmer@users.noreply.github.com>
Co-authored-by: Eliot Hall <60240707+chrehall68@users.noreply.github.com>
Co-authored-by: Rickard Edén <rickardeden@gmail.com>
Co-authored-by: XiaoYang <xyangk@gmail.com>
Co-authored-by: Oseltamivir <58582368+Oseltamivir@users.noreply.github.com>
Co-authored-by: mahiatlinux <110882203+mahiatlinux@users.noreply.github.com>
Co-authored-by: Sébastien De Greef <sebdg@binarycompany.com>
Co-authored-by: Alberto Ferrer <albertof@barrahome.org>
Co-authored-by: Thomas Viehmann <tv.github-private@beamnet.de>
Co-authored-by: Walter Korman <lemurware@gmail.com>
* Fix bug in save.py with interpreting quantization_method as a string that prevents GGUF from saving
* Implemented better list management and then forgot to actually call the new list variable, fixed
* Check type of given quantization method and return type error if not list or string
* Update save.py
---------
Co-authored-by: Daniel Han <danielhanchen@gmail.com>
Co-authored-by: Michael Han <107991372+shimmyshimmer@users.noreply.github.com>
Co-authored-by: Eliot Hall <60240707+chrehall68@users.noreply.github.com>
Co-authored-by: Rickard Edén <rickardeden@gmail.com>
Co-authored-by: XiaoYang <xyangk@gmail.com>
Co-authored-by: Oseltamivir <58582368+Oseltamivir@users.noreply.github.com>
Co-authored-by: mahiatlinux <110882203+mahiatlinux@users.noreply.github.com>
Co-authored-by: Sébastien De Greef <sebdg@binarycompany.com>
Co-authored-by: Alberto Ferrer <albertof@barrahome.org>
Co-authored-by: Thomas Viehmann <tv.github-private@beamnet.de>
Co-authored-by: Walter Korman <lemurware@gmail.com>
* Revert "Fix breaking bug in save.py with interpreting quantization_method as …" (#652)
This reverts commit 506cb68867296237e95bc53c32f1bfc9b1757960.
* Revert "Revert "Fix breaking bug in save.py with interpreting quantization_me…" (#653)
This reverts commit 2f48cc9af385579876fd45bd833169d1f1a2ea58.
* Update llama.py
* peft
* patch
* Update loader.py
* retrain
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* offload
* Update llama.py
* Create a starter script for command-line training to integrate in ML ops pipelines. (#623)
* Update chat_templates.py
* Ollama
* Update chat_templates.py
* Update chat_templates.py
* Update chat_templates.py
* Update chat_templates.py
* Update chat_templates.py
* Update chat_templates.py
* Update chat_templates.py
* Update chat_templates.py
* Update chat_templates.py
* Update chat_templates.py
* Ollama
* Update chat_templates.py
* ollama
* Update mapper.py
* Update chat_templates.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update chat_templates.py
* Update chat_templates.py
* Update chat_templates.py
* Update chat_templates.py
* Update llama.py
* Fixes
* clearer messages
* Update tokenizer_utils.py
* Update tokenizer_utils.py
* Update llama.py
* Update llama.py
* Update llama.py
* log
* Update __init__.py
* Update llama.py
* Update __init__.py
* Create Merge.png
* Create ollama.png
* Gemma2
* Update llama.py
* Update loader.py
* Update pyproject.toml
* Update pyproject.toml
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update _utils.py
* Revert Gemma2
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update rms_layernorm.py
* Update gemma2.py
* logit softcapping
* Update cross_entropy_loss.py
* Update llama.py
* Update llama.py
* Update gemma2.py
* Update gemma2.py
* Update cross_entropy_loss.py
* Update llama.py
* Update llama.py
* Update cross_entropy_loss.py
* Update cross_entropy_loss.py
* Update llama.py
* Update cross_entropy_loss.py
* Update cross_entropy_loss.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update llama.py
* Update gemma2.py
* Update llama.py
* Update llama.py
* Update gemma2.py
* Update gemma2.py
* Update llama.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update gemma2.py
* Update _utils.py
* Update _utils.py
* Update gemma2.py
* compile flags
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Update gemma2.py
* Update gemma2.py
* fixes
* Update _utils.py
* Fix generation
* Update llama.py
* Update llama.py
* Update _utils.py
* Update _utils.py
* Update _utils.py
* pad token
* Update gemma2.py
* pad token
* Update _utils.py
* Update llama.py
* Update gemma2.py
* edit warning
* Update tokenizer_utils.py
---------
Co-authored-by: Eliot Hall <60240707+chrehall68@users.noreply.github.com>
Co-authored-by: Rickard Edén <rickardeden@gmail.com>
Co-authored-by: XiaoYang <xyangk@gmail.com>
Co-authored-by: Oseltamivir <58582368+Oseltamivir@users.noreply.github.com>
Co-authored-by: mahiatlinux <110882203+mahiatlinux@users.noreply.github.com>
Co-authored-by: Sébastien De Greef <sebdg@binarycompany.com>
Co-authored-by: Alberto Ferrer <albertof@barrahome.org>
Co-authored-by: Thomas Viehmann <tv.github-private@beamnet.de>
Co-authored-by: Walter Korman <lemurware@gmail.com>
Co-authored-by: ArcadaLabs-Jason <52756218+ArcadaLabs-Jason@users.noreply.github.com>
Co-authored-by: Michael Han <107991372+shimmyshimmer@users.noreply.github.com>
|
2024-07-02 22:51:01 -07:00 |
|
Michael Han
|
bb66faaa33
|
ReadMe Revamp (#156)
* HF Perf Button
* Update README.md
Adding new buttons cleanup
* Update README.md
* Delete images/Discord.png
* Delete images/try live demo green.png
* new transparent logos
* Revamping page
* Revamp mainpage
* Update README.md
* Update README.md
* Update README.md
* Update README.md
* Update README.md
* Update README.md
* Update README.md
* finetune button
* Delete start free finetune button.png
* free finetune button
* Add files via upload
* Update README.md
* Update README.md
* Add files via upload
* Add files via upload
* Update README.md
* Add files via upload
* Update README.md
* Update README.md
* Update README.md
* Update README.md
* Update README.md
* Update README.md
* Update README.md
* Update README.md
* Update README.md
* Update README.md
* Squashed commit of the following:
commit efa0d2332e
Author: Daniel Han <danielhanchen@gmail.com>
Date: Sun Feb 4 17:35:56 2024 +1100
2x faster inference (#151)
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update llama.py
* Update fast_lora.py
* Update llama.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update swiglu.py
* Update fast_lora.py
* Update swiglu.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update save.py
* Update fast_lora.py
* Update utils.py
* Update llama.py
* Update fast_lora.py
* Update swiglu.py
* Update save.py
* Update save.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Revert "Update llama.py"
This reverts commit a208ec46e0.
* Update llama.py
* Works?
* Update pyproject.toml
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Swiglu
* Update swiglu.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update swiglu.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* attention_mask
* Update llama.py
* Update llama.py
* labels
* Update mistral.py
* Update llama.py
* attention mask
* Update save.py
* Update save.py
* Update mistral.py
* attention mask
* Update llama.py
* Update llama.py
* Update mistral.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update dpo.py
* Patch saving
* Update save.py
* Update save.py
* patch_saving_functions
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* print
* Mistral patch
* Update mistral.py
* Update save.py
* saving
* Update llama.py
* Update llama.py
* Fast inference repatch
* Update llama.py
* Update utils.py
* Update utils.py
* Update utils.py
* Update mistral.py
* Update __init__.py
* Fix inference
* Update mistral.py
* fast lm_head
* Remove fast path
* Update rope_embedding.py
* Update loader.py
* LlamaAttention_fast_forward_inference
* if past_key_value is not None and q_len == 1:
* revert inference
* Update loader.py
* past_key_value
* Update llama.py
* Update llama.py
* Fix SDPA
* Update llama.py
* padding
* Inference
* Update llama.py
* Revert
* Update mistral.py
* faster inference
* inference
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* inference
* Update llama.py
* Update utils.py
* faster inference
* Update llama.py
* revert
* lm_head
* Update llama.py
* inference
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update mistral.py
* Update llama.py
* faster inference
* Update llama.py
* fast inference
* Update llama.py
* Update llama.py
* Update mistral.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* torch compile
* past_key_values
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update utils.py
* Update utils.py
* Update utils.py
* Update utils.py
* Update llama.py
* fast inference + saving config.json
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update mistral.py
* fast inference again
* more temp matrices
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* fast inference
* Update mistral.py
* Update llama.py
* SDPA
* attention_mask
* New version
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update utils.py
* Update utils.py
commit 2f55935f94
Author: Daniel Han <danielhanchen@gmail.com>
Date: Wed Jan 31 04:03:37 2024 +1100
Hotfix - fix inference (#146)
* faster saving & inference
* Update llama.py
* Update save.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update mistral.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* fast inference
* Update llama.py
* Update save.py
* Update llama.py
* Mistral correct RoPE scaling
* Max sequence lengths
* Apache 2
* fast_linear_forward
* Update utils.py
* Update utils.py
* No print
* Update utils.py
* Update utils.py
* inference
* Update llama.py
* Fast inference RoPE
* Update llama.py
* Update llama.py
* RoPE
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* LoRA
* Fast LoRA saving
* Update llama.py
* hidden_states
* q_len == 1
* q_len issue
* Update mistral.py
* Update mistral.py
* incorrect inference
* Update to transformers 4.37
* Graceful FA2 error + torch 2.1.1
* Update mapper.py
* Update pyproject.toml
* Fix saving and bnb-4bit
* Update fast_lora.py
* Update fast_lora.py
* remove patching
* Update llama.py
* Update llama.py
* Update swiglu.py
* Repatch
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update llama.py
* Update fast_lora.py
* Update llama.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update swiglu.py
* Update fast_lora.py
* Update swiglu.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update save.py
* Update fast_lora.py
* Update utils.py
* Update llama.py
* Update fast_lora.py
* Update swiglu.py
* Update save.py
* Update save.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Revert "Update llama.py"
This reverts commit a208ec46e0.
* Update llama.py
* Works?
* Update pyproject.toml
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Swiglu
* Update swiglu.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update swiglu.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* attention_mask
* Update llama.py
* Update llama.py
* labels
* Update mistral.py
* Update llama.py
* attention mask
* Update save.py
* Update save.py
* Update mistral.py
* attention mask
* Update llama.py
* Update llama.py
* Update mistral.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update dpo.py
* Patch saving
* Update save.py
* Update save.py
* patch_saving_functions
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* print
* Mistral patch
* Update mistral.py
* Update save.py
* saving
* Update llama.py
* Update llama.py
* Fast inference repatch
* Update llama.py
* Update utils.py
* Update utils.py
* Update utils.py
* Update mistral.py
* Update __init__.py
* Fix inference
* Update mistral.py
* fast lm_head
* Remove fast path
* Update rope_embedding.py
* Update loader.py
* LlamaAttention_fast_forward_inference
* if past_key_value is not None and q_len == 1:
* revert inference
* Update loader.py
* past_key_value
commit a3a2ad9382
Author: Daniel Han <danielhanchen@gmail.com>
Date: Mon Jan 29 17:49:54 2024 +1100
Fix inference attention mask (#142)
* faster saving & inference
* Update llama.py
* Update save.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update mistral.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* fast inference
* Update llama.py
* Update save.py
* Update llama.py
* Mistral correct RoPE scaling
* Max sequence lengths
* Apache 2
* fast_linear_forward
* Update utils.py
* Update utils.py
* No print
* Update utils.py
* Update utils.py
* inference
* Update llama.py
* Fast inference RoPE
* Update llama.py
* Update llama.py
* RoPE
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* LoRA
* Fast LoRA saving
* Update llama.py
* hidden_states
* q_len == 1
* q_len issue
* Update mistral.py
* Update mistral.py
* incorrect inference
* Update to transformers 4.37
* Graceful FA2 error + torch 2.1.1
* Update mapper.py
* Update pyproject.toml
* Fix saving and bnb-4bit
* Update fast_lora.py
* Update fast_lora.py
* remove patching
* Update llama.py
* Update llama.py
* Update swiglu.py
* Repatch
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update llama.py
* Update fast_lora.py
* Update llama.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update swiglu.py
* Update fast_lora.py
* Update swiglu.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update save.py
* Update fast_lora.py
* Update utils.py
* Update llama.py
* Update fast_lora.py
* Update swiglu.py
* Update save.py
* Update save.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Revert "Update llama.py"
This reverts commit a208ec46e0.
* Update llama.py
* Works?
* Update pyproject.toml
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Swiglu
* Update swiglu.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update swiglu.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* attention_mask
* Update llama.py
* Update llama.py
* labels
* Update mistral.py
* Update llama.py
* attention mask
* Update save.py
* Update save.py
* Update mistral.py
* attention mask
* Update llama.py
* Update llama.py
* Update mistral.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update dpo.py
* Patch saving
* Update save.py
* Update save.py
* patch_saving_functions
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* print
* Mistral patch
* Update mistral.py
* Update save.py
* saving
* Update llama.py
* Update llama.py
commit 90309ca8dc
Author: Daniel Han <danielhanchen@gmail.com>
Date: Mon Jan 29 03:45:07 2024 +1100
Nightly (#140)
* faster saving & inference
* Update llama.py
* Update save.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update mistral.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* fast inference
* Update llama.py
* Update save.py
* Update llama.py
* Mistral correct RoPE scaling
* Max sequence lengths
* Apache 2
* fast_linear_forward
* Update utils.py
* Update utils.py
* No print
* Update utils.py
* Update utils.py
* inference
* Update llama.py
* Fast inference RoPE
* Update llama.py
* Update llama.py
* RoPE
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* LoRA
* Fast LoRA saving
* Update llama.py
* hidden_states
* q_len == 1
* q_len issue
* Update mistral.py
* Update mistral.py
* incorrect inference
* Update to transformers 4.37
* Graceful FA2 error + torch 2.1.1
* Update mapper.py
* Update pyproject.toml
* Fix saving and bnb-4bit
* Update fast_lora.py
* Update fast_lora.py
* remove patching
* Update llama.py
* Update llama.py
* Update swiglu.py
* Repatch
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update llama.py
* Update fast_lora.py
* Update llama.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update swiglu.py
* Update fast_lora.py
* Update swiglu.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update save.py
* Update fast_lora.py
* Update utils.py
* Update llama.py
* Update fast_lora.py
* Update swiglu.py
* Update save.py
* Update save.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Revert "Update llama.py"
This reverts commit a208ec46e0.
* Update llama.py
* Works?
* Update pyproject.toml
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Swiglu
* Update swiglu.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update swiglu.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* attention_mask
* Update llama.py
* Update llama.py
* labels
* Update mistral.py
* Update llama.py
* attention mask
* Update save.py
* Update save.py
* Update mistral.py
* attention mask
* Update llama.py
* Update llama.py
* Update mistral.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update dpo.py
* Patch saving
* Update save.py
* Update save.py
* patch_saving_functions
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* print
* Mistral patch
* Update mistral.py
* Update save.py
* saving
commit a16bc73e80
Author: Daniel Han <danielhanchen@gmail.com>
Date: Mon Jan 29 02:52:39 2024 +1100
Fix saving issues (#139)
* faster saving & inference
* Update llama.py
* Update save.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update mistral.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* fast inference
* Update llama.py
* Update save.py
* Update llama.py
* Mistral correct RoPE scaling
* Max sequence lengths
* Apache 2
* fast_linear_forward
* Update utils.py
* Update utils.py
* No print
* Update utils.py
* Update utils.py
* inference
* Update llama.py
* Fast inference RoPE
* Update llama.py
* Update llama.py
* RoPE
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* LoRA
* Fast LoRA saving
* Update llama.py
* hidden_states
* q_len == 1
* q_len issue
* Update mistral.py
* Update mistral.py
* incorrect inference
* Update to transformers 4.37
* Graceful FA2 error + torch 2.1.1
* Update mapper.py
* Update pyproject.toml
* Fix saving and bnb-4bit
* Update fast_lora.py
* Update fast_lora.py
* remove patching
* Update llama.py
* Update llama.py
* Update swiglu.py
* Repatch
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update llama.py
* Update fast_lora.py
* Update llama.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update swiglu.py
* Update fast_lora.py
* Update swiglu.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update save.py
* Update fast_lora.py
* Update utils.py
* Update llama.py
* Update fast_lora.py
* Update swiglu.py
* Update save.py
* Update save.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Revert "Update llama.py"
This reverts commit a208ec46e0.
* Update llama.py
* Works?
* Update pyproject.toml
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Swiglu
* Update swiglu.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update swiglu.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* attention_mask
* Update llama.py
* Update llama.py
* labels
* Update mistral.py
* Update llama.py
* attention mask
* Update save.py
* Update save.py
* Update mistral.py
* attention mask
* Update llama.py
* Update llama.py
* Update mistral.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update dpo.py
* Patch saving
* Update save.py
* Update save.py
* patch_saving_functions
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* print
commit af33224554
Author: Daniel Han <danielhanchen@gmail.com>
Date: Sun Jan 28 04:30:29 2024 +1100
1 more bug (#138)
* faster saving & inference
* Update llama.py
* Update save.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update mistral.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* fast inference
* Update llama.py
* Update save.py
* Update llama.py
* Mistral correct RoPE scaling
* Max sequence lengths
* Apache 2
* fast_linear_forward
* Update utils.py
* Update utils.py
* No print
* Update utils.py
* Update utils.py
* inference
* Update llama.py
* Fast inference RoPE
* Update llama.py
* Update llama.py
* RoPE
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* LoRA
* Fast LoRA saving
* Update llama.py
* hidden_states
* q_len == 1
* q_len issue
* Update mistral.py
* Update mistral.py
* incorrect inference
* Update to transformers 4.37
* Graceful FA2 error + torch 2.1.1
* Update mapper.py
* Update pyproject.toml
* Fix saving and bnb-4bit
* Update fast_lora.py
* Update fast_lora.py
* remove patching
* Update llama.py
* Update llama.py
* Update swiglu.py
* Repatch
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update llama.py
* Update fast_lora.py
* Update llama.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update swiglu.py
* Update fast_lora.py
* Update swiglu.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update save.py
* Update fast_lora.py
* Update utils.py
* Update llama.py
* Update fast_lora.py
* Update swiglu.py
* Update save.py
* Update save.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Revert "Update llama.py"
This reverts commit a208ec46e0.
* Update llama.py
* Works?
* Update pyproject.toml
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Swiglu
* Update swiglu.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update swiglu.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* attention_mask
* Update llama.py
* Update llama.py
* labels
* Update mistral.py
* Update llama.py
* attention mask
* Update save.py
* Update save.py
commit e2bbd3819e
Author: Daniel Han <danielhanchen@gmail.com>
Date: Sun Jan 28 04:20:06 2024 +1100
Fix bugs + more accurate Swiglu (#137)
* faster saving & inference
* Update llama.py
* Update save.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update mistral.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* fast inference
* Update llama.py
* Update save.py
* Update llama.py
* Mistral correct RoPE scaling
* Max sequence lengths
* Apache 2
* fast_linear_forward
* Update utils.py
* Update utils.py
* No print
* Update utils.py
* Update utils.py
* inference
* Update llama.py
* Fast inference RoPE
* Update llama.py
* Update llama.py
* RoPE
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* LoRA
* Fast LoRA saving
* Update llama.py
* hidden_states
* q_len == 1
* q_len issue
* Update mistral.py
* Update mistral.py
* incorrect inference
* Update to transformers 4.37
* Graceful FA2 error + torch 2.1.1
* Update mapper.py
* Update pyproject.toml
* Fix saving and bnb-4bit
* Update fast_lora.py
* Update fast_lora.py
* remove patching
* Update llama.py
* Update llama.py
* Update swiglu.py
* Repatch
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update llama.py
* Update fast_lora.py
* Update llama.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update swiglu.py
* Update fast_lora.py
* Update swiglu.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update save.py
* Update fast_lora.py
* Update utils.py
* Update llama.py
* Update fast_lora.py
* Update swiglu.py
* Update save.py
* Update save.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Revert "Update llama.py"
This reverts commit a208ec46e0.
* Update llama.py
* Works?
* Update pyproject.toml
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Swiglu
* Update swiglu.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update swiglu.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* attention_mask
* Update llama.py
* Update llama.py
* labels
* Update mistral.py
* Update llama.py
* attention mask
commit a81aff286f
Author: Daniel Han <danielhanchen@gmail.com>
Date: Sat Jan 27 04:50:22 2024 +1100
Inference bug fix (#134)
* faster saving & inference
* Update llama.py
* Update save.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update mistral.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* fast inference
* Update llama.py
* Update save.py
* Update llama.py
* Mistral correct RoPE scaling
* Max sequence lengths
* Apache 2
* fast_linear_forward
* Update utils.py
* Update utils.py
* No print
* Update utils.py
* Update utils.py
* inference
* Update llama.py
* Fast inference RoPE
* Update llama.py
* Update llama.py
* RoPE
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* LoRA
* Fast LoRA saving
* Update llama.py
* hidden_states
* q_len == 1
* q_len issue
* Update mistral.py
* Update mistral.py
* incorrect inference
* Update to transformers 4.37
* Graceful FA2 error + torch 2.1.1
* Update mapper.py
* Update pyproject.toml
* Fix saving and bnb-4bit
* Update fast_lora.py
* Update fast_lora.py
* remove patching
* Update llama.py
* Update llama.py
* Update swiglu.py
* Repatch
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update llama.py
* Update fast_lora.py
* Update llama.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update swiglu.py
* Update fast_lora.py
* Update swiglu.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update save.py
* Update fast_lora.py
* Update utils.py
* Update llama.py
* Update fast_lora.py
* Update swiglu.py
* Update save.py
* Update save.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Revert "Update llama.py"
This reverts commit a208ec46e0.
* Update llama.py
commit 7da0c50f75
Author: Daniel Han <danielhanchen@gmail.com>
Date: Sat Jan 27 04:47:54 2024 +1100
More bug fixes (#133)
* faster saving & inference
* Update llama.py
* Update save.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update mistral.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* fast inference
* Update llama.py
* Update save.py
* Update llama.py
* Mistral correct RoPE scaling
* Max sequence lengths
* Apache 2
* fast_linear_forward
* Update utils.py
* Update utils.py
* No print
* Update utils.py
* Update utils.py
* inference
* Update llama.py
* Fast inference RoPE
* Update llama.py
* Update llama.py
* RoPE
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* LoRA
* Fast LoRA saving
* Update llama.py
* hidden_states
* q_len == 1
* q_len issue
* Update mistral.py
* Update mistral.py
* incorrect inference
* Update to transformers 4.37
* Graceful FA2 error + torch 2.1.1
* Update mapper.py
* Update pyproject.toml
* Fix saving and bnb-4bit
* Update fast_lora.py
* Update fast_lora.py
* remove patching
* Update llama.py
* Update llama.py
* Update swiglu.py
* Repatch
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update llama.py
* Update fast_lora.py
* Update llama.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update swiglu.py
* Update fast_lora.py
* Update swiglu.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update fast_lora.py
* Update save.py
* Update fast_lora.py
* Update utils.py
* Update llama.py
* Update fast_lora.py
* Update swiglu.py
* Update save.py
* Update save.py
* Update llama.py
* Update llama.py
* Update llama.py
commit 62fae3aa74
Author: Daniel Han <danielhanchen@gmail.com>
Date: Fri Jan 26 04:19:17 2024 +1100
Fix bugs (#129)
* faster saving & inference
* Update llama.py
* Update save.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update mistral.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* fast inference
* Update llama.py
* Update save.py
* Update llama.py
* Mistral correct RoPE scaling
* Max sequence lengths
* Apache 2
* fast_linear_forward
* Update utils.py
* Update utils.py
* No print
* Update utils.py
* Update utils.py
* inference
* Update llama.py
* Fast inference RoPE
* Update llama.py
* Update llama.py
* RoPE
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* LoRA
* Fast LoRA saving
* Update llama.py
* hidden_states
* q_len == 1
* q_len issue
* Update mistral.py
* Update mistral.py
* incorrect inference
* Update to transformers 4.37
* Graceful FA2 error + torch 2.1.1
* Update mapper.py
* Update pyproject.toml
* Fix saving and bnb-4bit
* Update fast_lora.py
* Update fast_lora.py
* remove patching
* Update llama.py
* Update llama.py
* Update swiglu.py
* Repatch
* Update fast_lora.py
commit 04f8771821
Author: Daniel Han <danielhanchen@gmail.com>
Date: Tue Jan 23 03:55:24 2024 +1100
2-4x faster native HF inference (#119)
* faster saving & inference
* Update llama.py
* Update save.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update mistral.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* fast inference
* Update llama.py
* Update save.py
* Update llama.py
* Mistral correct RoPE scaling
* Max sequence lengths
* Apache 2
* fast_linear_forward
* Update utils.py
* Update utils.py
* No print
* Update utils.py
* Update utils.py
* inference
* Update llama.py
* Fast inference RoPE
* Update llama.py
* Update llama.py
* RoPE
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* LoRA
* Fast LoRA saving
commit 3a9b2dee98
Author: Daniel Han <danielhanchen@gmail.com>
Date: Sun Jan 21 22:20:22 2024 +1100
Hotfix (#118)
* faster saving & inference
* Update llama.py
* Update save.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update mistral.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
commit a6f4fb0075
Author: Daniel Han-Chen <danielhanchen@gmail.com>
Date: Sun Jan 21 05:00:37 2024 +1100
Update save.py
commit 705cac0357
Author: Daniel Han-Chen <danielhanchen@gmail.com>
Date: Sun Jan 21 04:21:54 2024 +1100
Update save.py
commit 16edcb3be2
Author: Daniel Han-Chen <danielhanchen@gmail.com>
Date: Sun Jan 21 04:13:03 2024 +1100
Update save.py
commit 3d05a74b12
Author: Daniel Han <danielhanchen@gmail.com>
Date: Sun Jan 21 03:43:49 2024 +1100
Fixed saving! (#113)
* Fix tokenizer, dropout, bias for LoRA
* Update loader.py
* Fix LoRA downcasting
* Update _utils.py
* Saving to GGUF
* fix
* colab_quantize_to_gguf
* move save modules
* save module
* Update __init__.py
* Update save.py
* Temp downgrade due to TRL issue
* Fix up bugs
* Faster saving + other changes
* Update llama.py
* Saving modules
* spelling
* Update llama.py
* Update save.py
* Update save.py
* Update loader.py
* Update llama.py
* patch saving
* Update save.py
* Update save.py
* Update save.py
* patch saving
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* original_model
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* saving to RAM leakage?
* Update save.py
* new_save_directory
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update pyproject.toml
* Update pyproject.toml
* Update pyproject.toml
* Quick fixes
* Update llama.py
* Update llama.py
* Update dpo.py
* Update dpo.py
* Update llama.py
* Update save.py
* getattr
* RSLoRA and LoftQ direct support
* Update llama.py
* Update llama.py
* Update llama.py
* Fix DPO + GGUF
* Fix quantization_method
* Fix quantization_config
* patch model
* Update llama.py
* Update llama.py
* Update llama.py
* Update save.py
* Update save.py
* tokenizer_save_settings
* Update save.py
* quantization and loftq
* Update save.py
* Update llama.py
* Update save.py
* upload_to_huggingface
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
commit bb05d6b6e2
Author: Daniel Han <danielhanchen@gmail.com>
Date: Sat Jan 20 23:23:00 2024 +1100
Hotfix for Jan 2024 Release (#110)
* Fix tokenizer, dropout, bias for LoRA
* Update loader.py
* Fix LoRA downcasting
* Update _utils.py
* Saving to GGUF
* fix
* colab_quantize_to_gguf
* move save modules
* save module
* Update __init__.py
* Update save.py
* Temp downgrade due to TRL issue
* Fix up bugs
* Faster saving + other changes
* Update llama.py
* Saving modules
* spelling
* Update llama.py
* Update save.py
* Update save.py
* Update loader.py
* Update llama.py
* patch saving
* Update save.py
* Update save.py
* Update save.py
* patch saving
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* original_model
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* saving to RAM leakage?
* Update save.py
* new_save_directory
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update pyproject.toml
* Update pyproject.toml
* Update pyproject.toml
* Quick fixes
* Update llama.py
* Update llama.py
* Update dpo.py
* Update dpo.py
* Update llama.py
* Update save.py
* getattr
* RSLoRA and LoftQ direct support
* Update llama.py
* Update llama.py
* Update llama.py
* Fix DPO + GGUF
* Fix quantization_method
* Fix quantization_config
* patch model
* Update llama.py
* Update llama.py
* Update llama.py
* Update save.py
* Update save.py
* tokenizer_save_settings
* Update save.py
* quantization and loftq
* Update save.py
* Update llama.py
* Update save.py
commit 12e75c93d0
Author: Daniel Han <danielhanchen@gmail.com>
Date: Sat Jan 20 04:25:06 2024 +1100
Quick fixes (#106)
* Fix tokenizer, dropout, bias for LoRA
* Update loader.py
* Fix LoRA downcasting
* Update _utils.py
* Saving to GGUF
* fix
* colab_quantize_to_gguf
* move save modules
* save module
* Update __init__.py
* Update save.py
* Temp downgrade due to TRL issue
* Fix up bugs
* Faster saving + other changes
* Update llama.py
* Saving modules
* spelling
* Update llama.py
* Update save.py
* Update save.py
* Update loader.py
* Update llama.py
* patch saving
* Update save.py
* Update save.py
* Update save.py
* patch saving
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* original_model
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* saving to RAM leakage?
* Update save.py
* new_save_directory
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update pyproject.toml
* Update pyproject.toml
* Update pyproject.toml
* Quick fixes
* Update llama.py
* Update llama.py
* Update dpo.py
* Update dpo.py
* Update llama.py
* Update save.py
* getattr
* RSLoRA and LoftQ direct support
* Update llama.py
* Update llama.py
* Update llama.py
* Fix DPO + GGUF
commit 52b5ef31e0
Author: Daniel Han-Chen <danielhanchen@gmail.com>
Date: Sat Jan 20 02:30:31 2024 +1100
Update _utils.py
commit 1a19c38675
Merge: 0a52390 0d6e52b
Author: Daniel Han-Chen <danielhanchen@gmail.com>
Date: Fri Jan 19 23:15:38 2024 +1100
Merge branch 'main' of https://github.com/unslothai/unsloth
commit 0a52390ac2
Author: Daniel Han-Chen <danielhanchen@gmail.com>
Date: Fri Jan 19 23:15:20 2024 +1100
Revert quantization methods
commit 0d6e52b5c7
Author: Daniel Han <danielhanchen@gmail.com>
Date: Fri Jan 19 22:57:22 2024 +1100
getattr issues (#103)
* Fix tokenizer, dropout, bias for LoRA
* Update loader.py
* Fix LoRA downcasting
* Update _utils.py
* Saving to GGUF
* fix
* colab_quantize_to_gguf
* move save modules
* save module
* Update __init__.py
* Update save.py
* Temp downgrade due to TRL issue
* Fix up bugs
* Faster saving + other changes
* Update llama.py
* Saving modules
* spelling
* Update llama.py
* Update save.py
* Update save.py
* Update loader.py
* Update llama.py
* patch saving
* Update save.py
* Update save.py
* Update save.py
* patch saving
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* original_model
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* saving to RAM leakage?
* Update save.py
* new_save_directory
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update pyproject.toml
* Update pyproject.toml
* Update pyproject.toml
* Quick fixes
* Update llama.py
* Update llama.py
* Update dpo.py
* Update dpo.py
* Update llama.py
* Update save.py
* getattr
commit b3fcea6421
Author: Daniel Han <danielhanchen@gmail.com>
Date: Fri Jan 19 22:52:30 2024 +1100
Quick fixes (#101)
* Fix tokenizer, dropout, bias for LoRA
* Update loader.py
* Fix LoRA downcasting
* Update _utils.py
* Saving to GGUF
* fix
* colab_quantize_to_gguf
* move save modules
* save module
* Update __init__.py
* Update save.py
* Temp downgrade due to TRL issue
* Fix up bugs
* Faster saving + other changes
* Update llama.py
* Saving modules
* spelling
* Update llama.py
* Update save.py
* Update save.py
* Update loader.py
* Update llama.py
* patch saving
* Update save.py
* Update save.py
* Update save.py
* patch saving
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* original_model
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* saving to RAM leakage?
* Update save.py
* new_save_directory
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update pyproject.toml
* Update pyproject.toml
* Update pyproject.toml
* Quick fixes
* Update llama.py
* Update llama.py
* Update dpo.py
* Update dpo.py
* Update llama.py
* Update save.py
commit d691516ab9
Author: Daniel Han <danielhanchen@gmail.com>
Date: Fri Jan 19 04:51:19 2024 +1100
2024 Release (#96)
* Fix tokenizer, dropout, bias for LoRA
* Update loader.py
* Fix LoRA downcasting
* Update _utils.py
* Saving to GGUF
* fix
* colab_quantize_to_gguf
* move save modules
* save module
* Update __init__.py
* Update save.py
* Temp downgrade due to TRL issue
* Fix up bugs
* Faster saving + other changes
* Update llama.py
* Saving modules
* spelling
* Update llama.py
* Update save.py
* Update save.py
* Update loader.py
* Update llama.py
* patch saving
* Update save.py
* Update save.py
* Update save.py
* patch saving
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* original_model
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* saving to RAM leakage?
* Update save.py
* new_save_directory
* Update save.py
* Update save.py
* Update save.py
* Update save.py
* Update pyproject.toml
* Update pyproject.toml
* Update pyproject.toml
commit 9e2dec16fb
Author: Daniel Han-Chen <danielhanchen@gmail.com>
Date: Fri Jan 19 03:41:00 2024 +1100
Update pyproject.toml
commit 396c7245dd
Author: Daniel Han-Chen <danielhanchen@gmail.com>
Date: Fri Jan 19 03:35:17 2024 +1100
Update pyproject.toml
commit 738e91591f
Author: Daniel Han <danielhanchen@gmail.com>
Date: Thu Jan 11 04:08:03 2024 +1100
Fix some bugs (#83)
* Fix tokenizer, dropout, bias for LoRA
* Update loader.py
* Fix LoRA downcasting
* Update _utils.py
* Saving to GGUF
* fix
* colab_quantize_to_gguf
* move save modules
* save module
* Update __init__.py
* Update save.py
* Temp downgrade due to TRL issue
* Fix up bugs
commit a1da50b5ce
Author: Daniel Han <danielhanchen@gmail.com>
Date: Wed Jan 10 23:10:48 2024 +1100
Update README.md (#81)
commit 606e8a9284
Author: shimmy <107991372+shimmyshimmer@users.noreply.github.com>
Date: Wed Jan 10 23:10:23 2024 +1100
Discord button redo (#80)
commit 0169294ffb
Author: shimmy <107991372+shimmyshimmer@users.noreply.github.com>
Date: Wed Jan 10 23:02:20 2024 +1100
Update logos (#79)
* HF Perf Button
* Update README.md
Adding new buttons cleanup
* Update README.md
* Delete images/Discord.png
* Delete images/try live demo green.png
* new transparent logos
* Revamping page
* Revamp mainpage
* Update README.md
* Update README.md
commit b2a8c33430
Author: Daniel Han <danielhanchen@gmail.com>
Date: Wed Jan 10 20:03:01 2024 +1100
Create FUNDING.yml (#78)
commit c9c1abf290
Author: Daniel Han-Chen <danielhanchen@gmail.com>
Date: Wed Jan 10 01:02:44 2024 +1100
fix_tokenizer
commit 6efffb46e4
Author: Daniel Han-Chen <danielhanchen@gmail.com>
Date: Tue Jan 9 23:40:43 2024 +1100
check_tokenizer
---------
Co-authored-by: Daniel Han <danielhanchen@gmail.com>
|
2024-02-07 02:00:12 +11:00 |
|
shimmy
|
606e8a9284
|
Discord button redo (#80)
|
2024-01-10 23:10:23 +11:00 |
|
shimmy
|
0169294ffb
|
Update logos (#79)
* HF Perf Button
* Update README.md
Adding new buttons cleanup
* Update README.md
* Delete images/Discord.png
* Delete images/try live demo green.png
* new transparent logos
* Revamping page
* Revamp mainpage
* Update README.md
* Update README.md
|
2024-01-10 23:02:20 +11:00 |
|
shimmy
|
8621977c8b
|
Replacing buttons (#75)
|
2024-01-09 20:17:18 +11:00 |
|
shimmy
|
38a524dfd1
|
Add files via upload (#74)
Updated buttons
|
2024-01-09 19:23:59 +11:00 |
|
Daniel Han
|
353432271d
|
Fix tokenizer + docs (#62)
* Patch tokenizer
* Update _utils.py
* Update _utils.py
* Update _utils.py
* Cleanup
* Add comments to functions
* Update rope_embedding.py
* Update rope_embedding.py
* Update llama.py
* New logos!
* Update README.md
|
2024-01-05 04:08:53 +11:00 |
|
Daniel Han
|
ef70177a24
|
Small fixes (#48)
* Fix generation for GQA
* Update _utils.py
* flash attn
* Update _utils.py
* Update llama.py
* Update mistral.py
* platform
* Update _utils.py
* Update llama.py
* Logo changed
* Update README.md
* Update README.md
|
2023-12-23 04:22:48 +11:00 |
|
Daniel Han-Chen
|
feec338c08
|
Torch version, docs, readme, general loader
|
2023-12-18 04:23:16 +11:00 |
|
Daniel Han
|
6416f235cf
|
Pre-release 2023 December version (Mistral, Prelim DPO, WSL, bug fixes) (#16)
* Immediate bug fixes
* Update README.md
* Update README.md
* Update llama.py
* Update llama.py
* Rope Scaling and max_seq_len will change
* Update llama.py
* new images
* Update README.md
* Images
* Update README.md
* Update pyproject.toml
* GQA
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
* Update llama.py
|
2023-12-06 02:59:22 +11:00 |
|
Daniel Han-Chen
|
d1c1748266
|
Sticker
|
2023-12-02 03:08:16 +11:00 |
|
Daniel Han-Chen
|
77c527b15d
|
Create Discord.png
|
2023-12-01 16:04:15 +11:00 |
|
Daniel Han-Chen
|
6f1da39738
|
Merge branch 'main' of https://github.com/unslothai/unsloth
|
2023-12-01 13:51:33 +11:00 |
|
Daniel Han-Chen
|
fc8c364e66
|
Create Slim Orca 2GPUs.png
|
2023-12-01 13:51:05 +11:00 |
|
Daniel Han
|
87b4e17d06
|
Add files via upload
|
2023-12-01 03:28:49 +11:00 |
|
Daniel Han
|
3276cf8c7a
|
Add files via upload
|
2023-12-01 01:34:43 +11:00 |
|
Daniel Han-Chen
|
30461b4057
|
images and exclusion
|
2023-12-01 01:22:09 +11:00 |
|
Daniel Han-Chen
|
71b59482cd
|
images
|
2023-12-01 01:20:24 +11:00 |
|
Daniel Han-Chen
|
13a8e1cd34
|
images
|
2023-12-01 00:56:06 +11:00 |
|