-
Notifications
You must be signed in to change notification settings - Fork 369
Added support for multiple LoRA adapters #232
Added support for multiple LoRA adapters #232
Conversation
This looks good, but I think
Apart from that, good to go from me. |
Alright i will move the loading logic back over and just iterate over the LoraAdapter's while loading the tensors. |
This should be ready, but i havent tested it yet. |
Old LoRA adapters fail with the following error:
Is this a side-effect of the new quantization formats? |
Can you provide more information about which models you tried it with? That sounds like something's misaligned because it comes from llm/crates/ggml/src/format/loader.rs Line 209 in 923a09b
|
Sure i used this LLama base model in q4_0 and these adapters: |
Alright, i passed the wrong paths to the LoRA adapter 🤦 . Its now working and tested it with the model and the two adapters linked above. |
Closes #227.
Moved the loading logic into the
LoraPatches
struct and added theLoraAdapter
struct which abstracts multiple LoRA patches.Changed "loar_adapter" to "lora_adapters" in the
ModelParameters
.