From 5a2d8a38acc4c8eb95129e841d4b70c36e0f1452 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E6=B8=B8=E9=9B=81?= Date: Tue, 3 Sep 2024 11:26:25 +0800 Subject: [PATCH] lora --- funasr/models/llm_asr/model.py | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) diff --git a/funasr/models/llm_asr/model.py b/funasr/models/llm_asr/model.py index 2432070ec..2e31e9f7b 100644 --- a/funasr/models/llm_asr/model.py +++ b/funasr/models/llm_asr/model.py @@ -974,8 +974,9 @@ class LLMASR4(nn.Module): if lora_init_param_path is not None: model = PeftModel.from_pretrained(model, lora_init_param_path) for name, param in model.named_parameters(): - if "lora_A" in name or "lora_B" in name: - param.requires_grad = True + if not lora_conf.get("freeze_lora", False): + if "lora_A" in name or "lora_B" in name: + param.requires_grad = True else: peft_config = LoraConfig(**lora_conf) model = get_peft_model(model, peft_config)