FunASR/funasr/train_utils/load_pretrained_model.py
2024-08-07 12:48:02 +08:00

130 lines
4.6 KiB
Python

from typing import Any
from typing import Dict
from typing import Union
from io import BytesIO
import os
import logging
import torch
import torch.nn
import torch.optim
import pdb
def load_pretrained_model(
path,
model: torch.nn.Module,
ignore_init_mismatch: bool = True,
map_location: str = "cpu",
oss_bucket=None,
scope_map=[],
excludes=None,
**kwargs,
):
"""Load a model state and set it to the model.
Args:
init_param: <file_path>:<src_key>:<dst_key>:<exclude_Keys>
Examples:
"""
obj = model
dst_state = obj.state_dict()
use_deepspeed = kwargs.get("use_deepspeed", False)
logging.info(f"ckpt: {path}, use_deepspeed: {use_deepspeed}")
if use_deepspeed and os.path.isdir(path):
ckpt_dir = os.path.dirname(path)
ckpt_name = os.path.basename(path)
if os.path.exists(f"{ckpt_dir}/zero_to_fp32.py"):
print("Detect zero_to_fp32, begin to convert fp32 model")
ckpt_fp32 = f"{ckpt_dir}/{ckpt_name[3:]}"
if os.path.exists(ckpt_fp32):
print(f"Detect zero_to_fp32 already exist! Loading it directly. {ckpt_fp32}")
src_state = torch.load(ckpt_fp32, map_location=map_location)
else:
with open(f"{ckpt_dir}/latest", "w") as latest:
latest.write(ckpt_name)
latest.flush()
from deepspeed.utils.zero_to_fp32 import (
get_fp32_state_dict_from_zero_checkpoint,
)
src_state = get_fp32_state_dict_from_zero_checkpoint(ckpt_dir) # already on cpu
if kwargs.get("save_deepspeed_zero_fp32", False):
print(
f'save_deepspeed_zero_fp32: {kwargs.get("save_deepspeed_zero_fp32", False)}, {ckpt_fp32}'
)
torch.save({"state_dict": src_state}, ckpt_fp32)
else:
print("Detect deepspeed without zero, load fp32 model directly")
for item in os.listdir(path):
if item.endswith(".pt"):
src_state = torch.load(f"{path}/{item}", map_location=map_location)
else:
src_state = torch.load(path, map_location=map_location)
src_state = src_state["state_dict"] if "state_dict" in src_state else src_state
src_state = src_state["model_state_dict"] if "model_state_dict" in src_state else src_state
src_state = src_state["model"] if "model" in src_state else src_state
if isinstance(scope_map, str):
scope_map = scope_map.split(",")
scope_map += ["module.", "None"]
logging.info(f"scope_map: {scope_map}")
if excludes is not None:
if isinstance(excludes, str):
excludes = excludes.split(",")
logging.info(f"excludes: {excludes}")
for k in dst_state.keys():
excludes_flag = False
if excludes is not None:
for k_ex in excludes:
if k.startswith(k_ex):
logging.info(f"key: {k} matching: {k_ex}, excluded")
excludes_flag = True
break
if excludes_flag:
continue
k_src = k
if scope_map is not None:
src_prefix = ""
dst_prefix = ""
for i in range(0, len(scope_map), 2):
src_prefix = scope_map[i] if scope_map[i].lower() != "none" else ""
dst_prefix = scope_map[i + 1] if scope_map[i + 1].lower() != "none" else ""
if dst_prefix == "" and (src_prefix + k) in src_state.keys():
k_src = src_prefix + k
if not k_src.startswith("module."):
logging.info(f"init param, map: {k} from {k_src} in ckpt")
elif (
k.startswith(dst_prefix)
and k.replace(dst_prefix, src_prefix, 1) in src_state.keys()
):
k_src = k.replace(dst_prefix, src_prefix, 1)
if not k_src.startswith("module."):
logging.info(f"init param, map: {k} from {k_src} in ckpt")
if k_src in src_state.keys():
if ignore_init_mismatch and dst_state[k].shape != src_state[k_src].shape:
logging.info(
f"ignore_init_mismatch:{ignore_init_mismatch}, dst: {k, dst_state[k].shape}, src: {k_src, src_state[k_src].shape}"
)
else:
dst_state[k] = src_state[k_src]
else:
print(f"Warning, miss key in ckpt: {k}, {path}")
flag = obj.load_state_dict(dst_state, strict=True)
logging.info(f"Loading ckpt: {path}, status: {flag}")