This commit is contained in:
游雁 2023-12-11 13:42:40 +08:00
parent 85cabd30a9
commit d77910eb6d
94 changed files with 3986 additions and 22 deletions

View File

@ -0,0 +1,113 @@
# network architecture
model: funasr.cli.models.paraformer:Paraformer
model_conf:
ctc_weight: 0.3
lsm_weight: 0.1
length_normalized_loss: false
predictor_weight: 1.0
sampling_ratio: 0.4
use_1st_decoder_loss: true
# encoder related
encoder: conformer
encoder_conf:
output_size: 256 # dimension of attention
attention_heads: 4
linear_units: 2048 # the number of units of position-wise feed forward
num_blocks: 12 # the number of encoder blocks
dropout_rate: 0.1
positional_dropout_rate: 0.1
attention_dropout_rate: 0.0
input_layer: conv2d # encoder architecture type
normalize_before: true
pos_enc_layer_type: rel_pos
selfattention_layer_type: rel_selfattn
activation_type: swish
macaron_style: true
use_cnn_module: true
cnn_module_kernel: 15
# decoder related
decoder: paraformer_decoder_san
decoder_conf:
attention_heads: 4
linear_units: 2048
num_blocks: 6
dropout_rate: 0.1
positional_dropout_rate: 0.1
self_attention_dropout_rate: 0.0
src_attention_dropout_rate: 0.0
# frontend related
frontend: wav_frontend
frontend_conf:
fs: 16000
window: hamming
n_mels: 80
frame_length: 25
frame_shift: 10
lfr_m: 1
lfr_n: 1
train_conf:
accum_grad: 1
grad_clip: 5
max_epoch: 150
val_scheduler_criterion:
- valid
- acc
best_model_criterion:
- - valid
- acc
- max
keep_nbest_models: 10
log_interval: 50
optim: adam
optim_conf:
lr: 0.0005
scheduler: warmuplr
scheduler_conf:
warmup_steps: 30000
specaug: specaug
specaug_conf:
apply_time_warp: true
time_warp_window: 5
time_warp_mode: bicubic
apply_freq_mask: true
freq_mask_width_range:
- 0
- 30
num_freq_mask: 2
apply_time_mask: true
time_mask_width_range:
- 0
- 40
num_time_mask: 2
predictor: cif_predictor
predictor_conf:
idim: 256
threshold: 1.0
l_order: 1
r_order: 1
tail_threshold: 0.45
dataset_conf:
data_names: speech,text
data_types: sound,text
shuffle: True
shuffle_conf:
shuffle_size: 2048
sort_size: 500
batch_conf:
batch_type: example
batch_size: 2
num_workers: 8
normalize: null

9
examples/aishell/run.sh Normal file
View File

@ -0,0 +1,9 @@
cmd="funasr_cli/cli/train_cli.py"
python $cmd \
--config-path "/Users/zhifu/funasr_github/test_local/funasr_cli_egs" \
--config-name "config.yaml" \
+token_list="/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt" \
+train_data_set_list="/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl" \
+output_dir="/nfs/zhifu.gzf/ckpt/funasr2/exp1"

View File

@ -0,0 +1,120 @@
# network architecture
model: funasr.cli.models.paraformer:Paraformer
model_conf:
ctc_weight: 0.0
lsm_weight: 0.1
length_normalized_loss: true
predictor_weight: 1.0
predictor_bias: 1
sampling_ratio: 0.75
# encoder
encoder: sanm
encoder_conf:
output_size: 512
attention_heads: 4
linear_units: 2048
num_blocks: 50
dropout_rate: 0.1
positional_dropout_rate: 0.1
attention_dropout_rate: 0.1
input_layer: pe
pos_enc_class: SinusoidalPositionEncoder
normalize_before: true
kernel_size: 11
sanm_shfit: 0
selfattention_layer_type: sanm
# decoder
decoder: paraformer_decoder_sanm
decoder_conf:
attention_heads: 4
linear_units: 2048
num_blocks: 16
dropout_rate: 0.1
positional_dropout_rate: 0.1
self_attention_dropout_rate: 0.1
src_attention_dropout_rate: 0.1
att_layer_num: 16
kernel_size: 11
sanm_shfit: 0
predictor: cif_predictor_v2
predictor_conf:
idim: 512
threshold: 1.0
l_order: 1
r_order: 1
tail_threshold: 0.45
# frontend related
frontend: wav_frontend
frontend_conf:
fs: 16000
window: hamming
n_mels: 80
frame_length: 25
frame_shift: 10
lfr_m: 7
lfr_n: 6
specaug: specaug_lfr
specaug_conf:
apply_time_warp: false
time_warp_window: 5
time_warp_mode: bicubic
apply_freq_mask: true
freq_mask_width_range:
- 0
- 30
lfr_rate: 6
num_freq_mask: 1
apply_time_mask: true
time_mask_width_range:
- 0
- 12
num_time_mask: 1
train_conf:
accum_grad: 1
grad_clip: 5
max_epoch: 150
val_scheduler_criterion:
- valid
- acc
best_model_criterion:
- - valid
- acc
- max
keep_nbest_models: 10
log_interval: 50
optim: adam
optim_conf:
lr: 0.0005
scheduler: warmuplr
scheduler_conf:
warmup_steps: 30000
dataset_conf:
data_names: speech,text
data_types: sound,text
shuffle: True
shuffle_conf:
shuffle_size: 2048
sort_size: 500
batch_conf:
batch_type: example
batch_size: 2
num_workers: 8
split_with_space: true
input_size: 560
ctc_conf:
dropout_rate: 0.0
ctc_type: builtin
reduce: true
ignore_nan_grad: true
normalize: null

View File

@ -0,0 +1,12 @@
cmd="funasr/cli/train_cli.py"
python $cmd \
+model_pretrain="/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch" \
+token_list="/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt" \
+train_data_set_list="/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl" \
+output_dir="/Users/zhifu/Downloads/ckpt/funasr2/exp2" \
+device="cpu"
#--config-path "/Users/zhifu/funasr_github/examples/industrial_data_pretraining/paraformer-large/conf" \
#--config-name "finetune.yaml" \

View File

@ -193,6 +193,7 @@ class Paraformer(nn.Module):
self.decoder.embed = None self.decoder.embed = None
self.use_1st_decoder_loss = use_1st_decoder_loss self.use_1st_decoder_loss = use_1st_decoder_loss
self.length_normalized_loss = length_normalized_loss
def forward( def forward(
self, self,
@ -302,6 +303,8 @@ class Paraformer(nn.Module):
stats["loss"] = torch.clone(loss.detach()) stats["loss"] = torch.clone(loss.detach())
# force_gatherable: to-device and to-tensor if scalar for DataParallel # force_gatherable: to-device and to-tensor if scalar for DataParallel
if self.length_normalized_loss:
batch_size = (text_lengths + self.predictor_bias).sum()
loss, stats, weight = force_gatherable((loss, stats, batch_size), loss.device) loss, stats, weight = force_gatherable((loss, stats, batch_size), loss.device)
return loss, stats, weight return loss, stats, weight

View File

@ -25,17 +25,25 @@ from funasr.utils.dynamic_import import dynamic_import
import torch.distributed as dist import torch.distributed as dist
from torch.nn.parallel import DistributedDataParallel as DDP from torch.nn.parallel import DistributedDataParallel as DDP
from torch.distributed.fsdp import FullyShardedDataParallel as FSDP from torch.distributed.fsdp import FullyShardedDataParallel as FSDP
from funasr.utils.download_from_hub import download_model
def preprocess_config(cfg: DictConfig): def preprocess_config(cfg: DictConfig):
for key, value in cfg.items(): for key, value in cfg.items():
if value == 'None': if value == 'None':
cfg[key] = None cfg[key] = None
@hydra.main(config_name=None, version_base=None)
def main_hydra(kwargs: DictConfig):
import pdb; pdb.set_trace()
if kwargs.get("model_pretrain"):
kwargs = download_model(**kwargs)
import pdb;
pdb.set_trace()
main(**kwargs)
@hydra.main() def main(**kwargs):
def main(kwargs: DictConfig):
# preprocess_config(kwargs) # preprocess_config(kwargs)
# import pdb; pdb.set_trace() # import pdb; pdb.set_trace()
# set random seed # set random seed
@ -76,10 +84,10 @@ def main(kwargs: DictConfig):
# init_param # init_param
init_param = kwargs.get("init_param", None) init_param = kwargs.get("init_param", None)
if init_param is not None: if init_param is not None:
init_param = eval(init_param) init_param = init_param
if isinstance(init_param, Sequence): if isinstance(init_param, Sequence):
init_param = (init_param,) init_param = (init_param,)
logging.info("init_param is not None: ", init_param) logging.info("init_param is not None: %s", init_param)
for p in init_param: for p in init_param:
logging.info(f"Loading pretrained params from {p}") logging.info(f"Loading pretrained params from {p}")
load_pretrained_model( load_pretrained_model(
@ -99,7 +107,7 @@ def main(kwargs: DictConfig):
freeze_param = eval(freeze_param) freeze_param = eval(freeze_param)
if isinstance(freeze_param, Sequence): if isinstance(freeze_param, Sequence):
freeze_param = (freeze_param,) freeze_param = (freeze_param,)
logging.info("freeze_param is not None: ", freeze_param) logging.info("freeze_param is not None: %s", freeze_param)
for t in freeze_param: for t in freeze_param:
for k, p in model.named_parameters(): for k, p in model.named_parameters():
if k.startswith(t + ".") or k == t: if k.startswith(t + ".") or k == t:
@ -138,9 +146,15 @@ def main(kwargs: DictConfig):
dataloader_tr = torch.utils.data.DataLoader(dataset_tr, dataloader_tr = torch.utils.data.DataLoader(dataset_tr,
collate_fn=dataset_tr.collator, collate_fn=dataset_tr.collator,
batch_sampler=batch_sampler, batch_sampler=batch_sampler,
num_workers=kwargs.get("num_workers", 0), num_workers=kwargs.get("dataset_conf").get("num_workers", 4),
pin_memory=True) pin_memory=True)
if (use_ddp or use_fsdp) and dist.get_rank() == 0 or not (use_ddp or use_fsdp) and local_rank == 0:
os.makedirs(kwargs.get("output_dir", "./"), exist_ok=True)
yaml_file = os.path.join(kwargs.get("output_dir", "./"), "config.yaml")
OmegaConf.save(config=kwargs, f=yaml_file)
logging.info("config.yaml is saved to: %s", yaml_file)
trainer = Trainer( trainer = Trainer(
model=model, model=model,
optim=optim, optim=optim,
@ -160,4 +174,4 @@ def main(kwargs: DictConfig):
if __name__ == "__main__": if __name__ == "__main__":
main() main_hydra()

View File

@ -2,6 +2,7 @@ import torch
import os import os
from funasr.torch_utils.device_funcs import to_device from funasr.torch_utils.device_funcs import to_device
import logging import logging
import time
from tqdm import tqdm from tqdm import tqdm
from contextlib import nullcontext from contextlib import nullcontext
import torch.distributed as dist import torch.distributed as dist
@ -53,20 +54,28 @@ class Trainer:
self.dataloader_train = dataloader_train self.dataloader_train = dataloader_train
self.dataloader_val = dataloader_val self.dataloader_val = dataloader_val
self.output_dir = kwargs.get('output_dir', './') self.output_dir = kwargs.get('output_dir', './')
self.resume = kwargs.get('resume', None) self.resume = kwargs.get('resume', True)
self.start_epoch = 1 self.start_epoch = 1
self.max_epoch = kwargs.get('max_epoch', 100) self.max_epoch = kwargs.get('max_epoch', 100)
self.local_rank = local_rank self.local_rank = local_rank
self.rank = dist.get_rank()
self.world_size = dist.get_world_size()
self.use_ddp = use_ddp self.use_ddp = use_ddp
self.use_fsdp = use_fsdp self.use_fsdp = use_fsdp
self.device = torch.device("cuda", local_rank) self.device = next(model.parameters()).device
self.kwargs = kwargs self.kwargs = kwargs
if self.resume: if self.resume:
self._resume_checkpoint(self.resume) self._resume_checkpoint(self.resume)
try:
rank = dist.get_rank()
world_size = dist.get_world_size()
except:
rank = 0
world_size = 1
logging.warning("distributed is not initialized, only single shard")
self.rank = rank
self.world_size = world_size
def _save_checkpoint(self, epoch): def _save_checkpoint(self, epoch):
""" """
Saves a checkpoint containing the model's state, the optimizer's state, Saves a checkpoint containing the model's state, the optimizer's state,
@ -117,6 +126,7 @@ class Trainer:
if dist.get_rank() == 0: if dist.get_rank() == 0:
self._save_checkpoint(epoch) self._save_checkpoint(epoch)
self.scheduler.step() self.scheduler.step()
break
def _train_epoch(self, epoch): def _train_epoch(self, epoch):
""" """
@ -132,13 +142,21 @@ class Trainer:
accum_grad = self.kwargs.get("accum_grad", 1) accum_grad = self.kwargs.get("accum_grad", 1)
# Initialize the gradient accumulation # Initialize the gradient accumulation
self.optim.zero_grad() self.optim.zero_grad()
speed_stats = {}
time5 = time.perf_counter()
for batch_idx, batch in enumerate(self.dataloader_train): for batch_idx, batch in enumerate(self.dataloader_train):
time1 = time.perf_counter()
speed_stats["data_load"] = f"{time1-time5:0.3f}"
# import pdb;
# pdb.set_trace()
batch = to_device(batch, self.device) batch = to_device(batch, self.device)
my_context = self.model.no_sync if batch_idx % accum_grad != 0 else nullcontext my_context = self.model.no_sync if batch_idx % accum_grad != 0 else nullcontext
with my_context(): with my_context():
time2 = time.perf_counter()
retval = self.model(**batch) retval = self.model(**batch)
time3 = time.perf_counter()
speed_stats["forward_time"] = f"{time3 - time2:0.3f}"
loss, stats, weight = retval loss, stats, weight = retval
stats = {k: v for k, v in stats.items() if v is not None} stats = {k: v for k, v in stats.items() if v is not None}
if self.use_ddp or self.use_fsdp: if self.use_ddp or self.use_fsdp:
@ -154,6 +172,8 @@ class Trainer:
# Scale the loss since we're not updating for every mini-batch # Scale the loss since we're not updating for every mini-batch
loss = loss / accum_grad loss = loss / accum_grad
loss.backward() loss.backward()
time4 = time.perf_counter()
speed_stats["backward_time"] = f"{time4 - time3:0.3f}"
# Perform an optimizer step only after accumulating enough gradients # Perform an optimizer step only after accumulating enough gradients
if (batch_idx + 1) % accum_grad == 0 or (batch_idx + 1) == len(self.dataloader_train): if (batch_idx + 1) % accum_grad == 0 or (batch_idx + 1) == len(self.dataloader_train):
@ -176,12 +196,26 @@ class Trainer:
self.scheduler.step() self.scheduler.step()
# Clear gradients for the next accumulation stage # Clear gradients for the next accumulation stage
self.optim.zero_grad() self.optim.zero_grad()
total_time = f"{time.perf_counter() - time5:0.3f}"
time5 = time.perf_counter()
speed_stats["optim_time"] = f"{time5 - time4:0.3f}"
speed_stats["total_time"] = total_time
pbar.update(1) pbar.update(1)
if self.local_rank == 0: if self.local_rank == 0:
pbar.set_description( description = (
f"Training Epoch: {epoch + 1}/{self.max_epoch}, step {batch_idx}/{len(self.dataloader_train)} (loss: {loss.detach().float():.3f}, {[(k, round(v.cpu().item(), 3)) for k, v in stats.items()]})") f"Epoch: {epoch + 1}/{self.max_epoch}, "
f"step {batch_idx}/{len(self.dataloader_train)}, "
f"{speed_stats}, "
f"(loss: {loss.detach().float():.3f}), "
f"{[(k, round(v.cpu().item(), 3)) for k, v in stats.items()]}"
)
pbar.set_description(description)
if batch_idx == 2:
break
pbar.close() pbar.close()
def _validate_epoch(self, epoch): def _validate_epoch(self, epoch):

View File

@ -6,6 +6,7 @@ import kaldiio
import librosa import librosa
import torchaudio import torchaudio
import time import time
import logging
def load_audio(audio_path: str, fs: int=16000): def load_audio(audio_path: str, fs: int=16000):
audio = None audio = None
@ -41,8 +42,7 @@ class IndexedDatasetJsonl(torch.utils.data.Dataset):
def __init__(self, path): def __init__(self, path):
super().__init__() super().__init__()
# data_parallel_size = dist.get_world_size()
data_parallel_size = 1
contents = [] contents = []
with open(path, encoding='utf-8') as fin: with open(path, encoding='utf-8') as fin:
for line in fin: for line in fin:
@ -66,12 +66,20 @@ class IndexedDatasetJsonl(torch.utils.data.Dataset):
self.contents = [] self.contents = []
total_num = len(contents) total_num = len(contents)
num_per_rank = total_num // data_parallel_size try:
# rank = dist.get_rank() rank = dist.get_rank()
rank = 0 world_size = dist.get_world_size()
except:
rank = 0
world_size = 1
logging.warning("distributed is not initialized, only single shard")
num_per_rank = total_num // world_size
# rank = 0
# import ipdb; ipdb.set_trace() # import ipdb; ipdb.set_trace()
self.contents = contents[rank * num_per_rank:(rank + 1) * num_per_rank] self.contents = contents[rank * num_per_rank:(rank + 1) * num_per_rank]
logging.info("in rank: {}, num of samplers: {}, total_num of samplers across ranks: {}".format(rank, len(self.contents), len(contents)))
def __len__(self): def __len__(self):
return len(self.contents) return len(self.contents)

View File

@ -0,0 +1,63 @@
import os
from omegaconf import OmegaConf
import torch
from funasr.utils.name_maps_from_hub import name_maps_ms, name_maps_hf
def download_model(**kwargs):
model_hub = kwargs.get("model_hub", "ms")
if model_hub == "ms":
kwargs = download_fr_ms(**kwargs)
return kwargs
def download_fr_ms(**kwargs):
model_or_path = kwargs.get("model_pretrain")
model_revision = kwargs.get("model_pretrain_revision")
if not os.path.exists(model_or_path):
model_or_path = get_or_download_model_dir(model_or_path, model_revision, third_party="funasr")
config = os.path.join(model_or_path, "config.yaml")
assert os.path.exists(config), "{} is not exist!".format(config)
cfg = OmegaConf.load(config)
kwargs = OmegaConf.merge(cfg, kwargs)
init_param = os.path.join(model_or_path, "model.pb")
kwargs["init_param"] = init_param
kwargs["token_list"] = os.path.join(model_or_path, "tokens.txt")
return kwargs
def get_or_download_model_dir(
model,
model_revision=None,
third_party=None):
""" Get local model directory or download model if necessary.
Args:
model (str): model id or path to local model directory.
model_revision (str, optional): model version number.
third_party (str, optional): in which third party library
this function is called.
"""
from modelscope.hub.check_model import check_local_model_is_latest
from modelscope.hub.snapshot_download import snapshot_download
from modelscope.utils.constant import Invoke, ThirdParty
if os.path.exists(model):
model_cache_dir = model if os.path.isdir(
model) else os.path.dirname(model)
check_local_model_is_latest(
model_cache_dir,
user_agent={
Invoke.KEY: Invoke.LOCAL_TRAINER,
ThirdParty.KEY: third_party
})
else:
model_cache_dir = snapshot_download(
model,
revision=model_revision,
user_agent={
Invoke.KEY: Invoke.TRAINER,
ThirdParty.KEY: third_party
})
return model_cache_dir

View File

@ -0,0 +1,16 @@
name_maps_ms = {
"paraformer-zh": "damo/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch",
"paraformer-zh-spk": "damo/speech_paraformer-large-vad-punc-spk_asr_nat-zh-cn",
"paraformer-en": "damo/speech_paraformer-large-vad-punc_asr_nat-en-16k-common-vocab10020",
"paraformer-en-spk": "damo/speech_paraformer-large-vad-punc_asr_nat-en-16k-common-vocab10020",
"paraformer-zh-streaming": "damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-online",
"fsmn-vad": "damo/speech_fsmn_vad_zh-cn-16k-common-pytorch",
"ct-punc": "damo/punc_ct-transformer_cn-en-common-vocab471067-large",
"fa-zh": "damo/speech_timestamp_prediction-v1-16k-offline",
}
name_maps_hf = {
}

View File

@ -0,0 +1 @@
{}

View File

@ -0,0 +1,154 @@
hydra:
run:
dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
sweep:
dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task: []
job:
name: train_cli
chdir: null
override_dirname: ''
id: ???
num: ???
config_name: null
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.1'
cwd: /Users/zhifu/funasr_github
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: /Users/zhifu/funasr_github/funasr/cli
schema: file
provider: main
- path: ''
schema: structured
provider: schema
output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/10-35-57
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@ -0,0 +1 @@
[]

View File

@ -0,0 +1 @@
{}

View File

@ -0,0 +1,154 @@
hydra:
run:
dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
sweep:
dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task: []
job:
name: train_cli
chdir: null
override_dirname: ''
id: ???
num: ???
config_name: null
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.1'
cwd: /Users/zhifu/funasr_github
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: /Users/zhifu/funasr_github/funasr/cli
schema: file
provider: main
- path: ''
schema: structured
provider: schema
output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/10-37-01
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@ -0,0 +1 @@
[]

View File

@ -0,0 +1 @@
{}

View File

@ -0,0 +1,154 @@
hydra:
run:
dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
sweep:
dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task: []
job:
name: train_cli
chdir: null
override_dirname: ''
id: ???
num: ???
config_name: null
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.1'
cwd: /Users/zhifu/funasr_github
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: /Users/zhifu/funasr_github/funasr/cli
schema: file
provider: main
- path: ''
schema: structured
provider: schema
output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/10-39-53
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@ -0,0 +1 @@
[]

View File

@ -0,0 +1,74 @@
model: funasr.cli.models.paraformer:Paraformer
model_conf:
ctc_weight: 0.3
lsm_weight: 0.1
length_normalized_loss: false
predictor_weight: 1.0
sampling_ratio: 0.4
use_1st_decoder_loss: true
frontend: wav_frontend
frontend_conf:
fs: 16000
window: hamming
n_mels: 80
frame_length: 25
frame_shift: 10
lfr_m: 1
lfr_n: 1
train_conf:
accum_grad: 1
grad_clip: 5
max_epoch: 150
val_scheduler_criterion:
- valid
- acc
best_model_criterion:
- - valid
- acc
- max
keep_nbest_models: 10
log_interval: 50
optim: adam
optim_conf:
lr: 0.0005
scheduler: warmuplr
scheduler_conf:
warmup_steps: 30000
specaug: specaug
specaug_conf:
apply_time_warp: true
time_warp_window: 5
time_warp_mode: bicubic
apply_freq_mask: true
freq_mask_width_range:
- 0
- 30
num_freq_mask: 2
apply_time_mask: true
time_mask_width_range:
- 0
- 40
num_time_mask: 2
predictor: cif_predictor
predictor_conf:
idim: 256
threshold: 1.0
l_order: 1
r_order: 1
tail_threshold: 0.45
dataset_conf:
data_names: speech,text
data_types: sound,text
shuffle: true
shuffle_conf:
shuffle_size: 2048
sort_size: 500
batch_conf:
batch_type: example
batch_size: 2
num_workers: 8
normalize: null
model_pretrain: damo/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch
token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
output_dir: /nfs/zhifu.gzf/ckpt/funasr2/exp2

View File

@ -0,0 +1,158 @@
hydra:
run:
dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
sweep:
dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task:
- +model_pretrain=damo/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2
job:
name: train_cli
chdir: null
override_dirname: +model_pretrain=damo/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
id: ???
num: ???
config_name: finetune.yaml
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.1'
cwd: /Users/zhifu/funasr_github
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: /Users/zhifu/funasr_github/examples/industrial_data_pretraining/paraformer-large/conf
schema: file
provider: main
- path: ''
schema: structured
provider: schema
output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/10-41-29
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@ -0,0 +1,4 @@
- +model_pretrain=damo/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2

View File

@ -0,0 +1,4 @@
model_pretrain: damo/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch
token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
output_dir: /nfs/zhifu.gzf/ckpt/funasr2/exp2

View File

@ -0,0 +1,158 @@
hydra:
run:
dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
sweep:
dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task:
- +model_pretrain=damo/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2
job:
name: train_cli
chdir: null
override_dirname: +model_pretrain=damo/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
id: ???
num: ???
config_name: null
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.1'
cwd: /Users/zhifu/funasr_github
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: /Users/zhifu/funasr_github/funasr/cli
schema: file
provider: main
- path: ''
schema: structured
provider: schema
output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/10-44-08
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@ -0,0 +1,4 @@
- +model_pretrain=damo/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2

View File

@ -0,0 +1,4 @@
model_pretrain: damo/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch
token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
output_dir: /nfs/zhifu.gzf/ckpt/funasr2/exp2

View File

@ -0,0 +1,158 @@
hydra:
run:
dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
sweep:
dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task:
- +model_pretrain=damo/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2
job:
name: train_cli
chdir: null
override_dirname: +model_pretrain=damo/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
id: ???
num: ???
config_name: null
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.1'
cwd: /Users/zhifu/funasr_github
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: /Users/zhifu/funasr_github/funasr/cli
schema: file
provider: main
- path: ''
schema: structured
provider: schema
output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/10-45-48
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@ -0,0 +1,4 @@
- +model_pretrain=damo/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2

View File

@ -0,0 +1,4 @@
model_pretrain: /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
output_dir: /nfs/zhifu.gzf/ckpt/funasr2/exp2

View File

@ -0,0 +1,158 @@
hydra:
run:
dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
sweep:
dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task:
- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2
job:
name: train_cli
chdir: null
override_dirname: +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
id: ???
num: ???
config_name: null
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.1'
cwd: /Users/zhifu/funasr_github
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: /Users/zhifu/funasr_github/funasr/cli
schema: file
provider: main
- path: ''
schema: structured
provider: schema
output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/11-26-16
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@ -0,0 +1,4 @@
- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2

View File

@ -0,0 +1,4 @@
model_pretrain: /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
output_dir: /nfs/zhifu.gzf/ckpt/funasr2/exp2

View File

@ -0,0 +1,155 @@
hydra:
run:
dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
sweep:
dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task:
- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2
job:
name: train_cli
chdir: null
override_dirname: +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
id: ???
num: ???
config_name: null
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: /Users/zhifu/funasr_github
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: ''
schema: structured
provider: schema
output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/11-32-18
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@ -0,0 +1,4 @@
- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2

View File

@ -0,0 +1 @@
[2023-12-11 11:32:54,061][root][INFO] - Loading pretrained params from /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb

View File

@ -0,0 +1,5 @@
model_pretrain: /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
output_dir: /nfs/zhifu.gzf/ckpt/funasr2/exp2
device: cpu

View File

@ -0,0 +1,156 @@
hydra:
run:
dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
sweep:
dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task:
- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2
- +device=cpu
job:
name: train_cli
chdir: null
override_dirname: +device=cpu,+model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
id: ???
num: ???
config_name: null
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: /Users/zhifu/funasr_github
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: ''
schema: structured
provider: schema
output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/11-33-27
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@ -0,0 +1,5 @@
- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2
- +device=cpu

View File

@ -0,0 +1 @@
[2023-12-11 11:33:45,313][root][INFO] - Loading pretrained params from /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb

View File

@ -0,0 +1,5 @@
model_pretrain: /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
output_dir: /nfs/zhifu.gzf/ckpt/funasr2/exp2
device: cpu

View File

@ -0,0 +1,156 @@
hydra:
run:
dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
sweep:
dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task:
- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2
- +device=cpu
job:
name: train_cli
chdir: null
override_dirname: +device=cpu,+model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
id: ???
num: ???
config_name: null
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: /Users/zhifu/funasr_github
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: ''
schema: structured
provider: schema
output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/11-36-25
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@ -0,0 +1,5 @@
- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2
- +device=cpu

View File

@ -0,0 +1,2 @@
[2023-12-11 11:36:30,624][root][INFO] - Loading pretrained params from /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb
[2023-12-11 11:36:31,819][root][WARNING] - distributed is not initialized, only single shard

View File

@ -0,0 +1,5 @@
model_pretrain: /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
output_dir: /Users/zhifu.gzf/ckpt/funasr2/exp2
device: cpu

View File

@ -0,0 +1,156 @@
hydra:
run:
dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
sweep:
dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task:
- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/Users/zhifu.gzf/ckpt/funasr2/exp2
- +device=cpu
job:
name: train_cli
chdir: null
override_dirname: +device=cpu,+model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/Users/zhifu.gzf/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
id: ???
num: ???
config_name: null
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: /Users/zhifu/funasr_github
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: ''
schema: structured
provider: schema
output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/11-37-13
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@ -0,0 +1,5 @@
- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/Users/zhifu.gzf/ckpt/funasr2/exp2
- +device=cpu

View File

@ -0,0 +1,2 @@
[2023-12-11 11:43:26,900][root][INFO] - Loading pretrained params from /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb
[2023-12-11 11:43:28,209][root][WARNING] - distributed is not initialized, only single shard

View File

@ -0,0 +1,5 @@
model_pretrain: /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
output_dir: /Users/zhifu/Downloads/ckpt/funasr2/exp2
device: cpu

View File

@ -0,0 +1,156 @@
hydra:
run:
dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
sweep:
dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task:
- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
- +device=cpu
job:
name: train_cli
chdir: null
override_dirname: +device=cpu,+model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
id: ???
num: ???
config_name: null
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: /Users/zhifu/funasr_github
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: ''
schema: structured
provider: schema
output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/11-44-36
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@ -0,0 +1,5 @@
- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
- +device=cpu

View File

@ -0,0 +1,2 @@
[2023-12-11 11:44:48,432][root][INFO] - Loading pretrained params from /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb
[2023-12-11 11:44:49,667][root][WARNING] - distributed is not initialized, only single shard

View File

@ -0,0 +1,5 @@
model_pretrain: /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
output_dir: /Users/zhifu/Downloads/ckpt/funasr2/exp2
device: cpu

View File

@ -0,0 +1,156 @@
hydra:
run:
dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
sweep:
dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task:
- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
- +device=cpu
job:
name: train_cli
chdir: null
override_dirname: +device=cpu,+model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
id: ???
num: ???
config_name: null
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: /Users/zhifu/funasr_github
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: ''
schema: structured
provider: schema
output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/11-48-01
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@ -0,0 +1,5 @@
- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
- +device=cpu

View File

@ -0,0 +1,3 @@
[2023-12-11 11:48:05,427][root][INFO] - Loading pretrained params from /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb
[2023-12-11 11:48:07,007][root][WARNING] - distributed is not initialized, only single shard
[2023-12-11 11:48:07,026][root][WARNING] - distributed is not initialized, only single shard

View File

@ -0,0 +1,5 @@
model_pretrain: /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
output_dir: /Users/zhifu/Downloads/ckpt/funasr2/exp2
device: cpu

View File

@ -0,0 +1,156 @@
hydra:
run:
dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
sweep:
dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task:
- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
- +device=cpu
job:
name: train_cli
chdir: null
override_dirname: +device=cpu,+model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
id: ???
num: ???
config_name: null
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: /Users/zhifu/funasr_github
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: ''
schema: structured
provider: schema
output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/13-06-39
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@ -0,0 +1,5 @@
- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
- +device=cpu

View File

@ -0,0 +1,5 @@
model_pretrain: /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
output_dir: /Users/zhifu/Downloads/ckpt/funasr2/exp2
device: cpu

View File

@ -0,0 +1,156 @@
hydra:
run:
dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
sweep:
dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task:
- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
- +device=cpu
job:
name: train_cli
chdir: null
override_dirname: +device=cpu,+model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
id: ???
num: ???
config_name: null
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: /Users/zhifu/funasr_github
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: ''
schema: structured
provider: schema
output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/13-07-13
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@ -0,0 +1,5 @@
- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
- +device=cpu

View File

@ -0,0 +1,3 @@
[2023-12-11 13:07:30,599][root][INFO] - Loading pretrained params from /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb
[2023-12-11 13:07:31,757][root][WARNING] - distributed is not initialized, only single shard
[2023-12-11 13:07:31,859][root][WARNING] - distributed is not initialized, only single shard

View File

@ -0,0 +1,5 @@
model_pretrain: /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
output_dir: /Users/zhifu/Downloads/ckpt/funasr2/exp2
device: cpu

View File

@ -0,0 +1,156 @@
hydra:
run:
dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
sweep:
dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task:
- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
- +device=cpu
job:
name: train_cli
chdir: null
override_dirname: +device=cpu,+model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
id: ???
num: ???
config_name: null
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: /Users/zhifu/funasr_github
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: ''
schema: structured
provider: schema
output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/13-12-46
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@ -0,0 +1,5 @@
- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
- +device=cpu

View File

@ -0,0 +1,5 @@
[2023-12-11 13:12:51,157][root][INFO] - init_param is not None: ('/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb',)
[2023-12-11 13:12:51,157][root][INFO] - Loading pretrained params from /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb
[2023-12-11 13:12:52,231][root][WARNING] - distributed is not initialized, only single shard
[2023-12-11 13:12:52,266][root][INFO] - config.yaml is saved to: /Users/zhifu/Downloads/ckpt/funasr2/exp2/config.yaml
[2023-12-11 13:12:52,266][root][WARNING] - distributed is not initialized, only single shard

View File

@ -0,0 +1,5 @@
model_pretrain: /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
output_dir: /Users/zhifu/Downloads/ckpt/funasr2/exp2
device: cpu

View File

@ -0,0 +1,156 @@
hydra:
run:
dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
sweep:
dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task:
- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
- +device=cpu
job:
name: train_cli
chdir: null
override_dirname: +device=cpu,+model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
id: ???
num: ???
config_name: null
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: /Users/zhifu/funasr_github
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: ''
schema: structured
provider: schema
output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/13-16-01
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@ -0,0 +1,5 @@
- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
- +device=cpu

View File

@ -0,0 +1,6 @@
[2023-12-11 13:16:27,835][root][INFO] - init_param is not None: ('/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb',)
[2023-12-11 13:16:27,835][root][INFO] - Loading pretrained params from /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb
[2023-12-11 13:16:28,828][root][WARNING] - distributed is not initialized, only single shard
[2023-12-11 13:16:28,828][root][INFO] - in rank: 0, num of samplers: 640, total_num of samplers across ranks: 640
[2023-12-11 13:16:28,835][root][INFO] - config.yaml is saved to: /Users/zhifu/Downloads/ckpt/funasr2/exp2/config.yaml
[2023-12-11 13:16:28,835][root][WARNING] - distributed is not initialized, only single shard

View File

@ -0,0 +1,5 @@
model_pretrain: /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
output_dir: /Users/zhifu/Downloads/ckpt/funasr2/exp2
device: cpu

View File

@ -0,0 +1,156 @@
hydra:
run:
dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
sweep:
dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task:
- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
- +device=cpu
job:
name: train_cli
chdir: null
override_dirname: +device=cpu,+model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
id: ???
num: ???
config_name: null
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: /Users/zhifu/funasr_github
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: ''
schema: structured
provider: schema
output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/13-19-05
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@ -0,0 +1,5 @@
- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
- +device=cpu

View File

@ -0,0 +1,6 @@
[2023-12-11 13:19:09,868][root][INFO] - init_param is not None: ('/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb',)
[2023-12-11 13:19:09,868][root][INFO] - Loading pretrained params from /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb
[2023-12-11 13:19:11,006][root][WARNING] - distributed is not initialized, only single shard
[2023-12-11 13:19:11,006][root][INFO] - in rank: 0, num of samplers: 640, total_num of samplers across ranks: 640
[2023-12-11 13:19:11,014][root][INFO] - config.yaml is saved to: /Users/zhifu/Downloads/ckpt/funasr2/exp2/config.yaml
[2023-12-11 13:19:11,014][root][WARNING] - distributed is not initialized, only single shard

View File

@ -0,0 +1,5 @@
model_pretrain: /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
output_dir: /Users/zhifu/Downloads/ckpt/funasr2/exp2
device: cpu

View File

@ -0,0 +1,156 @@
hydra:
run:
dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
sweep:
dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task:
- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
- +device=cpu
job:
name: train_cli
chdir: null
override_dirname: +device=cpu,+model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
id: ???
num: ???
config_name: null
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: /Users/zhifu/funasr_github
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: ''
schema: structured
provider: schema
output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/13-24-15
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@ -0,0 +1,5 @@
- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
- +device=cpu

View File

@ -0,0 +1,6 @@
[2023-12-11 13:24:21,397][root][INFO] - init_param is not None: ('/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb',)
[2023-12-11 13:24:21,397][root][INFO] - Loading pretrained params from /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb
[2023-12-11 13:24:21,980][root][WARNING] - distributed is not initialized, only single shard
[2023-12-11 13:24:21,980][root][INFO] - in rank: 0, num of samplers: 640, total_num of samplers across ranks: 640
[2023-12-11 13:24:21,987][root][INFO] - config.yaml is saved to: /Users/zhifu/Downloads/ckpt/funasr2/exp2/config.yaml
[2023-12-11 13:24:21,987][root][WARNING] - distributed is not initialized, only single shard

View File

@ -0,0 +1,5 @@
model_pretrain: /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
output_dir: /Users/zhifu/Downloads/ckpt/funasr2/exp2
device: cpu

View File

@ -0,0 +1,156 @@
hydra:
run:
dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
sweep:
dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task:
- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
- +device=cpu
job:
name: train_cli
chdir: null
override_dirname: +device=cpu,+model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
id: ???
num: ???
config_name: null
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: /Users/zhifu/funasr_github
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: ''
schema: structured
provider: schema
output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/13-37-51
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@ -0,0 +1,5 @@
- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
- +device=cpu

View File

@ -0,0 +1,6 @@
[2023-12-11 13:38:00,018][root][INFO] - init_param is not None: ('/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb',)
[2023-12-11 13:38:00,019][root][INFO] - Loading pretrained params from /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb
[2023-12-11 13:38:01,162][root][WARNING] - distributed is not initialized, only single shard
[2023-12-11 13:38:01,162][root][INFO] - in rank: 0, num of samplers: 640, total_num of samplers across ranks: 640
[2023-12-11 13:38:01,248][root][INFO] - config.yaml is saved to: /Users/zhifu/Downloads/ckpt/funasr2/exp2/config.yaml
[2023-12-11 13:38:01,249][root][WARNING] - distributed is not initialized, only single shard

View File

@ -0,0 +1,5 @@
model_pretrain: /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
output_dir: /Users/zhifu/Downloads/ckpt/funasr2/exp2
device: cpu

View File

@ -0,0 +1,156 @@
hydra:
run:
dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
sweep:
dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
subdir: ${hydra.job.num}
launcher:
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
sweeper:
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
max_batch_size: null
params: null
help:
app_name: ${hydra.job.name}
header: '${hydra.help.app_name} is powered by Hydra.
'
footer: 'Powered by Hydra (https://hydra.cc)
Use --hydra-help to view Hydra specific help
'
template: '${hydra.help.header}
== Configuration groups ==
Compose your configuration from those groups (group=option)
$APP_CONFIG_GROUPS
== Config ==
Override anything in the config (foo.bar=value)
$CONFIG
${hydra.help.footer}
'
hydra_help:
template: 'Hydra (${hydra.runtime.version})
See https://hydra.cc for more info.
== Flags ==
$FLAGS_HELP
== Configuration groups ==
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
to command line)
$HYDRA_CONFIG_GROUPS
Use ''--cfg hydra'' to Show the Hydra config.
'
hydra_help: ???
hydra_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][HYDRA] %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
root:
level: INFO
handlers:
- console
loggers:
logging_example:
level: DEBUG
disable_existing_loggers: false
job_logging:
version: 1
formatters:
simple:
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
handlers:
console:
class: logging.StreamHandler
formatter: simple
stream: ext://sys.stdout
file:
class: logging.FileHandler
formatter: simple
filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
root:
level: INFO
handlers:
- console
- file
disable_existing_loggers: false
env: {}
mode: RUN
searchpath: []
callbacks: {}
output_subdir: .hydra
overrides:
hydra:
- hydra.mode=RUN
task:
- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
- +device=cpu
job:
name: train_cli
chdir: null
override_dirname: +device=cpu,+model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
id: ???
num: ???
config_name: null
env_set: {}
env_copy: []
config:
override_dirname:
kv_sep: '='
item_sep: ','
exclude_keys: []
runtime:
version: 1.3.2
version_base: '1.3'
cwd: /Users/zhifu/funasr_github
config_sources:
- path: hydra.conf
schema: pkg
provider: hydra
- path: ''
schema: structured
provider: schema
output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/13-39-37
choices:
hydra/env: default
hydra/callbacks: null
hydra/job_logging: default
hydra/hydra_logging: default
hydra/hydra_help: default
hydra/help: default
hydra/sweeper: basic
hydra/launcher: basic
hydra/output: default
verbose: false

View File

@ -0,0 +1,5 @@
- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
- +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
- +device=cpu

View File

@ -0,0 +1,6 @@
[2023-12-11 13:39:40,353][root][INFO] - init_param is not None: ('/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb',)
[2023-12-11 13:39:40,353][root][INFO] - Loading pretrained params from /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb
[2023-12-11 13:39:41,780][root][WARNING] - distributed is not initialized, only single shard
[2023-12-11 13:39:41,780][root][INFO] - in rank: 0, num of samplers: 640, total_num of samplers across ranks: 640
[2023-12-11 13:39:41,861][root][INFO] - config.yaml is saved to: /Users/zhifu/Downloads/ckpt/funasr2/exp2/config.yaml
[2023-12-11 13:39:41,862][root][WARNING] - distributed is not initialized, only single shard