K_step: 1000 accumulate_grad_batches: 1 audio_num_mel_bins: 128 audio_sample_rate: 44100 binarization_args: shuffle: false with_align: true with_f0: true with_hubert: true with_spk_embed: false with_wav: false binarizer_cls: preprocessing.SVCpre.SVCBinarizer binary_data_dir: data/binary/Unnamed check_val_every_n_epoch: 10 choose_test_manually: false clip_grad_norm: 1 config_path: training/config_nsf.yaml content_cond_steps: [] cwt_add_f0_loss: false cwt_hidden_size: 128 cwt_layers: 2 cwt_loss: l1 cwt_std_scale: 0.8 datasets: - opencpop debug: false dec_ffn_kernel_size: 9 dec_layers: 4 decay_steps: 60000 decoder_type: fft dict_dir: '' diff_decoder_type: wavenet diff_loss_type: l2 dilation_cycle_length: 4 dropout: 0.1 ds_workers: 4 dur_enc_hidden_stride_kernel: - 0,2,3 - 0,2,3 - 0,1,3 dur_loss: mse dur_predictor_kernel: 3 dur_predictor_layers: 5 enc_ffn_kernel_size: 9 enc_layers: 4 encoder_K: 8 encoder_type: fft endless_ds: false f0_bin: 256 f0_max: 1100.0 f0_min: 40.0 ffn_act: gelu ffn_padding: SAME fft_size: 2048 fmax: 16000 fmin: 40 fs2_ckpt: '' gaussian_start: true gen_dir_name: '' gen_tgt_spk_id: -1 hidden_size: 256 hop_size: 512 hubert_gpu: true hubert_path: checkpoints/hubert/hubert_soft.pt infer: false keep_bins: 128 lambda_commit: 0.25 lambda_energy: 0.0 lambda_f0: 1.0 lambda_ph_dur: 0.3 lambda_sent_dur: 1.0 lambda_uv: 1.0 lambda_word_dur: 1.0 load_ckpt: '' log_interval: 100 loud_norm: false lr: 0.0008 max_beta: 0.02 max_epochs: 3000 max_eval_sentences: 1 max_eval_tokens: 60000 max_frames: 42000 max_input_tokens: 60000 max_sentences: 14 max_tokens: 128000 max_updates: 1000000 mel_loss: ssim:0.5|l1:0.5 mel_vmax: 1.5 mel_vmin: -6.0 min_level_db: -120 no_fs2: true norm_type: gn num_ckpt_keep: 10 num_heads: 2 num_sanity_val_steps: 1 num_spk: 1 num_test_samples: 0 num_valid_plots: 10 optimizer_adam_beta1: 0.9 optimizer_adam_beta2: 0.98 out_wav_norm: false pe_ckpt: checkpoints/0102_xiaoma_pe/model_ckpt_steps_60000.ckpt pe_enable: false perform_enhance: true pitch_ar: false pitch_enc_hidden_stride_kernel: - 0,2,5 - 0,2,5 - 0,2,5 pitch_extractor: parselmouth pitch_loss: l2 pitch_norm: log pitch_type: frame pndm_speedup: 10 pre_align_args: allow_no_txt: false denoise: false forced_align: mfa txt_processor: zh_g2pM use_sox: true use_tone: false pre_align_cls: data_gen.singing.pre_align.SingingPreAlign predictor_dropout: 0.5 predictor_grad: 0.1 predictor_hidden: -1 predictor_kernel: 5 predictor_layers: 5 prenet_dropout: 0.5 prenet_hidden_size: 256 pretrain_fs_ckpt: '' processed_data_dir: xxx profile_infer: false raw_data_dir: data/raw/Unnamed ref_norm_layer: bn rel_pos: true reset_phone_dict: true residual_channels: 384 residual_layers: 20 save_best: true save_ckpt: true save_codes: - configs - modules - src - utils save_f0: true save_gt: false schedule_type: linear seed: 1234 sort_by_len: true speaker_id: Unnamed spec_max: - 0.47615352272987366 - 0.6125704050064087 - 0.7518845796585083 - 0.900716245174408 - 0.8935521841049194 - 0.9057011604309082 - 0.9648348689079285 - 0.9044283032417297 - 0.9109272360801697 - 0.9744535088539124 - 0.9476388692855835 - 0.9883336424827576 - 1.0821290016174316 - 1.046391248703003 - 0.9829667806625366 - 1.0163493156433105 - 0.9825412631034851 - 1.0021960735321045 - 1.052114725112915 - 1.128888726234436 - 1.186057209968567 - 1.112004280090332 - 1.1282787322998047 - 1.051572322845459 - 1.1104764938354492 - 1.176831603050232 - 1.13348388671875 - 1.1075258255004883 - 1.1696264743804932 - 1.0231049060821533 - 0.9303848743438721 - 1.1257890462875366 - 1.1610286235809326 - 1.0335885286331177 - 1.0645352602005005 - 1.0619306564331055 - 1.1310148239135742 - 1.1191954612731934 - 1.1307402849197388 - 1.2094721794128418 - 1.2683185338974 - 1.1212272644042969 - 1.1781182289123535 - 1.1501952409744263 - 0.9884514808654785 - 0.9226155281066895 - 0.9469702839851379 - 1.023751139640808 - 1.1348609924316406 - 1.087107539176941 - 0.9899962544441223 - 1.061837077140808 - 1.0341650247573853 - 0.9019684195518494 - 0.7986546158790588 - 0.7983465194702148 - 0.7755436301231384 - 0.701917290687561 - 0.7639197707176208 - 0.7503461837768555 - 0.6701087951660156 - 0.5326520800590515 - 0.6320568323135376 - 0.4748716950416565 - 0.41016310453414917 - 0.4754445552825928 - 0.4267503023147583 - 0.391481876373291 - 0.3118276298046112 - 0.3193877339363098 - 0.3111794888973236 - 0.3342774212360382 - 0.1353837102651596 - 0.16596835851669312 - 0.1730986088514328 - 0.2325316220521927 - 0.17107760906219482 - 0.10877621918916702 - 0.2612082064151764 - 0.11200784891843796 - 0.14075303077697754 - 0.07312829792499542 - -0.011712555773556232 - 0.1741427332162857 - 0.19782507419586182 - 0.03305494412779808 - 0.004054426681250334 - 0.1011907309293747 - 0.1317272037267685 - 0.014256341382861137 - 0.019952761009335518 - -0.1253873109817505 - -0.14854255318641663 - -0.14063480496406555 - -0.1331133395433426 - -0.28339776396751404 - -0.38559386134147644 - -0.2798943519592285 - -0.19351321458816528 - -0.23238061368465424 - -0.2850213944911957 - -0.20320385694503784 - -0.24087588489055634 - -0.15823237597942352 - -0.13949760794639587 - -0.19627133011817932 - -0.1920071393251419 - -0.19384469091892242 - -0.22403620183467865 - -0.18197931349277496 - -0.28423866629600525 - -0.26859334111213684 - -0.3213472068309784 - -0.3303631842136383 - -0.3835512697696686 - -0.3256210386753082 - -0.3938714265823364 - -0.4373253881931305 - -0.4146285951137543 - -0.4861420691013336 - -0.4018196761608124 - -0.46770456433296204 - -0.4100344479084015 - -0.5364681482315063 - -0.5802102088928223 - -0.5856970548629761 - -0.47378262877464294 - -0.36258620023727417 spec_min: - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 - -4.999994277954102 spk_cond_steps: [] stop_token_weight: 5.0 task_cls: training.task.SVC_task.SVCTask test_ids: [] test_input_dir: '' test_num: 0 test_prefixes: - test test_set_name: test timesteps: 1000 train_set_name: train use_crepe: false use_denoise: false use_energy_embed: false use_gt_dur: false use_gt_f0: false use_midi: false use_nsf: true use_pitch_embed: true use_pos_embed: true use_spk_embed: false use_spk_id: false use_split_spk_id: false use_uv: false use_var_enc: false use_vec: false val_check_interval: 1000 valid_num: 0 valid_set_name: valid vocoder: network.vocoders.nsf_hifigan.NsfHifiGAN vocoder_ckpt: checkpoints/nsf_hifigan/model warmup_updates: 2000 wav2spec_eps: 1e-6 weight_decay: 0 win_size: 2048 work_dir: checkpoints/Unnamed