Extract spectrogram and f0: python make_spect_f0.py
Generate training metadata: python make_metadata.py
My code is based on the above step!
Who can help me?
import os
import sys
import pickle
import numpy as np
import soundfile as sf
from scipy import signal
from librosa.filters import mel
from numpy.random import RandomState
from pysptk import sptk
from utils import butter_highpass
from utils import speaker_normalization
from utils import pySTFT
import torch
from autovc.model_bl import D_VECTOR
from collections import OrderedDict
mel_basis = mel(16000, 1024, fmin=90, fmax=7600, n_mels=80).T
min_level = np.exp(-100 / 20 * np.log(10))
b, a = butter_highpass(30, 16000, order=5)
C = D_VECTOR(dim_input=80, dim_cell=768, dim_emb=256).eval().cuda()
c_checkpoint = torch.load('assets/3000000-BL.ckpt')
new_state_dict = OrderedDict()
for key, val in c_checkpoint['model_b'].items():
new_key = key[7:]
new_state_dict[new_key] = val
C.load_state_dict(new_state_dict)
num_uttrs = 1
len_crop = 128
Extract spectrogram and f0: python make_spect_f0.py
Generate training metadata: python make_metadata.py
My code is based on the above step! Who can help me?
C = D_VECTOR(dim_input=80, dim_cell=768, dim_emb=256).eval().cuda() c_checkpoint = torch.load('assets/3000000-BL.ckpt') new_state_dict = OrderedDict() for key, val in c_checkpoint['model_b'].items(): new_key = key[7:] new_state_dict[new_key] = val C.load_state_dict(new_state_dict) num_uttrs = 1 len_crop = 128
spk2gen = pickle.load(open('assets/spk2gen.pkl', "rb"))
Modify as needed
rootDir = 'assets/wavs' targetDir_f0 = 'assets/raptf0' targetDir = 'assets/spmel'
dirName, subdirList, _ = next(os.walk(rootDir)) print('Found directory: %s' % dirName) speakers = [] for subdir in sorted(subdirList): print(subdir)
with open(os.path.join(rootDir, 'train.pkl'), 'wb') as handle: pickle.dump(speakers, handle)