Closed AzamRabiee closed 6 years ago
That's weird. Could you share the log output?
Sorry, I meant stdout/stderr of train.py, not tf logs.
If nepoch equals 2000, then it should not stop at epoch 0.
Part of stdout:
Command line args: {'--checkpoint': None, '--checkpoint-dir': 'checkpoints', '--data-root': './data/cmu_arctic/', '--help': False, '--hparams': '', '--log-event-path': None, '--reset-optimizer': False, '--restore-parts': None, '--speaker-id': None} Hyperparameters: adam_beta1: 0.9 adam_beta2: 0.999 adam_eps: 1e-08 batch_size: 1 builder: wavenet checkpoint_interval: 10000 cin_channels: 80 clip_thresh: 1.0 dropout: 0.050000000000000044 fft_size: 1024 frame_shift_ms: None freq_axis_kernel_size: 3 gate_channels: 512 gin_channels: -1 hop_size: 256 initial_learning_rate: 0.001 kernel_size: 3 layers: 16 lr_schedule: noam_learning_rate_decay lr_schedule_kwargs: {} max_time_sec: None max_time_steps: 20000 min_level_db: -100 n_speakers: 7 name: wavenet_vocoder nepochs: 2000 num_mels: 80 num_workers: 2 pin_memory: True preset: presets: {} random_state: 1234 ref_level_db: 20 residual_channels: 256 sample_rate: 16000 save_optimizer_state: True silence_threshold: 2 skip_out_channels: 256 stacks: 2 test_eval_epoch_interval: 5 test_num_samples: None test_size: 0.0441 train_eval_interval: 10000 upsample_conditional_features: True upsample_scales: [16, 16] weight_decay: 0.0 weight_normalization: True Local conditioning enabled. Shape of a sample: (179, 80). [train]: length of the dataset is 7580 Speaker stats: {0: 1092, 3: 1081, 4: 1089, 6: 1079, 5: 1073, 1: 1095, 2: 1071} Local conditioning enabled. Shape of a sample: (123, 80). [test]: length of the dataset is 350 Speaker stats: {5: 59, 1: 37, 4: 43, 6: 53, 2: 61, 0: 46, 3: 51}
Sorry, Ive found the error: File "train.py", line 456, in eval_model save_waveplot(path, y_hat, y_target) File "train.py", line 405, in save_waveplot plt.figure(figsize=(16, 6)) ... _tkinter.TclError: no display name and no $DISPLAY environment variable
It seems there's no error messages. What problem are you seeing? I am not sure I understand you.
Ah, ok,
import matplotlib
matplotlib.use('Agg')
import matplotlib.pyplot as plt
I think this fixes your issue.
but your suggestion does not work. same error is raised! Anyway, I can go to more training epochs without saving the plot.
Did you put the snippet into the top of the train.py?
Yeap! It's solved by adding the snippet and 'export DISPLAY=mymachine.com:0.0' Thanks!
Is really 2000 nepochs required? I got almost good results with only 30 epochs!
No. I just set a big number. You can stop training anytime by hitting ctrl+c.
I'm closing this
@r9y9 Hi Ryuichi Thanks for your amazing work. After I read your readme.txt, I have a couple of questions.
Thank you so much.
@zctang I'm not sure if I understand you correctly. For the conditioning settings, you can find what I used at https://r9y9.github.io/wavenet_vocoder/.
2: 7 speakers were available when I run the experiments (http://festvox.org/cmu_arctic/cmu_arctic/). You can use 18 speakers if you want.
@zctang I'm not sure if I understand you correctly. For the conditioning settings, you can find what I used at https://r9y9.github.io/wavenet_vocoder/.
2: 7 speakers were available when I run the experiments (http://festvox.org/cmu_arctic/cmu_arctic/). You can use 18 speakers if you want.
@r9y9 Thanks, I understood the first question.However, for the second question, if I want use your code and also use 18 speakers, should I also need to fix the code from nnmnkwii? Because I saw the cmu_arctic.py in it.
Thanks for sharing your comprehensive code! I've just started reading and running your code. In hparams, nepochs is set to 2000; but it seems to stop at the first epoch, as it shows in tensorboard: it seems that the training loop is only depend on nepochs. Is there any other parameter to set for keep going the training process?