Other Open Access

ESPnet2 pretrained model, kan-bayashi/jsut_tts_train_conformer_fastspeech2_transformer_teacher_raw_phn_jaconv_pyopenjtalk_accent_train.loss.ave, fs=24000, lang=jp

kan-bayashi


DataCite XML Export

<?xml version='1.0' encoding='utf-8'?>
<resource xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns="http://datacite.org/schema/kernel-4" xsi:schemaLocation="http://datacite.org/schema/kernel-4 http://schema.datacite.org/meta/kernel-4.1/metadata.xsd">
  <identifier identifierType="DOI">10.5281/zenodo.4391409</identifier>
  <creators>
    <creator>
      <creatorName>kan-bayashi</creatorName>
    </creator>
  </creators>
  <titles>
    <title>ESPnet2 pretrained model, kan-bayashi/jsut_tts_train_conformer_fastspeech2_transformer_teacher_raw_phn_jaconv_pyopenjtalk_accent_train.loss.ave, fs=24000, lang=jp</title>
  </titles>
  <publisher>Zenodo</publisher>
  <publicationYear>2020</publicationYear>
  <subjects>
    <subject>ESPnet</subject>
    <subject>deep-learning</subject>
    <subject>python</subject>
    <subject>pytorch</subject>
    <subject>speech-recognition</subject>
    <subject>speech-synthesis</subject>
    <subject>speech-translation</subject>
    <subject>machine-translation</subject>
  </subjects>
  <dates>
    <date dateType="Issued">2020-12-24</date>
  </dates>
  <resourceType resourceTypeGeneral="Other"/>
  <alternateIdentifiers>
    <alternateIdentifier alternateIdentifierType="url">https://zenodo.org/record/4391409</alternateIdentifier>
  </alternateIdentifiers>
  <relatedIdentifiers>
    <relatedIdentifier relatedIdentifierType="URL" relationType="IsSupplementTo">https://github.com/espnet/espnet</relatedIdentifier>
    <relatedIdentifier relatedIdentifierType="DOI" relationType="IsVersionOf">10.5281/zenodo.4391408</relatedIdentifier>
    <relatedIdentifier relatedIdentifierType="URL" relationType="IsPartOf">https://zenodo.org/communities/espnet</relatedIdentifier>
  </relatedIdentifiers>
  <rightsList>
    <rights rightsURI="https://creativecommons.org/licenses/by/4.0/legalcode">Creative Commons Attribution 4.0 International</rights>
    <rights rightsURI="info:eu-repo/semantics/openAccess">Open Access</rights>
  </rightsList>
  <descriptions>
    <description descriptionType="Abstract">This model was trained by kan-bayashi using jsut/tts1 recipe in &lt;a href="https://github.com/espnet/espnet/"&gt;espnet&lt;/a&gt;.
&lt;p&gt;&amp;nbsp;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Python API&lt;/strong&gt;&lt;pre&gt;&lt;code class="language-python"&gt;See https://github.com/espnet/espnet_model_zoo&lt;/code&gt;&lt;/pre&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Evaluate in the recipe&lt;/strong&gt;&lt;pre&gt;&lt;code class="language-bash"&gt;git clone https://github.com/espnet/espnet
cd espnet
git checkout 73e3fbe6b0050a401c5945d8210f8fa206c0dcc0
pip install -e .
cd egs2/jsut/tts1
# Download the model file here
./run.sh --skip_data_prep false --skip_train true --download_model kan-bayashi/jsut_tts_train_conformer_fastspeech2_transformer_teacher_raw_phn_jaconv_pyopenjtalk_accent_train.loss.ave&lt;/code&gt;
&lt;/pre&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Config&lt;/strong&gt;&lt;pre&gt;&lt;code&gt;config: conf/tuning/train_conformer_fastspeech2.yaml
print_config: false
log_level: INFO
dry_run: false
iterator_type: sequence
output_dir: exp/tts_train_conformer_fastspeech2_transformer_teacher_raw_phn_jaconv_pyopenjtalk_accent
ngpu: 1
seed: 0
num_workers: 1
num_att_plot: 3
dist_backend: nccl
dist_init_method: env://
dist_world_size: 4
dist_rank: 0
local_rank: 0
dist_master_addr: localhost
dist_master_port: 43445
dist_launcher: null
multiprocessing_distributed: true
cudnn_enabled: true
cudnn_benchmark: false
cudnn_deterministic: true
collect_stats: false
write_collected_feats: false
max_epoch: 200
patience: null
val_scheduler_criterion:
- valid
- loss
early_stopping_criterion:
- valid
- loss
- min
best_model_criterion:
-   - valid
    - loss
    - min
-   - train
    - loss
    - min
keep_nbest_models: 5
grad_clip: 1.0
grad_clip_type: 2.0
grad_noise: false
accum_grad: 1
no_forward_run: false
resume: true
train_dtype: float32
use_amp: false
log_interval: null
unused_parameters: false
use_tensorboard: true
use_wandb: false
wandb_project: null
wandb_id: null
pretrain_path: null
init_param: []
freeze_param: []
num_iters_per_epoch: 500
batch_size: 20
valid_batch_size: null
batch_bins: 18000000
valid_batch_bins: null
train_shape_file:
- exp/tts_train_transformer_raw_phn_jaconv_pyopenjtalk_accent/decode_use_teacher_forcingtrue_train.loss.ave/stats/train/text_shape.phn
- exp/tts_train_transformer_raw_phn_jaconv_pyopenjtalk_accent/decode_use_teacher_forcingtrue_train.loss.ave/stats/train/speech_shape
valid_shape_file:
- exp/tts_train_transformer_raw_phn_jaconv_pyopenjtalk_accent/decode_use_teacher_forcingtrue_train.loss.ave/stats/valid/text_shape.phn
- exp/tts_train_transformer_raw_phn_jaconv_pyopenjtalk_accent/decode_use_teacher_forcingtrue_train.loss.ave/stats/valid/speech_shape
batch_type: numel
valid_batch_type: null
fold_length:
- 150
- 240000
sort_in_batch: descending
sort_batch: descending
multiple_iterator: false
chunk_length: 500
chunk_shift_ratio: 0.5
num_cache_chunks: 1024
train_data_path_and_name_and_type:
-   - dump/raw/tr_no_dev/text
    - text
    - text
-   - exp/tts_train_transformer_raw_phn_jaconv_pyopenjtalk_accent/decode_use_teacher_forcingtrue_train.loss.ave/tr_no_dev/durations
    - durations
    - text_int
-   - dump/raw/tr_no_dev/wav.scp
    - speech
    - sound
valid_data_path_and_name_and_type:
-   - dump/raw/dev/text
    - text
    - text
-   - exp/tts_train_transformer_raw_phn_jaconv_pyopenjtalk_accent/decode_use_teacher_forcingtrue_train.loss.ave/dev/durations
    - durations
    - text_int
-   - dump/raw/dev/wav.scp
    - speech
    - sound
allow_variable_data_keys: false
max_cache_size: 0.0
max_cache_fd: 32
valid_max_cache_size: null
optim: adam
optim_conf:
    lr: 1.0
scheduler: noamlr
scheduler_conf:
    model_size: 384
    warmup_steps: 4000
token_list:
- 
- 
- '1'
- '2'
- '0'
- '3'
- '4'
- '-1'
- '5'
- a
- o
- '-2'
- i
- '-3'
- u
- e
- k
- n
- t
- '6'
- r
- '-4'
- s
- N
- m
- '7'
- sh
- d
- g
- w
- '8'
- U
- '-5'
- I
- cl
- h
- y
- b
- '9'
- j
- ts
- ch
- '-6'
- z
- p
- '-7'
- f
- ky
- ry
- '-8'
- gy
- '-9'
- hy
- ny
- '-10'
- by
- my
- '-11'
- '-12'
- '-13'
- py
- '-14'
- '-15'
- v
- '10'
- '-16'
- '-17'
- '11'
- '-21'
- '-20'
- '12'
- '-19'
- '13'
- '-18'
- '14'
- dy
- '15'
- ty
- '-22'
- '16'
- '18'
- '19'
- '17'
- 
odim: null
model_conf: {}
use_preprocessor: true
token_type: phn
bpemodel: null
non_linguistic_symbols: null
cleaner: jaconv
g2p: pyopenjtalk_accent
feats_extract: fbank
feats_extract_conf:
    fs: 24000
    fmin: 80
    fmax: 7600
    n_mels: 80
    hop_length: 300
    n_fft: 2048
    win_length: 1200
normalize: global_mvn
normalize_conf:
    stats_file: exp/tts_train_transformer_raw_phn_jaconv_pyopenjtalk_accent/decode_use_teacher_forcingtrue_train.loss.ave/stats/train/feats_stats.npz
tts: fastspeech2
tts_conf:
    adim: 384
    aheads: 2
    elayers: 4
    eunits: 1536
    dlayers: 4
    dunits: 1536
    positionwise_layer_type: conv1d
    positionwise_conv_kernel_size: 3
    duration_predictor_layers: 2
    duration_predictor_chans: 256
    duration_predictor_kernel_size: 3
    postnet_layers: 5
    postnet_filts: 5
    postnet_chans: 256
    use_masking: true
    encoder_normalize_before: true
    decoder_normalize_before: true
    reduction_factor: 1
    encoder_type: conformer
    decoder_type: conformer
    conformer_pos_enc_layer_type: rel_pos
    conformer_self_attn_layer_type: rel_selfattn
    conformer_activation_type: swish
    use_macaron_style_in_conformer: true
    use_cnn_in_conformer: true
    conformer_enc_kernel_size: 7
    conformer_dec_kernel_size: 31
    init_type: xavier_uniform
    transformer_enc_dropout_rate: 0.2
    transformer_enc_positional_dropout_rate: 0.2
    transformer_enc_attn_dropout_rate: 0.2
    transformer_dec_dropout_rate: 0.2
    transformer_dec_positional_dropout_rate: 0.2
    transformer_dec_attn_dropout_rate: 0.2
    pitch_predictor_layers: 5
    pitch_predictor_chans: 256
    pitch_predictor_kernel_size: 5
    pitch_predictor_dropout: 0.5
    pitch_embed_kernel_size: 1
    pitch_embed_dropout: 0.0
    stop_gradient_from_pitch_predictor: true
    energy_predictor_layers: 2
    energy_predictor_chans: 256
    energy_predictor_kernel_size: 3
    energy_predictor_dropout: 0.5
    energy_embed_kernel_size: 1
    energy_embed_dropout: 0.0
    stop_gradient_from_energy_predictor: false
pitch_extract: dio
pitch_extract_conf:
    fs: 24000
    n_fft: 2048
    hop_length: 300
    f0max: 400
    f0min: 80
    reduction_factor: 1
pitch_normalize: global_mvn
pitch_normalize_conf:
    stats_file: exp/tts_train_transformer_raw_phn_jaconv_pyopenjtalk_accent/decode_use_teacher_forcingtrue_train.loss.ave/stats/train/pitch_stats.npz
energy_extract: energy
energy_extract_conf:
    fs: 24000
    n_fft: 2048
    hop_length: 300
    win_length: 1200
    reduction_factor: 1
energy_normalize: global_mvn
energy_normalize_conf:
    stats_file: exp/tts_train_transformer_raw_phn_jaconv_pyopenjtalk_accent/decode_use_teacher_forcingtrue_train.loss.ave/stats/train/energy_stats.npz
required:
- output_dir
- token_list
distributed: true&lt;/code&gt;&lt;/pre&gt;&lt;/li&gt;
&lt;/ul&gt;</description>
  </descriptions>
</resource>
33
158
views
downloads
All versions This version
Views 3333
Downloads 158158
Data volume 44.6 GB44.6 GB
Unique views 3232
Unique downloads 101101

Share

Cite as