Skip to content

Recipes for commonvoice ASR and LID #129

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
wants to merge 116 commits into
base: persephone-asr
Choose a base branch
from
Open
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
116 commits
Select commit Hold shift + click to select a range
48b1e4e
commonvoice speech recognition recipe
neillu23 Jan 23, 2023
beab75c
update slurm configuration for rockfish
Feb 1, 2023
046b5f7
update data preparation for different languge
Feb 1, 2023
beb2ed5
update config and add cer scripts
Feb 15, 2023
ff0fd55
temporal remove data preparation for duration
Feb 15, 2023
f179db4
Add combination for multiple languages
Feb 15, 2023
f816ed3
Add language identification task for commonvoice
Feb 20, 2023
b524b84
Add Class Weighted Sampler for ASR and utterance-wise LID
Mar 24, 2023
07ddda6
Remove the seg_weighted_mode for sequence-level task
Mar 25, 2023
a2eff8e
Merge remote-tracking branch 'hyp/persephone-asr' into persephone-ref…
Mar 25, 2023
396e020
Update the LID trainer for merging the new dataloader
Mar 27, 2023
2ecdebf
add commonvoice config for rnnt transducer
Mar 27, 2023
d33abe9
Add fine-tuning code for pruned RNN-T, LID, and Both
Apr 22, 2023
3b7e8ac
Add LID decode scripts
neillu23 Apr 22, 2023
85282ac
Merge pull request #127 from hyperion-ml/persephone-asr
jesus-villalba May 3, 2023
35391de
new vox2 dataprep
jesus-villalba May 4, 2023
ebef851
update the np.str to np.str_
May 4, 2023
720bd6e
update np.str to np.str_
neillu23 May 4, 2023
845d2e0
Merge branch 'persephone-refactor' of https://github.com/neillu23/hyp…
May 4, 2023
b112ebd
Add empty __init__.py
May 5, 2023
cf861bc
fix new vox2 dataprep durations, scp -> RecordingSet
jesus-villalba May 8, 2023
c408f74
some fixes in sre21
jesus-villalba May 11, 2023
9c28408
update lid configs and np.str to str
May 13, 2023
7f43376
FiLM transducer
May 13, 2023
20c13e7
Add FiLMed Transducer
May 14, 2023
f8c84a9
remove unused function
May 14, 2023
05474de
Add decode script and configurations
May 15, 2023
cb4f20e
remove redundant code
May 16, 2023
2b61053
update fine-tune and decoding scripts for rnnt_film_transducer
May 17, 2023
ca5327a
update language id trainer to use chunk for training
May 18, 2023
27d579c
sre21 16k recipe finished
jesus-villalba May 20, 2023
df8a24f
add initialize model for joint-training and film training
May 20, 2023
159ff07
Add transducer and languageid joint training
May 21, 2023
29fdfb7
add asr_lid run script
May 21, 2023
3d33522
update film model
May 21, 2023
0c710a4
Merge branch 'persephone-refactor' of https://github.com/neillu23/hyp…
May 21, 2023
f243ded
Merge branch 'persephone-asr' of https://github.com/hyperion-ml/hyper…
neillu23 May 21, 2023
97c77f3
Merge branch 'persephone-refactor' of https://github.com/neillu23/hyp…
neillu23 May 21, 2023
a5971ab
update transducer_languageid joint model
May 21, 2023
92f33d3
update ASR and LID joint training code
May 22, 2023
190ea29
update configuration
May 22, 2023
16f8b49
update config with mean pruned rnn loss
May 22, 2023
1ebb219
update config to use mean transducer loss
May 22, 2023
c474869
update film parameter name
May 22, 2023
1f7e70b
update more options for film model
May 22, 2023
c6f5dee
add reduction option in fine-tune argument
May 22, 2023
7331d32
Merge branch 'persephone-refactor' of https://github.com/neillu23/hyp…
May 22, 2023
8e82143
update configuration
May 22, 2023
7135a83
Merge branch 'persephone-refactor' of https://github.com/neillu23/hyp…
May 22, 2023
c6ec4e2
fix film bug
May 23, 2023
90ee23c
Merge branch 'persephone-refactor' of https://github.com/neillu23/hyp…
May 23, 2023
a3d14ae
Merge branch 'persephone-refactor' of https://github.com/neillu23/hyp…
neillu23 May 23, 2023
2787891
sre21 8k adapted to persephone branck
jesus-villalba May 25, 2023
7ea0eb0
update lid training for focal loss and hard negative sampling
Jun 4, 2023
1974cb2
Merge branch 'persephone-refactor' of https://github.com/neillu23/hyp…
neillu23 Jun 4, 2023
aed329b
update film transducer decoder for original joiner
Jun 4, 2023
05ccbe5
Merge branch 'persephone-refactor' of https://github.com/neillu23/hyp…
neillu23 Jun 4, 2023
5ec0dc7
add different loss for lid
Jun 5, 2023
2e7ddbd
Merge branch 'persephone-refactor' of https://github.com/neillu23/hyp…
Jun 5, 2023
22920dc
add new training mode: film-ft, and add lid-film-asr system
Jun 8, 2023
e12e9f5
update configuration
Jun 8, 2023
7e1fdf8
update model
Jun 8, 2023
4dfe23c
update model for film_transducer_lid
Jun 8, 2023
9e59d74
add activation option for film
Jun 9, 2023
1f56469
add configuration
Jun 9, 2023
92a9c41
Merge branch 'persephone-refactor' of https://github.com/neillu23/hyp…
neillu23 Jun 11, 2023
35cb6f3
add config for film-asr-lid model
Jun 11, 2023
0b4bb6b
Merge branch 'persephone-refactor' of https://github.com/neillu23/hyp…
neillu23 Jun 11, 2023
250bacf
add new configs for film model
Jun 13, 2023
410100c
update config
Jun 13, 2023
2a8c3c4
update config for film ASR
Jun 18, 2023
1907869
add config for film ASR
Jun 18, 2023
1352974
Merge branch 'persephone-refactor' of https://github.com/neillu23/hyp…
neillu23 Jun 20, 2023
87822f6
add bias initialization
Jun 23, 2023
27eea76
add new config
Jun 24, 2023
9d25b5a
Merge branch 'persephone-refactor' of https://github.com/neillu23/hyp…
neillu23 Jun 25, 2023
f809855
first config for film-lid model
Jun 28, 2023
8c73fa4
add run script for film-asr-lid
Jun 28, 2023
982499e
update joint-training code
Jun 30, 2023
63a2bd9
added config 2.0 to vox v2
jesus-villalba Jun 30, 2023
5c75b6f
Merge remote-tracking branch 'upstream/persephone-refactor' into pers…
Jun 30, 2023
ac71e9a
update joint-training of LID-FILM-ASR
Jul 1, 2023
3d973ab
Merge branch 'persephone-refactor' of https://github.com/neillu23/hyp…
neillu23 Jul 1, 2023
28e61e3
update lid-film-asr training
Jul 2, 2023
b387ddd
update config
neillu23 Jul 2, 2023
d16338e
Merge branch 'persephone-refactor' of https://github.com/neillu23/hyp…
neillu23 Jul 2, 2023
acbfc06
update joint training for ASR-LID
Jul 4, 2023
f2e5aad
Merge branch 'persephone-refactor' of https://github.com/neillu23/hyp…
neillu23 Jul 5, 2023
47fae72
merge commit
Jul 5, 2023
562498f
update decode code
neillu23 Jul 5, 2023
27e96ba
Merge branch 'persephone-refactor' of https://github.com/neillu23/hyp…
Jul 5, 2023
458e65e
add rnn_original for film-rnn
Jul 7, 2023
c1d193a
finished experiments of models 2.0 in voxceleb/v2
jesus-villalba Jul 19, 2023
26eca97
add configs for commonvoice speaker verification
Aug 7, 2023
89efce4
voxceleb v1.2 works up to snorm backend
jesus-villalba Sep 4, 2023
77bbad4
Add new parameters for feat_fusion_end
Sep 7, 2023
89c6e20
finished vox v1.2 except plda
jesus-villalba Sep 8, 2023
44f085a
introduce entry points
jesus-villalba Sep 10, 2023
6105476
make it work with cuda 11
jesus-villalba Sep 11, 2023
e4a5be1
Merge pull request #137 from hyperion-ml/persephone-entry
jesus-villalba Sep 11, 2023
392cd30
started vox/v2.1 recipe and fix some readmes
jesus-villalba Sep 11, 2023
ed35173
vox/v2.1 recipe done, not tested
jesus-villalba Sep 11, 2023
8760d05
implemented lora in w2v2, not tested
jesus-villalba Sep 12, 2023
09ba2f2
Merge branch 'persephone-refactor' of https://github.com/hyperion-ml/…
Sep 13, 2023
844b6e3
Merge branch 'persephone-refactor' of https://github.com/neillu23/hyp…
Sep 13, 2023
71f629d
add lora into ASR (haven't tested)
Sep 13, 2023
a75610e
vox2.1 working and lora
jesus-villalba Sep 13, 2023
d4823db
Merge branch 'persephone-refactor' of https://github.com/hyperion-ml/…
Sep 13, 2023
c23103e
lora in wavlm and hubert
jesus-villalba Sep 13, 2023
81c540b
fix bug in w2v constructors with lora
jesus-villalba Sep 15, 2023
95ed74d
Merge branch 'persephone-refactor' of https://github.com/hyperion-ml/…
Sep 15, 2023
a54c963
update default argument of lora_merge_weights to false
Sep 23, 2023
6a72173
update config for 4 langs experiment
Sep 25, 2023
e15b227
Add FiLM inside the Wav2vec2
Sep 30, 2023
9022d8a
update FiLM Wav2vec2
Sep 30, 2023
27fffa0
add charachter based model for ASR
Oct 7, 2023
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
15 changes: 11 additions & 4 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -26,13 +26,20 @@ The full API is described in the documentation page [https://hyperion-ml.readthe
### Prerequisites

We use anaconda or miniconda, though you should be able to make it work in other python distributions
To start, you should create a new enviroment and install PyTorch>=1.9, (older versions are not supported any longer) e.g.:
To start, you should create a new enviroment and install PyTorch:
```
conda create --name ${your_env} python=3.8
conda create --name ${your_env} python=3.11
conda activate ${your_env}
conda install pytorch==1.10.1 torchvision==0.11.2 torchaudio==0.10.1 cudatoolkit=10.2 -c pytorch
conda install pytorch torchvision torchaudio pytorch-cuda=11.8 -c pytorch -c nvidia
```
In next Hyperion versions, we will upgrade to Pytorch>=1.9 and drop compatibility with older PyTorch versions.

For systems with cuda 10.2 driver:
```
conda create --name ${your_env} python=3.10
conda activate ${your_env}
conda install pytorch==1.12.1 torchvision==0.13.1 torchaudio==0.12.1 cudatoolkit=10.2 -c pytorch
```


### Installing Hyperion

Expand Down
33 changes: 33 additions & 0 deletions egs/commonvoice/v1/cmd.sh
Original file line number Diff line number Diff line change
@@ -0,0 +1,33 @@
# you can change cmd.sh depending on what type of queue you are using.
# If you have no queueing system and want to run on a local machine, you
# can change all instances 'queue.pl' to run.pl (but be careful and run
# commands one by one: most recipes will exhaust the memory on your
# machine). queue.pl works with GridEngine (qsub). slurm.pl works
# with slurm. Different queues are configured differently, with different
# queue names and different ways of specifying things like memory;
# to account for these differences you can create and edit the file
# conf/queue.conf to match your queue's configuration. Search for
# conf/queue.conf in http://kaldi-asr.org/doc/queue.html for more information,
# or search for the string 'default_config' in utils/queue.pl or utils/slurm.pl.

if [ "$(hostname -d)" == "cm.gemini" ];then
export train_cmd="queue.pl --config conf/coe_gpu_long.conf --mem 4G"
export cuda_cmd="queue.pl --config conf/coe_gpu_long.conf --mem 20G"
#export cuda_cmd="queue.pl --config conf/coe_gpu_v100.conf --mem 20G"
export cuda_cmd="queue.pl --config conf/coe_gpu_rtx.conf --mem 40G"
export cuda_eval_cmd="queue.pl --config conf/coe_gpu_short.conf --mem 4G"
#export cuda_eval_cmd="queue.pl --config conf/coe_gpu_rtx.conf --mem 10G"
#export cuda_eval_cmd="queue.pl --config conf/coe_gpu_long.conf --mem 4G"
elif [ "$(hostname -d)" == "cm.cluster" ];then
export train_cmd="slurm.pl --config conf/slurm.conf --mem 4G"
export cuda_cmd="slurm.pl --config conf/slurm.conf --mem 20G"
export cuda_eval_cmd="$train_cmd"
else
export train_cmd="run.pl"
export cuda_cmd="run.pl"
export cuda_eval_cmd="$train_cmd"
#export train_cmd="queue.pl --config conf/clsp.conf --mem 4G "
#export cuda_cmd="queue.pl --config conf/clsp.conf --mem 20G"
#export cuda_eval_cmd="$train_cmd"
fi

16 changes: 16 additions & 0 deletions egs/commonvoice/v1/conf/clsp.conf
Original file line number Diff line number Diff line change
@@ -0,0 +1,16 @@

# Default configuration
command sbatch --export=PATH
#command qsub -v PATH -cwd -S /bin/bash -j y -l arch=*64* -V
option mem=* --mem-per-cpu $0
# option mem=* -l mem_free=$0,ram_free=$0
option mem=0 # Do not add anything to qsub_opts
option num_threads=* -pe smp $0
option num_threads=1 # Do not add anything to qsub_opts
option max_jobs_run=* -tc $0
default gpu=0
option gpu=0
option gpu=* -p GPU-shared --gres=gpu:$0 -c $0 # Recommend allocating more CPU than, or equal to the number of GPU
#option gpu=0 -l 'hostname=b[1]*|c0[123456789]*|c1[1345679]*|c2[12357]*'
#option gpu=* -l 'hostname=c0[123456789]*|c1[1345679]*|c2[12357]*,gpu=$0'

11 changes: 11 additions & 0 deletions egs/commonvoice/v1/conf/coe_gpu_bigmem.conf
Original file line number Diff line number Diff line change
@@ -0,0 +1,11 @@

# Default configuration
command qsub -v PATH -cwd -S /bin/bash -j y -sync y -l arch=*64* -V
option mem=* -l mem_free=$0
option mem=0 # Do not add anything to qsub_opts
option num_threads=* -l num_proc=$0
option num_threads=1 # Do not add anything to qsub_opts
option max_jobs_run=* -tc $0
default gpu=0
option gpu=0 -q all.q -l h_rt=100:00:00 -l hostname=r[2-7]*
option gpu=* -l gpu=$0,h_rt=500:00:00 -q gpu.q -l hostname=r[237]n[01][0123456789]*
13 changes: 13 additions & 0 deletions egs/commonvoice/v1/conf/coe_gpu_long.conf
Original file line number Diff line number Diff line change
@@ -0,0 +1,13 @@

# Default configuration
command qsub -v PATH -cwd -S /bin/bash -j y -sync y -l arch=*64* -V
option mem=* -l mem_free=$0
option mem=0 # Do not add anything to qsub_opts
option num_threads=* -l num_proc=$0
option num_threads=1 # Do not add anything to qsub_opts
option max_jobs_run=* -tc $0
default gpu=0
option gpu=0 -q all.q -l h_rt=100:00:00 -l hostname=r[1-9]*
option gpu=* -l gpu=$0,h_rt=500:00:00 -q gpu.q -l hostname=r[1-9]*


11 changes: 11 additions & 0 deletions egs/commonvoice/v1/conf/coe_gpu_rtx.conf
Original file line number Diff line number Diff line change
@@ -0,0 +1,11 @@

# Default configuration
command qsub -v PATH -cwd -S /bin/bash -j y -sync y -l arch=*64* -V
option mem=* -l mem_free=$0
option mem=0 # Do not add anything to qsub_opts
option num_threads=* -l num_proc=$0
option num_threads=1 # Do not add anything to qsub_opts
option max_jobs_run=* -tc $0
default gpu=0
option gpu=0 -q all.q -l h_rt=100:00:00
option gpu=* -l gpu=$0,h_rt=500:00:00 -q gpu.q@@rtx
11 changes: 11 additions & 0 deletions egs/commonvoice/v1/conf/coe_gpu_short.conf
Original file line number Diff line number Diff line change
@@ -0,0 +1,11 @@

# Default configuration
command qsub -v PATH -cwd -S /bin/bash -j y -sync y -l arch=*64* -V
option mem=* -l mem_free=$0
option mem=0 # Do not add anything to qsub_opts
option num_threads=* -l num_proc=$0
option num_threads=1 # Do not add anything to qsub_opts
option max_jobs_run=* -tc $0
default gpu=0
option gpu=0 -q all.q -l h_rt=100:00:00 -l hostname=r[1-9]*
option gpu=* -l gpu=$0,h_rt=00:59:00 -q gpu_short.q -l hostname=r[17]*
11 changes: 11 additions & 0 deletions egs/commonvoice/v1/conf/coe_gpu_v100.conf
Original file line number Diff line number Diff line change
@@ -0,0 +1,11 @@

# Default configuration
command qsub -v PATH -cwd -S /bin/bash -j y -sync y -l arch=*64* -V
option mem=* -l mem_free=$0
option mem=0 # Do not add anything to qsub_opts
option num_threads=* -l num_proc=$0
option num_threads=1 # Do not add anything to qsub_opts
option max_jobs_run=* -tc $0
default gpu=0
option gpu=0 -q all.q -l h_rt=100:00:00
option gpu=* -l gpu=$0,h_rt=500:00:00 -q gpu.q@@v100
2 changes: 2 additions & 0 deletions egs/commonvoice/v1/conf/infer.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,2 @@
beam_width: 5
decoding_method: time_sync_beam_search
35 changes: 35 additions & 0 deletions egs/commonvoice/v1/conf/reverb_noise_aug.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,35 @@
reverb_aug:
reverb_prob: 0.45
max_reverb_context: 0.5
rir_types:
smallroom:
weight: 1
rir_path: scp:data/rirs_smallroom/rirs.scp
rir_norm: max
mediumroom:
weight: 1
rir_path: scp:data/rirs_mediumroom/rirs.scp
rir_norm: max
realroom:
weight: 1
rir_path: scp:data/rirs_real/rirs.scp
rir_norm: max
noise_aug:
noise_prob: 0.7
noise_types:
noise:
weight: 1
noise_path: data/musan_noise_proc_audio/wav.scp
min_snr: 0
max_snr: 18
music:
weight: 1
noise_path: data/musan_music_proc_audio/wav.scp
min_snr: 3
max_snr: 18
babble:
weight: 1
noise_path: data/musan_speech_babble/wav.scp
min_snr: 3
max_snr: 18

15 changes: 15 additions & 0 deletions egs/commonvoice/v1/conf/slurm.conf
Original file line number Diff line number Diff line change
@@ -0,0 +1,15 @@
# Default configuration
command sbatch --export=PATH
option name=* --job-name $0
default time=24:00:00
option time=* --time $0
option mem=* --mem-per-cpu $0
option mem=0
option num_threads=* --cpus-per-task $0
option num_threads=1 --cpus-per-task 1
option num_nodes=* --nodes $0
default gpu=0
option gpu=0
option gpu=* -p a100 -A jvillal7_gpu --ntasks-per-node 4 --gres=gpu:$0 -c $0 # Recommend allocating more CPU than, or equal to the number of GPU
# note: the --max-jobs-run option is supported as a special case
# by slurm.pl and you don't have to handle it in the config file.
Original file line number Diff line number Diff line change
@@ -0,0 +1,91 @@
data:
train:
dataset:
wav_scale: 1
aug_cfgs:
- conf/reverb_noise_aug.yaml
return_segment_info:
- text
- language
sampler:
#sampler_type: 'bucketing_seg_sampler'
sampler_type: 'class_weighted_random_bucketing_seg_sampler'
max_batch_length: 40
max_audio_length: 15.
min_batch_size: 1
drop_last: false
# for class_weighted_random_bucketing_seg_sampler
base_sampler_type: class_weighted_seg_sampler
weight_mode: "data-prior"
class_name: "language"
weight_exponent: 0.3
num_chunks_per_seg_epoch: 0.1

data_loader:
num_workers: 1
val:
dataset:
aug_cfgs:
- conf/reverb_noise_aug.yaml
wav_scale: 1
return_segment_info:
- text
- language
sampler:
#sampler_type: 'bucketing_seg_sampler'
sampler_type: 'class_weighted_random_bucketing_seg_sampler'
max_batch_length: 40
max_audio_length: 15.
min_batch_size: 1
drop_last: true
# for class_weighted_random_bucketing_seg_sampler
base_sampler_type: class_weighted_seg_sampler
weight_mode: "data-prior"
class_name: "language"
weight_exponent: 0.3
num_chunks_per_seg_epoch: 0.5
data_loader:
num_workers: 1
model:
hf_feats:
pretrained_model_path: facebook/wav2vec2-xls-r-300m
transducer:
decoder:
prune_range: 15
rnnt_loss: k2_pruned
simple_loss_scale: 0.2
predictor:
embed_dim: 1024
num_layers: 2
hid_feats: 512
embed_dropout_rate: 0.4
rnn_dropout_rate: 0.4
rnn_type: lstm
joiner:
joiner_type: film_joiner
hid_feats: 512
feat_fusion_method: film-weighted-avg
feat_fusion_start: 2
trainer:
optim:
opt_type: sgd
lr: 0.002
momentum: 0.9
weight_decay: 4e-4
lrsched:
lrsch_type: exp_lr
decay_rate: 0.5
decay_steps: 45000
hold_steps: 30000
min_lr: 4e-5
warmup_steps: 6000
update_lr_on_opt_step: true
grad_clip: 100
use_amp: true
log_interval: 1000
epochs: 120
# eff_batch_size: 1024
eff_batch_size: 128
train_mode: hf-feats-frozen-nograd


Original file line number Diff line number Diff line change
@@ -0,0 +1,91 @@
data:
train:
dataset:
wav_scale: 1
aug_cfgs:
- conf/reverb_noise_aug.yaml
return_segment_info:
- text
- language
sampler:
#sampler_type: 'bucketing_seg_sampler'
sampler_type: 'class_weighted_random_bucketing_seg_sampler'
max_batch_length: 40
max_audio_length: 15.
min_batch_size: 1
drop_last: false
# for class_weighted_random_bucketing_seg_sampler
base_sampler_type: class_weighted_seg_sampler
weight_mode: "data-prior"
class_name: "language"
weight_exponent: 0.3
num_chunks_per_seg_epoch: 0.1

data_loader:
num_workers: 1
val:
dataset:
aug_cfgs:
- conf/reverb_noise_aug.yaml
wav_scale: 1
return_segment_info:
- text
- language
sampler:
#sampler_type: 'bucketing_seg_sampler'
sampler_type: 'class_weighted_random_bucketing_seg_sampler'
max_batch_length: 40
max_audio_length: 15.
min_batch_size: 1
drop_last: true
# for class_weighted_random_bucketing_seg_sampler
base_sampler_type: class_weighted_seg_sampler
weight_mode: "data-prior"
class_name: "language"
weight_exponent: 0.3
num_chunks_per_seg_epoch: 1.0
data_loader:
num_workers: 1
model:
hf_feats:
pretrained_model_path: facebook/wav2vec2-xls-r-300m
transducer:
decoder:
prune_range: 15
rnnt_loss: k2_pruned
simple_loss_scale: 0.2
condition_size: 256
predictor:
embed_dim: 1024
num_layers: 2
hid_feats: 512
embed_dropout_rate: 0.4
rnn_dropout_rate: 0.4
rnn_type: lstm_residual
joiner:
hid_feats: 512
feat_fusion_method: film-weighted-avg
feat_fusion_start: 2
trainer:
optim:
opt_type: sgd
lr: 0.002
momentum: 0.9
weight_decay: 4e-4
lrsched:
lrsch_type: exp_lr
decay_rate: 0.5
decay_steps: 45000
hold_steps: 30000
min_lr: 4e-5
warmup_steps: 6000
update_lr_on_opt_step: true
grad_clip: 100
use_amp: true
log_interval: 1000
epochs: 120
# eff_batch_size: 1024
eff_batch_size: 128
train_mode: hf-feats-frozen-nograd


Loading