Skip to content
代码片段 群组 项目
提交 af6cd310 编辑于 作者: Paden Tomasello's avatar Paden Tomasello
浏览文件

Cleanup paths

上级 5ff3ea1e
无相关合并请求
......@@ -36,13 +36,16 @@ First, create the audio file manifests and label files:
python examples/audio_nlp/nlu/generate_manifests.py --stop_root $STOP_DOWNLOAD_DIR/stop --output $FAIRSEQ_DATASET_OUTPUT/
```
Modify `examples/audio_nlp/nlu/create_dict_stop.sh` so that `fairseq_root` points to your installation of fairseq and `data_root` points to `$FAIRSEQ_DATASET_OUTPUT`
Run `./examples/audio_nlp/nlu/create_dict_stop.sh` to generate the fairseq dictionaries.
Run `./examples/audio_nlp/nlu/create_dict_stop.sh $FAIRSEQ_DATASET_OUTPUT` to generate the fairseq dictionaries.
## Training an End-to-end NLU Model
Download a wav2vec or hubert model from [link](https://github.com/facebookresearch/fairseq/tree/main/examples/hubert) or [link](https://github.com/facebookresearch/fairseq/tree/main/examples/wav2vec)
```
python fairseq_cli/hydra-train --config-dir examples/audio_nlp/nlu/configs/ --config-name nlu_finetuning task.data=$FAIRSEQ_DATA_OUTPUT
python fairseq_cli/hydra-train --config-dir examples/audio_nlp/nlu/configs/ --config-name nlu_finetuning task.data=$FAIRSEQ_DATA_OUTPUT model.w2v_path=$PRETRAINED_MODEL_PATH
```
......@@ -47,7 +47,7 @@ lr_scheduler:
model:
_name: wav2vec_seq2seq
w2v_path: /private/home/padentomasello/models/wav2vec2/wav2vec_small.pt
w2v_path: ???
autoregressive: true
apply_mask: true
mask_prob: 0.5
......
......@@ -4,9 +4,9 @@
### for model training within fairseq
fairseq_root="/private/home/padentomasello/code/fairseq-py"
fairseq_root="."
data_root="/private/home/padentomasello/data/stop/fairseq_test/stop/"
data_root=$1
train_prefix="${data_root}/train"
valid_prefix="${data_root}/eval"
test_prefix="${data_root}/test"
......
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册