pip3 install -r requirements.txt
# pip3 install -r requirements_2.txt # Oscar's local env settings
HuBERT Unit:long-t5-base-SQA-hubert-100
mHuBERT Unit:long-t5-base-SQA-mhubert-1000
Datasets: NMSQA
T5-series Model:long-T5
Training Script:
python3 main.py
Datasets
Unit Datasets: GSQA/speech-alpaca-gpt4-unit Speech Datasets GSQA/spoken-alpaca-gpt4
T5-series Model:long-T5 alpaca-TQA-init T5-series Model: LongT5-alpaca-TQA
login GSQA authorized huggingface account
$ huggingface-cli login
login wandb account to record training figures
$ wandb login --relogin
# select one of the aux_task in choices to fill after --aux_task
$ python3 main_multiTask.py --aux_task qt,at,qu
(choices=['qt,qu', 'qt,at,qu', "qu,at", "at"])
3. after finish training, push model to https://huggingface.co/GSQA
ASR Model:Whisper --> TBD
Evaluating Script:
# stpe1: run
python3 whisper_evaluate.py --model /path/to/the/huggingface/model --auto_split_dataset
# (for more optional arguments check whisper_evaluate.py)
# step 2: for alpaca dataset BertScore, run
python3 BertScore_eval.py
# (remember to change the evaluation file path first)
# step 2: for dataset with context, run
python3 eval_score.py # Remember to check the name of output files.
# Note: Please put the best reported score to Overleaf Table.