You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
If you are using your own data, it must be formatted as one directory with 6 files: train.source, train.target, val.source, val.target, test.source, test.target.
38
39
The `.source` files are the input, the `.target` files are the desired output.
39
40
40
-
### Evaluation Commands
41
-
42
-
To create summaries for each article in dataset, we use `run_eval.py`, here are a few commands that run eval for different tasks and models.
43
-
If 'translation' is in your task name, the computed metric will be BLEU. Otherwise, ROUGE will be used.
44
-
45
-
For t5, you need to specify --task translation_{src}_to_{tgt} as follows:
46
-
```bash
47
-
export DATA_DIR=wmt_en_ro
48
-
python run_eval.py t5_base \
49
-
$DATA_DIR/val.source mbart_val_generations.txt \
50
-
--reference_path $DATA_DIR/val.target \
51
-
--score_path enro_bleu.json \
52
-
--task translation_en_to_ro \
53
-
--n_obs 100 \
54
-
--device cuda \
55
-
--fp16 \
56
-
--bs 32
57
-
```
58
-
59
-
This command works for MBART, although the BLEU score is suspiciously low.
0 commit comments