By default, translation is done using beam search. The
-beam_size option can be used to trade-off translation time and search accuracy, with
-beam_size 1 giving greedy search. The small default beam size is often enough in practice.
Beam search can also be used to provide an approximate n-best list of translations by setting
-n_best greater than 1. For analysis, the translation command also takes an oracle/gold
-tgt file and will output a comparison of scores.
The beam search provides a built-in filter based on unknown words:
-max_num_unks. Hypotheses with more unknown words than this value are dropped.
As dropped hypotheses temporarily reduce the beam size, the
-pre_filter_factor is a way to increase the number of considered hypotheses before applying filters.
The beam search also supports various normalization techniques that are disabled by default and can be used to biased the scores generated by the model:
where is the source, is the current target, and the functions as defined below. An additional penalty on end of sentence tokens can also be added to prioritize longer sentences.
Scores are normalized by the following formula as defined in Wu et al. (2016):
where is the current target length and is the length normalization coefficient
Scores are penalized by the following formula as defined in Wu et al. (2016):
where is the attention probability of the -th target word on the -th source word , is the source length, is the current target length and is the coverage normalization coefficient
End of sentence normalization¶
The score of the end of sentence token is penalized by the following formula:
where is the source length, is the current target length and is the end of sentence normalization coefficient
Decoding with auxiliary language model¶
Beam search can use an additional language model to modify score of each option as defined in Gulcehre et al. (2015) as "Shallow Fusion":
Where is the language model log-probability of the sequence and is defined by
To activate the language model, simply use
The language model cannot use bidirectional RNN and needs to share the same vocabulary (tokens and features) than the translation model.
Decoding with lexical constraints¶
Beam search can be constrained to produce translations from a phrase table in its output, whenever corresponding source tokens are found on the input side. The implementation is based on the Grid Beam Search approach described in Hokamp and Liu (2017). For the moment, only simple one-token constraints are accepted.
To activate lexical constraints, use
-lexical_constraints option together with
-phrase_table to specify the list of constraints.
-limit_lexical_constraints option may be used to produce each constraint translation token only as many times as the correspoding source token is found in the input.
Output attention to a file¶
-save_attention FILE can be used to save attention state to a file during translation. The format of the file is as following (compatible with NEMATUS):
sentence id ||| target words ||| score ||| source words ||| number of source words ||| number of target words ALIGNMENT FOR T_1 ALIGNMENT FOR T_2 ... ALIGNMENT FOR T_n
T_n are the target words - each alignement line is space separated probability to source word.
Visualizing the beam search¶
To visualize the beam search exploration, you can use the option
-save_beam_to beam.json. It will save a JSON serialization of the beam search history.
This option requires the
This representation can then be visualized dynamically using the
generate_beam_viz.py script from the
git clone https://github.com/OpenNMT/VisTools.git cd VisTools mkdir out/ python generate_beam_viz.py -d ~/OpenNMT/beam.json -o out/ firefox out/000000.html