usage: translate.py [-h] [-md] -model MODEL -src SRC [-tgt TGT]
                    [-output OUTPUT] [-beam_size BEAM_SIZE]
                    [-batch_size BATCH_SIZE]
                    [-max_sent_length MAX_SENT_LENGTH] [-replace_unk]
                    [-verbose] [-n_best N_BEST] [-gpu GPU]


optional arguments:

-h, --help

show this help message and exit


print Markdown-formatted help text and exit.

-model MODEL

Path to model .pt file

-src SRC

Source sequence to decode (one line per sequence)

-tgt TGT

True target sequence (optional)

-output OUTPUT

Path to output the predictions (each line will be the decoded sequence

-beam_size BEAM_SIZE

Beam size

-batch_size BATCH_SIZE

Batch size

-max_sent_length MAX_SENT_LENGTH

Maximum sentence length.


Replace the generated UNK tokens with the source token that had the highest
attention weight. If phrase_table is provided, it will lookup the identified
source token and give the corresponding target token. If it is not provided (or
the identified source token does not exist in the table) then it will copy the
source token


Print scores and predictions for each sentence

-n_best N_BEST

If verbose is set, will output the n_best decoded sentences

-gpu GPU

Device to run on