translation_server.lua options:

  • -h
    This help.
  • -md
    Dump help in Markdown format.
  • -config <string>
    Load options from this file.
  • -save_config <string>
    Save options to this file.

Server options

  • -host <string> (default:
    Host to run the server on.
  • -port <string> (default: 5556)
    Port to run the server on.

Translator options

  • -model <string>
    Path to the serialized model file.
  • -beam_size <number> (default: 5)
    Beam size.
  • -batch_size <number> (default: 30)
    Batch size.
  • -max_sent_length <number> (default: 250)
    Maximum output sentence length.
  • -replace_unk
    Replace the generated tokens with the source token that has the highest attention weight. If -phrase_table is provided, it will lookup the identified source token and give the corresponding target token. If it is not provided (or the identified source token does not exist in the table) then it will copy the source token
  • -phrase_table <string>
    Path to source-target dictionary to replace <unk> tokens.
  • -n_best <number> (default: 1)
    If > 1, it will also output an n-best list of decoded sentences.
  • -max_num_unks <number> (default: inf)
    All sequences with more <unk>s than this will be ignored during beam search.
  • -pre_filter_factor <number> (default: 1)
    Optional, set this only if filter is being used. Before applying filters, hypotheses with top beam_size * pre_filter_factor scores will be considered. If the returned hypotheses voilate filters, then set this to a larger value to consider more.
  • -length_norm <number> (default: 0)
    Length normalization coefficient (alpha). If set to 0, no length normalization.
  • -coverage_norm <number> (default: 0)
    Coverage normalization coefficient (beta). An extra coverage term multiplied by beta is added to hypotheses scores. If is set to 0, no coverage normalization.
  • -eos_norm <number> (default: 0)
    End of sentence normalization coefficient (gamma). If set to 0, no EOS normalization.
  • -dump_input_encoding
    Instead of generating target tokens conditional on the source tokens, we print the representation (encoding/embedding) of the input.

Cuda options

  • -gpuid <string> (default: 0)
    List of comma-separated GPU identifiers (1-indexed). CPU is used when set to 0.
  • -fallback_to_cpu
    If GPU can't be used, rollback on the CPU.
  • -fp16
    Use half-precision float on GPU.
  • -no_nccl
    Disable usage of nccl in parallel mode.

Logger options

  • -log_file <string>
    Output logs to a file under this path instead of stdout.
  • -disable_logs
    If set, output nothing.
  • -log_level <string> (accepted: DEBUG, INFO, WARNING, ERROR; default: INFO)
    Output logs at this level and above.