From 1113f83f980fe4a3e0cebe7d876b1b2cddd464c2 Mon Sep 17 00:00:00 2001 From: Rico Sennrich Date: Wed, 18 Sep 2013 15:13:56 +0200 Subject: [PATCH] documentation/permissions --- contrib/server/client_multimodel.py | 0 scripts/training/create_count_tables.py | 4 ++-- scripts/training/flexibility_score.py | 2 +- 3 files changed, 3 insertions(+), 3 deletions(-) mode change 100644 => 100755 contrib/server/client_multimodel.py diff --git a/contrib/server/client_multimodel.py b/contrib/server/client_multimodel.py old mode 100644 new mode 100755 diff --git a/scripts/training/create_count_tables.py b/scripts/training/create_count_tables.py index 29d18f2dd..064576969 100755 --- a/scripts/training/create_count_tables.py +++ b/scripts/training/create_count_tables.py @@ -7,9 +7,9 @@ # The benefit is that models can be combined quickly, with the same results as if we trained a model on the concatenation of all data (excepting differences in word alignment). # Also, each model can be given a weight, which is applied to all frequencies of the model for the combination. -# Note: the input phrase table must have alignment information (--phrase-word-alignment in train-model.perl); +# Note: the input phrase table must have alignment information; # it must be unsmoothed; -# additionally, the online model type requires the lexical counts files lex.counts.e2f and lex.counts.f2e to be in the same folder (--write-lexical-counts in train-model.perl) +# additionally, the phrase table type PhraseDictionaryMultiModelCounts requires the lexical counts files lex.counts.e2f and lex.counts.f2e (obtained by using the option --write-lexical-counts in train-model.perl) # The results may differ from training on the concatenation of all data due to differences in word alignment, and rounding errors. diff --git a/scripts/training/flexibility_score.py b/scripts/training/flexibility_score.py index 66f104605..2b9e3b694 100755 --- a/scripts/training/flexibility_score.py +++ b/scripts/training/flexibility_score.py @@ -1,7 +1,7 @@ #!/usr/bin/env python # -*- coding: utf-8 -*- # add flexibility scores to a phrase table half -# you usually don't have to call this script directly, but you can run train_model.perl with the option "--flexibility-score" (will only affect steps 5 and 6) +# you usually don't have to call this script directly; to add flexibility scores to your model, run train-model.perl with the option "--flexibility-score" (will only affect steps 5 and 6) # usage: python flexibility_score.py extract.context(.inv).sorted [--Inverse] [--Hierarchical] < phrasetable > output_file # author: Rico Sennrich