mosesdecoder/mert/evaluator.cpp
Jeroen Vermeulen 38d790cac0 Add cross-platform randomizer module.
The code uses two mechanisms for generating random numbers: srand()/rand(),
which is not thread-safe, and srandom()/random(), which is POSIX-specific.

Here I add a util/random.cc module that centralizes these calls, and unifies
some common usage patterns.  If the implementation is not good enough, we can
now change it in a single place.

To keep things simple, this uses the portable srand()/rand() but protects them
with a lock to avoid concurrency problems.

The hard part was to keep the regression tests passing: they rely on fixed
sequences of random numbers, so a small code change could break them very
thoroughly.  Util::rand(), for wide types like size_t, calls std::rand() not
once but twice.  This behaviour was generalized into utils::wide_rand() and
friends.
2015-04-23 23:46:04 +07:00

342 lines
10 KiB
C++

#include <cstdlib>
#include <fstream>
#include <iostream>
#include <string>
#include <vector>
#include <algorithm>
#include <getopt.h>
#include <cmath>
#if defined __MINGW32__
#include <ctime>
#endif // defined
#include "Scorer.h"
#include "ScorerFactory.h"
#include "Timer.h"
#include "Util.h"
#include "Data.h"
#include "util/random.hh"
using namespace std;
using namespace MosesTuning;
namespace
{
Scorer* g_scorer = NULL;
bool g_has_more_files = false;
bool g_has_more_scorers = false;
const float g_alpha = 0.05;
class EvaluatorUtil
{
public:
static void evaluate(const string& candFile, int bootstrap, bool nbest_mode);
static float average(const vector<float>& list);
static string int2string(int n);
static vector<ScoreStats> loadNBest(const string& nBestFile);
static vector<ScoreStats> loadCand(const string& candFile);
private:
EvaluatorUtil() {}
~EvaluatorUtil() {}
};
// load hypothesis from candidate output
vector<ScoreStats> EvaluatorUtil::loadCand(const string& candFile)
{
ifstream cand(candFile.c_str());
if (!cand.good()) throw runtime_error("Error opening candidate file");
vector<ScoreStats> entries;
// Loading sentences and preparing statistics
ScoreStats scoreentry;
string line;
while (getline(cand, line)) {
g_scorer->prepareStats(entries.size(), line, scoreentry);
entries.push_back(scoreentry);
}
return entries;
}
// load 1-best hypothesis from n-best file (useful if relying on alignment/tree information)
vector<ScoreStats> EvaluatorUtil::loadNBest(const string& nBestFile)
{
vector<ScoreStats> entries;
Data data(g_scorer);
data.loadNBest(nBestFile, true);
const ScoreDataHandle & score_data = data.getScoreData();
for (size_t i = 0; i != score_data->size(); i++) {
entries.push_back(score_data->get(i, 0));
}
return entries;
}
void EvaluatorUtil::evaluate(const string& candFile, int bootstrap, bool nbest_input)
{
vector<ScoreStats> entries;
if (nbest_input) {
entries = loadNBest(candFile);
} else {
entries = loadCand(candFile);
}
int n = entries.size();
if (bootstrap) {
vector<float> scores;
for (int i = 0; i < bootstrap; ++i) {
ScoreData scoredata(g_scorer);
for (int j = 0; j < n; ++j) {
const int randomIndex = util::rand_excl(n);
scoredata.add(entries[randomIndex], j);
}
g_scorer->setScoreData(&scoredata);
candidates_t candidates(n, 0);
float score = g_scorer->score(candidates);
scores.push_back(score);
}
float avg = average(scores);
sort(scores.begin(), scores.end());
int lbIdx = scores.size() * (g_alpha / 2);
int rbIdx = scores.size() * (1 - g_alpha / 2);
float lb = scores[lbIdx];
float rb = scores[rbIdx];
if (g_has_more_files) cout << candFile << "\t";
if (g_has_more_scorers) cout << g_scorer->getName() << "\t";
cout.setf(ios::fixed, ios::floatfield);
cout.precision(4);
cout << avg << "\t[" << lb << "," << rb << "]" << endl;
} else {
ScoreData scoredata(g_scorer);
for (int sid = 0; sid < n; ++sid) {
scoredata.add(entries[sid], sid);
}
g_scorer->setScoreData(&scoredata);
candidates_t candidates(n, 0);
float score = g_scorer->score(candidates);
if (g_has_more_files) cout << candFile << "\t";
if (g_has_more_scorers) cout << g_scorer->getName() << "\t";
cout.setf(ios::fixed, ios::floatfield);
cout.precision(4);
cout << score << endl;
}
}
string EvaluatorUtil::int2string(int n)
{
stringstream ss;
ss << n;
return ss.str();
}
float EvaluatorUtil::average(const vector<float>& list)
{
float sum = 0;
for (vector<float>::const_iterator it = list.begin(); it != list.end(); ++it)
sum += *it;
return sum / list.size();
}
void usage()
{
cerr << "usage: evaluator [options] --reference ref1[,ref2[,ref3...]] --candidate cand1[,cand2[,cand3...]] " << endl;
cerr << "[--sctype|-s] the scorer type (default BLEU)" << endl;
cerr << "[--scconfig|-c] configuration string passed to scorer" << endl;
cerr << "\tThis is of the form NAME1:VAL1,NAME2:VAL2 etc " << endl;
cerr << "[--reference|-R] comma separated list of reference files" << endl;
cerr << "[--candidate|-C] comma separated list of candidate files" << endl;
cerr << "[--nbest|-n] comma separated list of nbest files (only 1-best is evaluated)" << endl;
cerr << "[--factors|-f] list of factors passed to the scorer (e.g. 0|2)" << endl;
cerr << "[--filter|-l] filter command which will be used to preprocess the sentences" << endl;
cerr << "[--bootstrap|-b] number of booststraped samples (default 0 - no bootstraping)" << endl;
cerr << "[--rseed|-r] the random seed for bootstraping (defaults to system clock)" << endl;
cerr << "[--help|-h] print this message and exit" << endl;
cerr << endl;
cerr << "Evaluator is able to compute more metrics at once. To do this," << endl;
cerr << "specify more --sctype arguments. You can also specify more --scconfig strings." << endl;
cerr << endl;
cerr << "The example below prints BLEU score, PER score and interpolated" << endl;
cerr << "score of CDER and PER with the given weights." << endl;
cerr << endl;
cerr << "./evaluator \\" << endl;
cerr << "\t--sctype BLEU --scconfig reflen:closest \\" << endl;
cerr << "\t--sctype PER \\" << endl;
cerr << "\t--sctype CDER,PER --scconfig weights:0.25+0.75 \\" << endl;
cerr << "\t--candidate CANDIDATE \\" << endl;
cerr << "\t--reference REFERENCE" << endl;
cerr << endl;
cerr << "If you specify only one scorer and one candidate file, only the final score" << endl;
cerr << "will be printed to stdout. Otherwise each line will contain metric name" << endl;
cerr << "and/or filename and the final score. Since most of the metrics prints some" << endl;
cerr << "debuging info, consider redirecting stderr to /dev/null." << endl;
exit(1);
}
static struct option long_options[] = {
{"sctype", required_argument, 0, 's'},
{"scconfig", required_argument, 0, 'c'},
{"reference", required_argument, 0, 'R'},
{"candidate", required_argument, 0, 'C'},
{"nbest", required_argument, 0, 'n'},
{"bootstrap", required_argument, 0, 'b'},
{"rseed", required_argument, 0, 'r'},
{"factors", required_argument, 0, 'f'},
{"filter", required_argument, 0, 'l'},
{"help", no_argument, 0, 'h'},
{0, 0, 0, 0}
};
// Options used in evaluator.
struct ProgramOption {
vector<string> scorer_types;
vector<string> scorer_configs;
string reference;
string candidate;
string nbest;
vector<string> scorer_factors;
vector<string> scorer_filter;
int bootstrap;
int seed;
bool has_seed;
ProgramOption()
: reference(""),
candidate(""),
nbest(""),
bootstrap(0),
seed(0),
has_seed(false) { }
};
void ParseCommandOptions(int argc, char** argv, ProgramOption* opt)
{
int c;
int option_index;
int last_scorer_index = -1;
while ((c = getopt_long(argc, argv, "s:c:R:C:n:b:r:f:l:h", long_options, &option_index)) != -1) {
switch(c) {
case 's':
opt->scorer_types.push_back(string(optarg));
opt->scorer_configs.push_back(string(""));
opt->scorer_factors.push_back(string(""));
opt->scorer_filter.push_back(string(""));
last_scorer_index++;
break;
case 'c':
if (last_scorer_index == -1) throw runtime_error("You need to specify a scorer before its config string.");
opt->scorer_configs[last_scorer_index] = string(optarg);
break;
case 'R':
opt->reference = string(optarg);
break;
case 'C':
opt->candidate = string(optarg);
break;
case 'n':
opt->nbest = string(optarg);
break;
case 'b':
opt->bootstrap = atoi(optarg);
break;
case 'r':
opt->seed = strtol(optarg, NULL, 10);
opt->has_seed = true;
break;
case 'f':
if (last_scorer_index == -1) throw runtime_error("You need to specify a scorer before its list of factors.");
opt->scorer_factors[last_scorer_index] = string(optarg);
break;
case 'l':
if (last_scorer_index == -1) throw runtime_error("You need to specify a scorer before its filter.");
opt->scorer_filter[last_scorer_index] = string(optarg);
break;
default:
usage();
}
}
// Add default scorer if no scorer provided
if (opt->scorer_types.size() == 0) {
opt->scorer_types.push_back(string("BLEU"));
opt->scorer_configs.push_back(string(""));
opt->scorer_factors.push_back(string(""));
opt->scorer_filter.push_back(string(""));
}
}
void InitSeed(const ProgramOption *opt)
{
if (opt->has_seed) {
cerr << "Seeding random numbers with " << opt->seed << endl;
util::rand_init(opt->seed);
} else {
cerr << "Seeding random numbers with system clock " << endl;
util::rand_init();
}
}
} // anonymous namespace
int main(int argc, char** argv)
{
ResetUserTime();
ProgramOption option;
ParseCommandOptions(argc, argv, &option);
if (option.bootstrap) {
InitSeed(&option);
}
try {
vector<string> refFiles;
vector<string> candFiles;
if (option.reference.length() == 0) throw runtime_error("You have to specify at least one reference file.");
split(option.reference, ',', refFiles);
if (option.candidate.length() == 0 && option.nbest.length() == 0) throw runtime_error("You have to specify at least one candidate (or n-best) file.");
if (option.candidate.length() > 0 && option.nbest.length() > 0) throw runtime_error("You can either specify candidate files or n-best files, but not both.");
bool nbest_input = option.nbest.length() > 0;
if (nbest_input)
split(option.nbest, ',', candFiles);
else
split(option.candidate, ',', candFiles);
if (candFiles.size() > 1) g_has_more_files = true;
if (option.scorer_types.size() > 1) g_has_more_scorers = true;
for (vector<string>::const_iterator fileIt = candFiles.begin(); fileIt != candFiles.end(); ++fileIt) {
for (size_t i = 0; i < option.scorer_types.size(); i++) {
g_scorer = ScorerFactory::getScorer(option.scorer_types[i], option.scorer_configs[i]);
g_scorer->setFactors(option.scorer_factors[i]);
g_scorer->setFilter(option.scorer_filter[i]);
g_scorer->setReferenceFiles(refFiles);
EvaluatorUtil::evaluate(*fileIt, option.bootstrap, nbest_input);
delete g_scorer;
}
}
return EXIT_SUCCESS;
} catch (const exception& e) {
cerr << "Exception: " << e.what() << endl;
return EXIT_FAILURE;
}
}