#include "BleuScorer.h" #include #include #include #include #include #include #include #include "util/check.hh" #include "Ngram.h" #include "Reference.h" #include "Util.h" #include "ScoreDataIterator.h" #include "FeatureDataIterator.h" #include "Vocabulary.h" using namespace std; namespace { // configure regularisation const char KEY_REFLEN[] = "reflen"; const char REFLEN_AVERAGE[] = "average"; const char REFLEN_SHORTEST[] = "shortest"; const char REFLEN_CLOSEST[] = "closest"; } // namespace namespace MosesTuning { BleuScorer::BleuScorer(const string& config) : StatisticsBasedScorer("BLEU", config), m_ref_length_type(CLOSEST) { const string reflen = getConfig(KEY_REFLEN, REFLEN_CLOSEST); if (reflen == REFLEN_AVERAGE) { m_ref_length_type = AVERAGE; } else if (reflen == REFLEN_SHORTEST) { m_ref_length_type = SHORTEST; } else if (reflen == REFLEN_CLOSEST) { m_ref_length_type = CLOSEST; } else { throw runtime_error("Unknown reference length strategy: " + reflen); } } BleuScorer::~BleuScorer() {} size_t BleuScorer::CountNgrams(const string& line, NgramCounts& counts, unsigned int n, bool is_testing) { assert(n > 0); vector encoded_tokens; // When performing tokenization of a hypothesis translation, we don't have // to update the Scorer's word vocabulary. However, the tokenization of // reference translations requires modifying the vocabulary, which means // this procedure might be slower than the tokenization the hypothesis // translation. if (is_testing) { TokenizeAndEncodeTesting(line, encoded_tokens); } else { TokenizeAndEncode(line, encoded_tokens); } const size_t len = encoded_tokens.size(); vector ngram; for (size_t k = 1; k <= n; ++k) { //ngram order longer than sentence - no point if (k > len) { continue; } for (size_t i = 0; i < len - k + 1; ++i) { ngram.clear(); ngram.reserve(len); for (size_t j = i; j < i+k && j < len; ++j) { ngram.push_back(encoded_tokens[j]); } counts.Add(ngram); } } return len; } void BleuScorer::setReferenceFiles(const vector& referenceFiles) { // Make sure reference data is clear m_references.reset(); mert::VocabularyFactory::GetVocabulary()->clear(); //load reference data for (size_t i = 0; i < referenceFiles.size(); ++i) { TRACE_ERR("Loading reference from " << referenceFiles[i] << endl); if (!OpenReference(referenceFiles[i].c_str(), i)) { throw runtime_error("Unable to open " + referenceFiles[i]); } } } bool BleuScorer::OpenReference(const char* filename, size_t file_id) { ifstream ifs(filename); if (!ifs) { cerr << "Cannot open " << filename << endl; return false; } return OpenReferenceStream(&ifs, file_id); } bool BleuScorer::OpenReferenceStream(istream* is, size_t file_id) { if (is == NULL) return false; string line; size_t sid = 0; while (getline(*is, line)) { line = preprocessSentence(line); if (file_id == 0) { Reference* ref = new Reference; m_references.push_back(ref); // Take ownership of the Reference object. } if (m_references.size() <= sid) { cerr << "Reference " << file_id << "has too many sentences." << endl; return false; } NgramCounts counts; size_t length = CountNgrams(line, counts, kBleuNgramOrder); //for any counts larger than those already there, merge them in for (NgramCounts::const_iterator ci = counts.begin(); ci != counts.end(); ++ci) { const NgramCounts::Key& ngram = ci->first; const NgramCounts::Value newcount = ci->second; NgramCounts::Value oldcount = 0; m_references[sid]->get_counts()->Lookup(ngram, &oldcount); if (newcount > oldcount) { m_references[sid]->get_counts()->operator[](ngram) = newcount; } } //add in the length m_references[sid]->push_back(length); if (sid > 0 && sid % 100 == 0) { TRACE_ERR("."); } ++sid; } return true; } void BleuScorer::prepareStats(size_t sid, const string& text, ScoreStats& entry) { if (sid >= m_references.size()) { stringstream msg; msg << "Sentence id (" << sid << ") not found in reference set"; throw runtime_error(msg.str()); } NgramCounts testcounts; // stats for this line vector stats(kBleuNgramOrder * 2); string sentence = preprocessSentence(text); const size_t length = CountNgrams(sentence, testcounts, kBleuNgramOrder, true); const int reference_len = CalcReferenceLength(sid, length); stats.push_back(reference_len); //precision on each ngram type for (NgramCounts::const_iterator testcounts_it = testcounts.begin(); testcounts_it != testcounts.end(); ++testcounts_it) { const NgramCounts::Value guess = testcounts_it->second; const size_t len = testcounts_it->first.size(); NgramCounts::Value correct = 0; NgramCounts::Value v = 0; if (m_references[sid]->get_counts()->Lookup(testcounts_it->first, &v)) { correct = min(v, guess); } stats[len * 2 - 2] += correct; stats[len * 2 - 1] += guess; } entry.set(stats); } statscore_t BleuScorer::calculateScore(const vector& comps) const { CHECK(comps.size() == kBleuNgramOrder * 2 + 1); float logbleu = 0.0; for (int i = 0; i < kBleuNgramOrder; ++i) { if (comps[2*i] == 0) { return 0.0; } logbleu += log(comps[2*i]) - log(comps[2*i+1]); } logbleu /= kBleuNgramOrder; // reflength divided by test length const float brevity = 1.0 - static_cast(comps[kBleuNgramOrder * 2]) / comps[1]; if (brevity < 0.0) { logbleu += brevity; } return exp(logbleu); } int BleuScorer::CalcReferenceLength(size_t sentence_id, size_t length) { switch (m_ref_length_type) { case AVERAGE: return m_references[sentence_id]->CalcAverage(); break; case CLOSEST: return m_references[sentence_id]->CalcClosest(length); break; case SHORTEST: return m_references[sentence_id]->CalcShortest(); break; default: cerr << "unknown reference types." << endl; exit(1); } } void BleuScorer::DumpCounts(ostream* os, const NgramCounts& counts) const { for (NgramCounts::const_iterator it = counts.begin(); it != counts.end(); ++it) { *os << "("; const NgramCounts::Key& keys = it->first; for (size_t i = 0; i < keys.size(); ++i) { if (i != 0) { *os << " "; } *os << keys[i]; } *os << ") : " << it->second << ", "; } *os << endl; } float smoothedSentenceBleu (const std::vector& stats, float smoothing, bool smoothBP) { CHECK(stats.size() == kBleuNgramOrder * 2 + 1); float logbleu = 0.0; for (int j = 0; j < kBleuNgramOrder; j++) { logbleu += log(stats[2 * j] + smoothing) - log(stats[2 * j + 1] + smoothing); } logbleu /= kBleuNgramOrder; const float reflength = stats[(kBleuNgramOrder * 2)] + (smoothBP ? smoothing : 0.0f); const float brevity = 1.0 - reflength / stats[1]; if (brevity < 0.0) { logbleu += brevity; } return exp(logbleu); } float sentenceLevelBackgroundBleu(const std::vector& sent, const std::vector& bg) { // Sum sent and background std::vector stats; CHECK(sent.size()==bg.size()); CHECK(sent.size()==kBleuNgramOrder*2+1); for(size_t i=0; i& stats) { CHECK(stats.size() == kBleuNgramOrder * 2 + 1); float logbleu = 0.0; for (int j = 0; j < kBleuNgramOrder; j++) { logbleu += log(stats[2 * j]) - log(stats[2 * j + 1]); } logbleu /= kBleuNgramOrder; const float brevity = 1.0 - stats[(kBleuNgramOrder * 2)] / stats[1]; if (brevity < 0.0) { logbleu += brevity; } return exp(logbleu); } vector BleuScorer::ScoreNbestList(const string& scoreFile, const string& featureFile) { vector scoreFiles; vector featureFiles; scoreFiles.push_back(scoreFile); featureFiles.push_back(featureFile); vector featureDataIters; vector scoreDataIters; for (size_t i = 0; i < featureFiles.size(); ++i) { featureDataIters.push_back(FeatureDataIterator(featureFiles[i])); scoreDataIters.push_back(ScoreDataIterator(scoreFiles[i])); } vector > hypotheses; if (featureDataIters[0] == FeatureDataIterator::end()) { cerr << "Error: at the end of feature data iterator" << endl; exit(1); } for (size_t i = 0; i < featureFiles.size(); ++i) { if (featureDataIters[i] == FeatureDataIterator::end()) { cerr << "Error: Feature file " << i << " ended prematurely" << endl; exit(1); } if (scoreDataIters[i] == ScoreDataIterator::end()) { cerr << "Error: Score file " << i << " ended prematurely" << endl; exit(1); } if (featureDataIters[i]->size() != scoreDataIters[i]->size()) { cerr << "Error: features and scores have different size" << endl; exit(1); } for (size_t j = 0; j < featureDataIters[i]->size(); ++j) { hypotheses.push_back(pair(i,j)); } } // score the nbest list vector bleuScores; for (size_t i=0; i < hypotheses.size(); ++i) { pair translation = hypotheses[i]; float bleu = smoothedSentenceBleu(scoreDataIters[translation.first]->operator[](translation.second)); bleuScores.push_back(bleu); } return bleuScores; } }