mosesdecoder/moses/Hypothesis.h

311 lines
10 KiB
C
Raw Normal View History

// $Id$
// vim:tabstop=2
/***********************************************************************
Moses - factored phrase-based language decoder
Copyright (C) 2006 University of Edinburgh
This library is free software; you can redistribute it and/or
modify it under the terms of the GNU Lesser General Public
License as published by the Free Software Foundation; either
version 2.1 of the License, or (at your option) any later version.
This library is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
Lesser General Public License for more details.
You should have received a copy of the GNU Lesser General Public
License along with this library; if not, write to the Free Software
Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
***********************************************************************/
#ifndef moses_Hypothesis_h
#define moses_Hypothesis_h
#include <iostream>
#include <memory>
#include <vector>
#include "Phrase.h"
#include "TypeDef.h"
#include "WordsBitmap.h"
#include "Sentence.h"
#include "Phrase.h"
#include "GenerationDictionary.h"
#include "ScoreComponentCollection.h"
#include "InputType.h"
#include "ObjectPool.h"
namespace Moses
{
class SquareMatrix;
class StaticData;
class TranslationOption;
class WordsRange;
class Hypothesis;
Feature function overhaul. Each feature function is computed in one of three ways: 1) Stateless feature functions from the phrase table/generation table: these are computed when the TranslationOption is created. They become part of the ScoreBreakdown object contained in the TranslationOption and are added to the feature value vector when a hypothesis is extended. 2) Stateless feature functions that are computed during state exploration. Currently, only WordPenalty falls into this category, but these functions implement a method Evaluate which do does not receive a Hypothesis or any contextual information. 3) Stateful feature functions: these features receive the arc information (translation option), compute some value and then return some context information. The context information created by a particular feature function is passed back to it as the previous context when a hypothesis originating at the node where the previous edge terminates is created. States in the search space may be recombined if the context information is identical. The context information must be stored in an object implementing the FFState interface. TODO: 1) the command line interface / MERT interface needs to go to named parameters that are otherwise opaque 2) StatefulFeatureFunction's Evaluate method should just take a TranslationOption and a context object. It is not good that it takes a hypothesis, because then people may be tempted to access information about the "previous" hypothesis without "declaring" this dependency. 3) Future cost estimates should be handled using feature functions. All stateful feature functions need some kind of future cost estimate. 4) Philipp's poor-man's cube pruning is broken. git-svn-id: https://mosesdecoder.svn.sourceforge.net/svnroot/mosesdecoder/trunk@2087 1f5c12ca-751b-0410-a591-d2e778427230
2009-02-06 18:43:06 +03:00
class FFState;
2013-05-24 22:11:15 +04:00
class StatelessFeatureFunction;
class StatefulFeatureFunction;
class Manager;
class LexicalReordering;
typedef std::vector<Hypothesis*> ArcList;
/** Used to store a state in the beam search
for the best translation. With its link back to the previous hypothesis
m_prevHypo, we can trace back to the sentence start to read of the
(partial) translation to this point.
The expansion of hypotheses is handled in the class Manager, which
stores active hypothesis in the search in hypothesis stacks.
***/
class Hypothesis
{
friend std::ostream& operator<<(std::ostream&, const Hypothesis&);
protected:
static ObjectPool<Hypothesis> s_objectPool;
const Hypothesis* m_prevHypo; /*! backpointer to previous hypothesis (from which this one was created) */
// const Phrase &m_targetPhrase; /*! target phrase being created at the current decoding step */
const TargetPhrase &m_targetPhrase; /*! target phrase being created at the current decoding step */
Phrase const* m_sourcePhrase; /*! input sentence */
WordsBitmap m_sourceCompleted; /*! keeps track of which words have been translated so far */
//TODO: how to integrate this into confusion network framework; what if
//it's a confusion network in the end???
InputType const& m_sourceInput;
WordsRange m_currSourceWordsRange; /*! source word positions of the last phrase that was used to create this hypothesis */
WordsRange m_currTargetWordsRange; /*! target word positions of the last phrase that was used to create this hypothesis */
bool m_wordDeleted;
float m_totalScore; /*! score so far */
float m_futureScore; /*! estimated future cost to translate rest of sentence */
2013-04-24 16:44:11 +04:00
ScoreComponentCollection m_scoreBreakdown; /*! scores for this hypothesis */
std::vector<const FFState*> m_ffStates;
const Hypothesis *m_winningHypo;
ArcList *m_arcList; /*! all arcs that end at the same trellis point as this hypothesis */
const TranslationOption *m_transOpt;
Manager& m_manager;
int m_id; /*! numeric ID of this hypothesis, used for logging */
/*! used by initial seeding of the translation process */
Hypothesis(Manager& manager, InputType const& source, const TargetPhrase &emptyTarget);
/*! used when creating a new hypothesis using a translation option (phrase translation) */
Hypothesis(const Hypothesis &prevHypo, const TranslationOption &transOpt);
public:
static ObjectPool<Hypothesis> &GetObjectPool() {
return s_objectPool;
}
~Hypothesis();
/** return the subclass of Hypothesis most appropriate to the given translation option */
static Hypothesis* Create(const Hypothesis &prevHypo, const TranslationOption &transOpt, const Phrase* constraint);
static Hypothesis* Create(Manager& manager, const WordsBitmap &initialCoverage);
/** return the subclass of Hypothesis most appropriate to the given target phrase */
static Hypothesis* Create(Manager& manager, InputType const& source, const TargetPhrase &emptyTarget);
/** return the subclass of Hypothesis most appropriate to the given translation option */
Hypothesis* CreateNext(const TranslationOption &transOpt, const Phrase* constraint) const;
void PrintHypothesis() const;
const InputType& GetInput() const {
return m_sourceInput;
}
/** return target phrase used to create this hypothesis */
// const Phrase &GetCurrTargetPhrase() const
const TargetPhrase &GetCurrTargetPhrase() const {
return m_targetPhrase;
}
/** return input positions covered by the translation option (phrasal translation) used to create this hypothesis */
inline const WordsRange &GetCurrSourceWordsRange() const {
return m_currSourceWordsRange;
}
inline const WordsRange &GetCurrTargetWordsRange() const {
return m_currTargetWordsRange;
}
Manager& GetManager() const {
return m_manager;
}
/** output length of the translation option used to create this hypothesis */
inline size_t GetCurrTargetLength() const {
return m_currTargetWordsRange.GetNumWordsCovered();
}
void CalcScore(const SquareMatrix &futureScore);
int GetId()const {
return m_id;
}
const Hypothesis* GetPrevHypo() const;
/** length of the partial translation (from the start of the sentence) */
inline size_t GetSize() const {
return m_currTargetWordsRange.GetEndPos() + 1;
}
inline const Phrase* GetSourcePhrase() const {
return m_sourcePhrase;
}
std::string GetSourcePhraseStringRep(const std::vector<FactorType> factorsToPrint) const;
std::string GetTargetPhraseStringRep(const std::vector<FactorType> factorsToPrint) const;
std::string GetSourcePhraseStringRep() const;
std::string GetTargetPhraseStringRep() const;
/** curr - pos is relative from CURRENT hypothesis's starting index
* (ie, start of sentence would be some negative number, which is
* not allowed- USE WITH CAUTION) */
inline const Word &GetCurrWord(size_t pos) const {
return m_targetPhrase.GetWord(pos);
}
inline const Factor *GetCurrFactor(size_t pos, FactorType factorType) const {
return m_targetPhrase.GetFactor(pos, factorType);
}
/** recursive - pos is relative from start of sentence */
inline const Word &GetWord(size_t pos) const {
const Hypothesis *hypo = this;
while (pos < hypo->GetCurrTargetWordsRange().GetStartPos()) {
hypo = hypo->GetPrevHypo();
CHECK(hypo != NULL);
}
return hypo->GetCurrWord(pos - hypo->GetCurrTargetWordsRange().GetStartPos());
}
inline const Factor* GetFactor(size_t pos, FactorType factorType) const {
return GetWord(pos)[factorType];
}
/***
* \return The bitmap of source words we cover
*/
inline const WordsBitmap &GetWordsBitmap() const {
return m_sourceCompleted;
}
inline bool IsSourceCompleted() const {
return m_sourceCompleted.IsComplete();
}
int RecombineCompare(const Hypothesis &compare) const;
void ToStream(std::ostream& out) const {
if (m_prevHypo != NULL) {
m_prevHypo->ToStream(out);
}
out << (Phrase) GetCurrTargetPhrase();
}
void ToStringStream(std::stringstream& out) const {
if (m_prevHypo != NULL) {
m_prevHypo->ToStream(out);
}
out << (Phrase) GetCurrTargetPhrase();
}
std::string GetOutputString() const {
std::stringstream out;
ToStringStream(out);
return out.str();
}
TO_STRING();
inline void SetWinningHypo(const Hypothesis *hypo) {
m_winningHypo = hypo;
}
inline const Hypothesis *GetWinningHypo() const {
return m_winningHypo;
}
void AddArc(Hypothesis *loserHypo);
void CleanupArcList();
//! returns a list alternative previous hypotheses (or NULL if n-best support is disabled)
inline const ArcList* GetArcList() const {
return m_arcList;
}
const ScoreComponentCollection& GetScoreBreakdown() const {
2013-04-24 16:44:11 +04:00
return m_scoreBreakdown;
}
float GetTotalScore() const {
return m_totalScore;
}
float GetScore() const {
return m_totalScore-m_futureScore;
}
2012-07-02 18:57:54 +04:00
const FFState* GetFFState(int idx) const {
return m_ffStates[idx];
}
void SetFFState(int idx, FFState* state) {
m_ffStates[idx] = state;
}
2012-07-02 18:57:54 +04:00
// Added by oliver.wilson@ed.ac.uk for async lm stuff.
2013-04-22 16:03:07 +04:00
void EvaluateWith(const StatefulFeatureFunction &sfff, int state_idx);
void EvaluateWith(const StatelessFeatureFunction &slff);
//! target span that trans opt would populate if applied to this hypo. Used for alignment check
size_t GetNextStartPos(const TranslationOption &transOpt) const;
std::vector<std::vector<unsigned int> > *GetLMStats() const {
return NULL;
}
const TranslationOption &GetTranslationOption() const {
return *m_transOpt;
}
};
std::ostream& operator<<(std::ostream& out, const Hypothesis& hypothesis);
// sorting helper
struct CompareHypothesisTotalScore {
bool operator()(const Hypothesis* hypo1, const Hypothesis* hypo2) const {
return hypo1->GetTotalScore() > hypo2->GetTotalScore();
}
};
#ifdef USE_HYPO_POOL
#define FREEHYPO(hypo) \
{ \
ObjectPool<Hypothesis> &pool = Hypothesis::GetObjectPool(); \
pool.freeObject(hypo); \
} \
#else
#define FREEHYPO(hypo) delete hypo
#endif
/** defines less-than relation on hypotheses.
* The particular order is not important for us, we need just to figure out
* which hypothesis are equal based on:
* the last n-1 target words are the same
* and the covers (source words translated) are the same
* Directly using RecombineCompare is unreliable because the Compare methods
* of some states are based on archictecture-dependent pointer comparisons.
* That's why we use the hypothesis IDs instead.
*/
class HypothesisRecombinationOrderer
{
public:
bool operator()(const Hypothesis* hypoA, const Hypothesis* hypoB) const {
return (hypoA->RecombineCompare(*hypoB) < 0);
}
};
}
#endif