mirror of
https://github.com/Helsinki-NLP/OPUS-MT-train.git
synced 2024-10-05 16:47:21 +03:00
changes on mahti
This commit is contained in:
parent
6f2c9d8bd5
commit
d45e7aaf7c
@ -557,8 +557,8 @@ ifeq (${CONTINUE_EXISTING},1)
|
||||
MODEL_LATEST = $(firstword \
|
||||
${shell ls -t ${WORKDIR}/*${MODEL_VARIANT}.${PRE_SRC}-${PRE_TRG}.${subst -align,,${MODELTYPE}}.model[0-9].npz \
|
||||
${WORKDIR}/*${MODEL_VARIANT}.${PRE_SRC}-${PRE_TRG}.${subst -align,,${MODELTYPE}}-align.model[0-9].npz \
|
||||
${WORKDIR}/*${MODEL_VARIANT}.${PRE_SRC}-${PRE_TRG}.${subst -align,,${MODELTYPE}}.best-perplexity.npz \
|
||||
${WORKDIR}/*${MODEL_VARIANT}.${PRE_SRC}-${PRE_TRG}.${subst -align,,${MODELTYPE}}-align.best-perplexity.npz \
|
||||
${WORKDIR}/*${MODEL_VARIANT}.${PRE_SRC}-${PRE_TRG}.${subst -align,,${MODELTYPE}}.model[0-9].npz.best-perplexity.npz \
|
||||
${WORKDIR}/*${MODEL_VARIANT}.${PRE_SRC}-${PRE_TRG}.${subst -align,,${MODELTYPE}}-align.model[0-9].npz.best-perplexity.npz \
|
||||
2>/dev/null | grep -v 'tuned4' })
|
||||
MODEL_LATEST_VOCAB = $(shell echo "${MODEL_LATEST}" | \
|
||||
sed 's|\.${PRE_SRC}-${PRE_TRG}\..*$$|.${PRE_SRC}-${PRE_TRG}.vocab.yml|')
|
||||
|
16
lib/data.mk
16
lib/data.mk
@ -119,6 +119,14 @@ ifeq (${USE_PIVOTING},1)
|
||||
endif
|
||||
|
||||
|
||||
# additional data sets that might be available ...
|
||||
ifeq (${USE_EXTRA_BITEXTS},1)
|
||||
EXTRA_BITEXTS_SRC = ${sort ${wildcard ${DATADIR}/extra/${SRC}-${TRG}/*.${SRCEXT}.gz}}
|
||||
EXTRA_BITEXTS_TRG = ${patsubst %.${SRCEXT}.gz,%.${TRGEXT}.gz,${EXTRA_BITEXTS_SRC}}
|
||||
endif
|
||||
|
||||
|
||||
|
||||
print-datasets:
|
||||
-@for s in ${SRCLANGS}; do \
|
||||
for t in ${TRGLANGS}; do \
|
||||
@ -144,6 +152,10 @@ print-datasets-current-langpair:
|
||||
@echo "pivot-based translation data:"
|
||||
@echo ${PIVOTING_SRC}
|
||||
@echo ${PIVOTING_TRG}
|
||||
@echo "extra bitexts:"
|
||||
@echo ${EXTRA_BITEXTS_SRC}
|
||||
@echo ${EXTRA_BITEXTS_TRG}
|
||||
|
||||
|
||||
##-------------------------------------------------------------
|
||||
## data sets (train/dev/test)
|
||||
@ -153,9 +165,9 @@ print-datasets-current-langpair:
|
||||
## with some basic pre-processing (see lib/preprocess.mk)
|
||||
|
||||
CLEAN_TRAIN_SRC = ${patsubst %,${DATADIR}/${PRE}/%.${LANGPAIR}.${CLEAN_TRAINDATA_TYPE}.${SRCEXT}.gz,${TRAINSET}} \
|
||||
${BACKTRANS_SRC} ${FORWARDTRANS_SRC} ${FORWARDTRANSMONO_SRC} ${PIVOTING_SRC}
|
||||
${BACKTRANS_SRC} ${FORWARDTRANS_SRC} ${FORWARDTRANSMONO_SRC} ${PIVOTING_SRC} ${EXTRA_BITEXTS_SRC}
|
||||
CLEAN_TRAIN_TRG = ${patsubst %,${DATADIR}/${PRE}/%.${LANGPAIR}.${CLEAN_TRAINDATA_TYPE}.${TRGEXT}.gz,${TRAINSET}} \
|
||||
${BACKTRANS_TRG} ${FORWARDTRANS_TRG} ${FORWARDTRANSMONO_TRG} ${PIVOTING_TRG}
|
||||
${BACKTRANS_TRG} ${FORWARDTRANS_TRG} ${FORWARDTRANSMONO_TRG} ${PIVOTING_TRG} ${EXTRA_BITEXTS_TRG}
|
||||
|
||||
CLEAN_DEV_SRC = ${patsubst %,${DATADIR}/${PRE}/%.${LANGPAIR}.${CLEAN_DEVDATA_TYPE}.${SRCEXT}.gz,${DEVSET}}
|
||||
CLEAN_DEV_TRG = ${patsubst %.${SRCEXT}.gz,%.${TRGEXT}.gz,${CLEAN_DEV_SRC}}
|
||||
|
1
lib/env/mahti.mk
vendored
1
lib/env/mahti.mk
vendored
@ -8,6 +8,7 @@ DATA_PREPARE_HPCPARAMS = CPUJOB_HPC_CORES=2 CPUJOB_HPC_MEM=16g
|
||||
DATA_ALIGN_HPCPARAMS = CPUJOB_HPC_CORES=128 CPUJOB_HPC_JOBS=20 CPUJOB_HPC_MEM=128g
|
||||
|
||||
|
||||
# CSCPROJECT = project_2002982
|
||||
CSCPROJECT = project_2002688
|
||||
# CSCPROJECT = project_2005625
|
||||
WORKHOME = ${shell realpath ${PWD}/work}
|
||||
|
@ -245,6 +245,11 @@ listallmodels:
|
||||
${MAKE} DATASET=${DATASET}+bt USE_BACKTRANS=1 SHUFFLE_TRAINING_DATA=1 ${@:-bt=}
|
||||
|
||||
|
||||
## include additional bitexts
|
||||
%-xb:
|
||||
${MAKE} DATASET=${DATASET}+xb USE_EXTRA_BITEXTS=1 SHUFFLE_TRAINING_DATA=1 ${@:-xb=}
|
||||
|
||||
|
||||
## adding a pivot language to the model
|
||||
## --> add pivot language to each side (source and target)
|
||||
## --> only start the task if the pivot language adds anything on either side
|
||||
|
@ -190,6 +190,14 @@ elg-ukr2lit-tiny11:
|
||||
${MAKE} STUDENT_DATA=pft-pbt-bt SRCLANGS=ukr TRGLANGS=lit train-tiny11-student
|
||||
|
||||
|
||||
elg-pol2ukr-student2:
|
||||
${MAKE} MARIAN_EXTRA=--no-restore-corpus MARIAN_EARLY_STOPPING=15 CHECK_TRAINDATA_SIZE=1 CLEAN_CORPUS_TRAINING_DATA=1 STUDENT_DATA=pft-pbt-bt-xb SRCLANGS=pol TRGLANGS=ukr train-tiny11-student
|
||||
|
||||
elg-ukr2pol-student2:
|
||||
${MAKE} MARIAN_EXTRA=--no-restore-corpus MARIAN_EARLY_STOPPING=15 CHECK_TRAINDATA_SIZE=1 CLEAN_CORPUS_TRAINING_DATA=1 STUDENT_DATA=pft-pbt-bt-xb SRCLANGS=ukr TRGLANGS=pol train-tiny11-student
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
elg-ces_slk2ukr-tiny11:
|
||||
@ -233,22 +241,31 @@ elg-ukr2deu-student3:
|
||||
|
||||
|
||||
|
||||
elg-fin2ukr-student2:
|
||||
${MAKE} SUBWORD_VOCAB_SIZE=16000 MARIAN_EARLY_STOPPING=15 SRCLANGS=fin TRGLANGS=ukr CHECK_TRAINDATA_SIZE=1 CLEAN_CORPUS_TRAINING_DATA=1 STUDENT_DATA=ftbest-ftmono-nopar train-tiny11-student
|
||||
|
||||
|
||||
elg-fin2ukr-student:
|
||||
${MAKE} SRCLANGS=fin TRGLANGS=ukr CHECK_TRAINDATA_SIZE=1 CLEAN_CORPUS_TRAINING_DATA=1 STUDENT_DATA=ftbest-ftmono-nopar train-tiny11-student
|
||||
${MAKE} MARIAN_EARLY_STOPPING=15 SRCLANGS=fin TRGLANGS=ukr CHECK_TRAINDATA_SIZE=1 CLEAN_CORPUS_TRAINING_DATA=1 STUDENT_DATA=ftbest-ftmono-nopar train-tiny11-student
|
||||
|
||||
elg-ukr2fin-student:
|
||||
${MAKE} SRCLANGS=ukr TRGLANGS=fin CHECK_TRAINDATA_SIZE=1 CLEAN_CORPUS_TRAINING_DATA=1 STUDENT_DATA=ftbest-ftmono-nopar train-tiny11-student
|
||||
${MAKE} MARIAN_EXTRA=--no-restore-corpus MARIAN_EARLY_STOPPING=15 SRCLANGS=ukr TRGLANGS=fin CHECK_TRAINDATA_SIZE=1 CLEAN_CORPUS_TRAINING_DATA=1 STUDENT_DATA=ftbest-ftmono-nopar train-tiny11-student
|
||||
|
||||
elg-zle2fin-tiny11:
|
||||
${MAKE} MARIAN_EXTRA=--no-restore-corpus \
|
||||
elg-zle2fin-student:
|
||||
${MAKE} MARIAN_EXTRA=--no-restore-corpus MARIAN_EARLY_STOPPING=15 \
|
||||
DATA_PREPARE_HPCPARAMS='${DATA_PREPARE_HPCPARAMS} CPUJOB_HPC_DISK=1000' \
|
||||
DATA_ALIGN_HPCPARAMS="${DATA_ALIGN_HPCPARAMS} CPUJOB_HPC_DISK=1000" \
|
||||
CHECK_TRAINDATA_SIZE=1 CLEAN_CORPUS_TRAINING_DATA=1 \
|
||||
STUDENT_DATA=ftbest-ftmono-nopar SRCLANGS="ukr rus" TRGLANGS=fin \
|
||||
LANGPAIRSTR="zle-ukr" train-tiny11-student
|
||||
LANGPAIRSTR="zle-fin-tiny" train-tiny11-student
|
||||
|
||||
|
||||
elg-fin2rus-student:
|
||||
${MAKE} MARIAN_EARLY_STOPPING=15 SRCLANGS=fin TRGLANGS=rus CHECK_TRAINDATA_SIZE=1 CLEAN_CORPUS_TRAINING_DATA=1 STUDENT_DATA=ftbest-ftmono-nopar train-tiny11-student
|
||||
|
||||
elg-rus2fin-student:
|
||||
${MAKE} MARIAN_EARLY_STOPPING=15 SRCLANGS=rus TRGLANGS=fin CHECK_TRAINDATA_SIZE=1 CLEAN_CORPUS_TRAINING_DATA=1 STUDENT_DATA=ftbest-ftmono-nopar train-tiny11-student
|
||||
|
||||
|
||||
|
||||
|
||||
@ -309,7 +326,10 @@ elg-dist-missing:
|
||||
|
||||
|
||||
elg-zle2fin-pivot:
|
||||
${MAKE} MODELTYPE=transformer-big tatoeba-zle2fin-trainjob-pbt-pft-bt
|
||||
${MAKE} MARIAN_EXTRA=--no-restore-corpus MODELTYPE=transformer-big tatoeba-zle2fin-trainjob-pbt-pft-bt
|
||||
|
||||
elg-fin2zle-pivot:
|
||||
${MAKE} MARIAN_EXTRA=--no-restore-corpus MODELTYPE=transformer-big tatoeba-fin2zle-trainjob-pbt-pft-bt
|
||||
|
||||
|
||||
|
||||
@ -404,6 +424,12 @@ elg-eval-big2zle:
|
||||
${MAKE} MODELTYPE=transformer-big tatoeba-$${l}2zle-eval-testsets; \
|
||||
done
|
||||
|
||||
elg-eng2zle-xb:
|
||||
${MAKE} MARIAN_EXTRA=--no-restore-corpus MODELTYPE=transformer-big CONTINUE_EXISTING=1 tatoeba-eng2zle-trainjob-bt-xb
|
||||
|
||||
elg-zle2eng-xb:
|
||||
${MAKE} MARIAN_EARLY_STOPPING=25 MARIAN_EXTRA=--no-restore-corpus MODELTYPE=transformer-big CONTINUE_EXISTING=1 tatoeba-zle2eng-trainjob-bt-xb
|
||||
|
||||
|
||||
elg-pivot-eval:
|
||||
for l in dan swe fin deu ron tur; do \
|
||||
|
12
lib/tasks.mk
12
lib/tasks.mk
@ -345,3 +345,15 @@ job1-step3:
|
||||
${MAKE} all-bt
|
||||
|
||||
|
||||
|
||||
|
||||
print-info:
|
||||
@echo "model file: ${MODEL_START}"
|
||||
@echo "source vocab: ${MODEL_SRCVOCAB}"
|
||||
@echo "target vocab: ${MODEL_TRGVOCAB}"
|
||||
@echo "final model file: ${MODEL_FINAL}"
|
||||
@echo "latest compatible model: ${MODEL_LATEST}"
|
||||
ls -t ${WORKDIR}/*${MODEL_VARIANT}.${PRE_SRC}-${PRE_TRG}.${subst -align,,${MODELTYPE}}.model[0-9].npz \
|
||||
${WORKDIR}/*${MODEL_VARIANT}.${PRE_SRC}-${PRE_TRG}.${subst -align,,${MODELTYPE}}-align.model[0-9].npz \
|
||||
${WORKDIR}/*${MODEL_VARIANT}.${PRE_SRC}-${PRE_TRG}.${subst -align,,${MODELTYPE}}.best-perplexity.npz \
|
||||
${WORKDIR}/*${MODEL_VARIANT}.${PRE_SRC}-${PRE_TRG}.${subst -align,,${MODELTYPE}}-align.best-perplexity.npz
|
||||
|
@ -348,7 +348,6 @@ tatoeba-%-train:
|
||||
TATOEBA_TRGLANG_GROUP="`langgroup -n $$t`" \
|
||||
all-groupsize-limits; )
|
||||
|
||||
|
||||
## start the training job
|
||||
## - create config file
|
||||
## - create data sets
|
||||
@ -383,6 +382,17 @@ tatoeba-%-pivotlang:
|
||||
fi
|
||||
|
||||
|
||||
tatoeba-%-info:
|
||||
-( s=$(firstword $(subst 2, ,$(patsubst tatoeba-%-info,%,$@))); \
|
||||
t=$(lastword $(subst 2, ,$(patsubst tatoeba-%-info,%,$@))); \
|
||||
S="${call find-srclanggroup,${patsubst tatoeba-%-info,%,$@},${PIVOT}}"; \
|
||||
T="${call find-trglanggroup,${patsubst tatoeba-%-info,%,$@},${PIVOT}}"; \
|
||||
${MAKE} LANGPAIRSTR=$$s-$$t SRCLANGS="$$S" TRGLANGS="$$T" \
|
||||
TATOEBA_SRCLANG_GROUP="`langgroup -n $$s`" \
|
||||
TATOEBA_TRGLANG_GROUP="`langgroup -n $$t`" \
|
||||
print-info; )
|
||||
|
||||
|
||||
## evaluate with the model-specific test set
|
||||
tatoeba-%-eval:
|
||||
( s=$(firstword $(subst 2, ,$(patsubst tatoeba-%-eval,%,$@))); \
|
||||
|
Loading…
Reference in New Issue
Block a user