# Train Opus-MT models This package includes scripts for training NMT models using MarianNMT and OPUS data for [OPUS-MT](https://github.com/Helsinki-NLP/Opus-MT). More details are given in the [Makefile](Makefile) but documentation needs to be improved. Also, the targets require a specific environment and right now only work well on the CSC HPC cluster in Finland. ## Structure Essential files for making new models: * `Makefile`: top-level makefile * `lib/env.mk`: system-specific environment (now based on CSC machines) * `lib/config.mk`: essential model configuration * `lib/data.mk`: data pre-processing tasks * `lib/generic.mk`: generic implicit rules that can extend other tasks * `lib/dist.mk`: make packages for distributing models (CSC ObjectStorage based) * `lib/slurm.mk`: submit jobs with SLURM There are also make targets for specific models and tasks. Look into `lib/models/` to see what has been defined already. Note that this frequently changes! There is, for example: * `lib/models/multilingua.mk`: various multilingual models * `lib/models/celtic.mk`: data and models for Celtic languages * `lib/models/doclevel.mk`: experimental document-level models Run this if you want to train a model, for example for translating English to French: ``` make SRCLANG=en TRGLANG=fr train ``` To evaluate the model with the automatically generated test data (from the Tatoeba corpus as a default) run: ``` make SRCLANG=en TRGLANG=fr eval ``` For multilingual (more than one language on either side) models run, for example: ``` make SRCLANG="de en" TRGLANG="fr es pt" train make SRCLANG="de en" TRGLANG="fr es pt" eval ``` Note that data pre-processing should run on CPUs and training/testing on GPUs. To speed up things you can process data sets in parallel using the jobs flag of make, for example using 8 threads: ``` make -j 8 SRCLANG=en TRGLANG=fr data ``` ## Upload to Object Storage ``` swift upload OPUS-MT --changed --skip-identical name-of-file swift post OPUS-MT --read-acl ".r:*" ```