From 33b29442982f2569155f2fe31e1a4cecbc508bfa Mon Sep 17 00:00:00 2001 From: Abdulrhman Alkhodiry Date: Sun, 30 Jun 2024 10:41:10 +0300 Subject: [PATCH] Update dependencies and fix convert_model (#458) * feat: Update dependencies in Cargo.toml Update the dependencies in Cargo.toml to their latest versions: - rust_tokenizers: 8.1.1 - tch: 0.16.0 (with features = ["download-libtorch"]) - serde_json: 1 - serde: 1 (with features = ["derive"]) - ordered-float: 4.2.0 - uuid: 1 (with features = ["v4"]) - thiserror: 1 - half: 2 - regex: 1.6 - cached-path: 0.6 (with default-features = false and optional = true) - dirs: 5 (optional = true) - lazy_static: 1 (optional = true) - ort: 1.16.3 (optional = true, default-features = false, features = ["half"]) - ndarray: 0.15 (optional = true) - tokenizers: 0.19.1 (optional = true, default-features = false, features = ["onig"]) ``` * chore: Update .gitignore and requirements.txt, and improve convert_model.py Update .gitignore to exclude the /models/ and /.venv/ directories, and the convert_model.log file. Remove the requirements.txt file. In convert_model.py: - Add a new function, `zipfile_factory`, to handle zip file creation. - Update the logger configuration to log debug messages to a file named `convert_model.log`. * delete duplicate requirements file * update CI req file path * missing requests dependency --------- Co-authored-by: Abdulrhman Alkhodiry Co-authored-by: Guillaume Becquin --- .github/workflows/continuous-integration.yml | 2 +- .gitignore | 4 +- Cargo.toml | 23 +- README.md | 434 +++++++++++++------ requirements.txt | 3 - src/models/bart/decoder.rs | 1 + src/models/prophetnet/encoder.rs | 1 + src/models/reformer/encoder.rs | 1 + src/models/reformer/reformer_model.rs | 1 + src/models/t5/encoder.rs | 1 + utils/convert_model.py | 23 +- utils/requirements.txt | 11 + 12 files changed, 351 insertions(+), 154 deletions(-) delete mode 100644 requirements.txt create mode 100644 utils/requirements.txt diff --git a/.github/workflows/continuous-integration.yml b/.github/workflows/continuous-integration.yml index 666f4a4..07349c6 100644 --- a/.github/workflows/continuous-integration.yml +++ b/.github/workflows/continuous-integration.yml @@ -174,7 +174,7 @@ jobs: with: python-version: '3.10' - run: | - pip install -r requirements.txt --progress-bar off + pip install -r ./utils/requirements.txt --progress-bar off python ./utils/download-dependencies_distilbert.py fmt: diff --git a/.gitignore b/.gitignore index d9d3001..ad47214 100644 --- a/.gitignore +++ b/.gitignore @@ -17,4 +17,6 @@ Cargo.lock /target #**/*.rs.bk -/resources/ +/models/ +/.venv/ +convert_model.log \ No newline at end of file diff --git a/Cargo.toml b/Cargo.toml index 273ce3a..b63f11a 100644 --- a/Cargo.toml +++ b/Cargo.toml @@ -76,10 +76,10 @@ features = ["doc-only"] [dependencies] rust_tokenizers = "8.1.1" -tch = "0.15.0" +tch = { version = "0.16.0", features = ["download-libtorch"] } serde_json = "1" serde = { version = "1", features = ["derive"] } -ordered-float = "3" +ordered-float = "4.2.0" uuid = { version = "1", features = ["v4"] } thiserror = "1" half = "2" @@ -88,9 +88,13 @@ regex = "1.6" cached-path = { version = "0.6", default-features = false, optional = true } dirs = { version = "5", optional = true } lazy_static = { version = "1", optional = true } -ort = {version="~1.15.2", optional = true, default-features = false, features = ["half"]} -ndarray = {version="0.15", optional = true} -tokenizers = {version="0.15", optional=true, default-features = false, features = ["onig"]} +ort = { version = "1.16.3", optional = true, default-features = false, features = [ + "half", +] } +ndarray = { version = "0.15", optional = true } +tokenizers = { version = "0.19.1", optional = true, default-features = false, features = [ + "onig", +] } [dev-dependencies] anyhow = "1" @@ -98,9 +102,12 @@ csv = "1" criterion = "0.5" tokio = { version = "1.35", features = ["sync", "rt-multi-thread", "macros"] } tempfile = "3" -itertools = "0.12" -tracing-subscriber = { version = "0.3", default-features = false, features = [ "env-filter", "fmt" ] } -ort = {version="~1.15.5", features = ["load-dynamic"]} +itertools = "0.13.0" +tracing-subscriber = { version = "0.3", default-features = false, features = [ + "env-filter", + "fmt", +] } +ort = { version = "1.16.3", features = ["load-dynamic"] } [[example]] name = "onnx-masked-lm" diff --git a/README.md b/README.md index 0d28157..1ec95bf 100644 --- a/README.md +++ b/README.md @@ -5,10 +5,21 @@ [![Documentation](https://docs.rs/rust-bert/badge.svg)](https://docs.rs/rust-bert) ![License](https://img.shields.io/crates/l/rust_bert.svg) -Rust-native state-of-the-art Natural Language Processing models and pipelines. Port of Hugging Face's [Transformers library](https://github.com/huggingface/transformers), using [tch-rs](https://github.com/LaurentMazare/tch-rs) or [onnxruntime bindings](https://github.com/pykeio/ort) and pre-processing from [rust-tokenizers](https://github.com/guillaume-be/rust-tokenizers). Supports multi-threaded tokenization and GPU inference. -This repository exposes the model base architecture, task-specific heads (see below) and [ready-to-use pipelines](#ready-to-use-pipelines). [Benchmarks](#benchmarks) are available at the end of this document. +Rust-native state-of-the-art Natural Language Processing models and pipelines. +Port of Hugging Face's +[Transformers library](https://github.com/huggingface/transformers), using +[tch-rs](https://github.com/LaurentMazare/tch-rs) or +[onnxruntime bindings](https://github.com/pykeio/ort) and pre-processing from +[rust-tokenizers](https://github.com/guillaume-be/rust-tokenizers). Supports +multi-threaded tokenization and GPU inference. This repository exposes the model +base architecture, task-specific heads (see below) and +[ready-to-use pipelines](#ready-to-use-pipelines). [Benchmarks](#benchmarks) are +available at the end of this document. + +Get started with tasks including question answering, named entity recognition, +translation, summarization, text generation, conversational agents and more in +just a few lines of code: -Get started with tasks including question answering, named entity recognition, translation, summarization, text generation, conversational agents and more in just a few lines of code: ```rust let qa_model = QuestionAnsweringModel::new(Default::default())?; @@ -19,84 +30,104 @@ Get started with tasks including question answering, named entity recognition, t ``` Output: + ``` [Answer { score: 0.9976, start: 13, end: 21, answer: "Amsterdam" }] ``` The tasks currently supported include: - - Translation - - Summarization - - Multi-turn dialogue - - Zero-shot classification - - Sentiment Analysis - - Named Entity Recognition - - Part of Speech tagging - - Question-Answering - - Language Generation - - Masked Language Model - - Sentence Embeddings - - Keywords extraction + +- Translation +- Summarization +- Multi-turn dialogue +- Zero-shot classification +- Sentiment Analysis +- Named Entity Recognition +- Part of Speech tagging +- Question-Answering +- Language Generation +- Masked Language Model +- Sentence Embeddings +- Keywords extraction
Expand to display the supported models/tasks matrix -| |**Sequence classification**|**Token classification**|**Question answering**|**Text Generation**|**Summarization**|**Translation**|**Masked LM**|**Sentence Embeddings**| -:-----:|:----:|:----:|:-----:|:----:|:-----:|:----:|:----:|:----: -DistilBERT|✅|✅|✅| | | |✅| ✅| -MobileBERT|✅|✅|✅| | | |✅| | -DeBERTa|✅|✅|✅| | | |✅| | -DeBERTa (v2)|✅|✅|✅| | | |✅| | -FNet|✅|✅|✅| | | |✅| | -BERT|✅|✅|✅| | | |✅| ✅| -RoBERTa|✅|✅|✅| | | |✅| ✅| -GPT| | | |✅ | | | | | -GPT2| | | |✅ | | | | | -GPT-Neo| | | |✅ | | | | | -GPT-J| | | |✅ | | | | | -BART|✅| | |✅ |✅| | | | -Marian| | | | | |✅| | | -MBart|✅| | |✅ | | | | | -M2M100| | | |✅ | | | | | -NLLB| | | |✅ | | | | | -Electra | |✅| | | | |✅| | -ALBERT |✅|✅|✅| | | |✅| ✅ | -T5 | | | |✅ |✅|✅| | ✅ | -LongT5 | | | |✅ |✅|| | | -XLNet|✅|✅|✅|✅ | | |✅| | -Reformer|✅| |✅|✅ | | |✅| | -ProphetNet| | | |✅ |✅ | | | | -Longformer|✅|✅|✅| | | |✅| | -Pegasus| | | | |✅| | | | +| | **Sequence classification** | **Token classification** | **Question answering** | **Text Generation** | **Summarization** | **Translation** | **Masked LM** | **Sentence Embeddings** | +| :----------: | :-------------------------: | :----------------------: | :--------------------: | :-----------------: | :---------------: | :-------------: | :-----------: | :---------------------: | +| DistilBERT | ✅ | ✅ | ✅ | | | | ✅ | ✅ | +| MobileBERT | ✅ | ✅ | ✅ | | | | ✅ | | +| DeBERTa | ✅ | ✅ | ✅ | | | | ✅ | | +| DeBERTa (v2) | ✅ | ✅ | ✅ | | | | ✅ | | +| FNet | ✅ | ✅ | ✅ | | | | ✅ | | +| BERT | ✅ | ✅ | ✅ | | | | ✅ | ✅ | +| RoBERTa | ✅ | ✅ | ✅ | | | | ✅ | ✅ | +| GPT | | | | ✅ | | | | | +| GPT2 | | | | ✅ | | | | | +| GPT-Neo | | | | ✅ | | | | | +| GPT-J | | | | ✅ | | | | | +| BART | ✅ | | | ✅ | ✅ | | | | +| Marian | | | | | | ✅ | | | +| MBart | ✅ | | | ✅ | | | | | +| M2M100 | | | | ✅ | | | | | +| NLLB | | | | ✅ | | | | | +| Electra | | ✅ | | | | | ✅ | | +| ALBERT | ✅ | ✅ | ✅ | | | | ✅ | ✅ | +| T5 | | | | ✅ | ✅ | ✅ | | ✅ | +| LongT5 | | | | ✅ | ✅ | | | | +| XLNet | ✅ | ✅ | ✅ | ✅ | | | ✅ | | +| Reformer | ✅ | | ✅ | ✅ | | | ✅ | | +| ProphetNet | | | | ✅ | ✅ | | | | +| Longformer | ✅ | ✅ | ✅ | | | | ✅ | | +| Pegasus | | | | | ✅ | | | | +
## Getting started -This library relies on the [tch](https://github.com/LaurentMazare/tch-rs) crate for bindings to the C++ Libtorch API. -The libtorch library is required can be downloaded either automatically or manually. The following provides a reference on how to set-up your environment -to use these bindings, please refer to the [tch](https://github.com/LaurentMazare/tch-rs) for detailed information or support. +This library relies on the [tch](https://github.com/LaurentMazare/tch-rs) crate +for bindings to the C++ Libtorch API. The libtorch library is required can be +downloaded either automatically or manually. The following provides a reference +on how to set-up your environment to use these bindings, please refer to the +[tch](https://github.com/LaurentMazare/tch-rs) for detailed information or +support. -Furthermore, this library relies on a cache folder for downloading pre-trained models. -This cache location defaults to `~/.cache/.rustbert`, but can be changed by setting the `RUSTBERT_CACHE` environment variable. Note that the language models used by this library are in the order of the 100s of MBs to GBs. +Furthermore, this library relies on a cache folder for downloading pre-trained +models. This cache location defaults to `~/.cache/.rustbert`, but can be changed +by setting the `RUSTBERT_CACHE` environment variable. Note that the language +models used by this library are in the order of the 100s of MBs to GBs. ### Manual installation (recommended) -1. Download `libtorch` from https://pytorch.org/get-started/locally/. This package requires `v2.2`: if this version is no longer available on the "get started" page, -the file should be accessible by modifying the target link, for example `https://download.pytorch.org/libtorch/cu121/libtorch-cxx11-abi-shared-with-deps-2.2.0%2Bcu121.zip` for a Linux version with CUDA12. **NOTE:** When using `rust-bert` as dependency from [crates.io](https://crates.io), please check the required `LIBTORCH` on the published package [readme](https://crates.io/crates/rust-bert) as it may differ from the version documented here (applying to the current repository version). +1. Download `libtorch` from https://pytorch.org/get-started/locally/. This + package requires `v2.2`: if this version is no longer available on the "get + started" page, the file should be accessible by modifying the target link, + for example + `https://download.pytorch.org/libtorch/cu121/libtorch-cxx11-abi-shared-with-deps-2.2.0%2Bcu121.zip` + for a Linux version with CUDA12. **NOTE:** When using `rust-bert` as + dependency from [crates.io](https://crates.io), please check the required + `LIBTORCH` on the published package + [readme](https://crates.io/crates/rust-bert) as it may differ from the + version documented here (applying to the current repository version). 2. Extract the library to a location of your choice 3. Set the following environment variables + ##### Linux: + ```bash export LIBTORCH=/path/to/libtorch export LD_LIBRARY_PATH=${LIBTORCH}/lib:$LD_LIBRARY_PATH ``` ##### Windows + ```powershell $Env:LIBTORCH = "X:\path\to\libtorch" $Env:Path += ";X:\path\to\libtorch\lib" ``` #### macOS + Homebrew + ```bash brew install pytorch jq export LIBTORCH=$(brew --cellar pytorch)/$(brew info --json pytorch | jq -r '.[0].installed[0].version') @@ -105,13 +136,19 @@ export LD_LIBRARY_PATH=${LIBTORCH}/lib:$LD_LIBRARY_PATH ### Automatic installation -Alternatively, you can let the `build` script automatically download the `libtorch` library for you. The `download-libtorch` feature flag needs to be enabled. -The CPU version of libtorch will be downloaded by default. To download a CUDA version, please set the environment variable `TORCH_CUDA_VERSION` to `cu118`. -Note that the libtorch library is large (order of several GBs for the CUDA-enabled version) and the first build may therefore take several minutes to complete. +Alternatively, you can let the `build` script automatically download the +`libtorch` library for you. The `download-libtorch` feature flag needs to be +enabled. The CPU version of libtorch will be downloaded by default. To download +a CUDA version, please set the environment variable `TORCH_CUDA_VERSION` to +`cu118`. Note that the libtorch library is large (order of several GBs for the +CUDA-enabled version) and the first build may therefore take several minutes to +complete. ### Verifying installation -Verify your installation (and linking with libtorch) by adding the `rust-bert` dependency to your `Cargo.toml` or by cloning the rust-bert source and running an example: +Verify your installation (and linking with libtorch) by adding the `rust-bert` +dependency to your `Cargo.toml` or by cloning the rust-bert source and running +an example: ```bash git clone git@github.com:guillaume-be/rust-bert.git @@ -121,41 +158,73 @@ cargo run --example sentence_embeddings ## ONNX Support (Optional) -ONNX support can be enabled via the optional `onnx` feature. This crate then leverages the [ort](https://github.com/pykeio/ort) crate with bindings to the onnxruntime C++ library. We refer the user to this page project for further installation instructions/support. -1. Enable the optional `onnx` feature. The `rust-bert` crate does not include any optional dependencies for `ort`, the end user should select the set of features that would be adequate for pulling the required `onnxruntime` C++ library. -2. The current recommended installation is to use dynamic linking by pointing to an existing library location. Use the `load-dynamic` cargo feature for `ort`. -3. set the `ORT_DYLIB_PATH` to point to the location of downloaded onnxruntime library (`onnxruntime.dll`/`libonnxruntime.so`/`libonnxruntime.dylib` depending on the operating system). These can be downloaded from the [release page](https://github.com/microsoft/onnxruntime/releases) of the onnxruntime project +ONNX support can be enabled via the optional `onnx` feature. This crate then +leverages the [ort](https://github.com/pykeio/ort) crate with bindings to the +onnxruntime C++ library. We refer the user to this page project for further +installation instructions/support. -Most architectures (including encoders, decoders and encoder-decoders) are supported. the library aims at keeping compatibility with models exported using the [optimum](https://github.com/huggingface/optimum) library. A detailed guide on how to export a Transformer model to ONNX using optimum is available at https://huggingface.co/docs/optimum/main/en/exporters/onnx/usage_guides/export_a_model -The resources used to create ONNX models are similar to those based on Pytorch, replacing the pytorch by the ONNX model. Since ONNX models are less flexible than their Pytorch counterparts in the handling of optional arguments, exporting a decoder or encoder-decoder model to ONNX will usually result in multiple files. These files are expected (but not all are necessary) for use in this library as per the table below: +1. Enable the optional `onnx` feature. The `rust-bert` crate does not include + any optional dependencies for `ort`, the end user should select the set of + features that would be adequate for pulling the required `onnxruntime` C++ + library. +2. The current recommended installation is to use dynamic linking by pointing to + an existing library location. Use the `load-dynamic` cargo feature for `ort`. +3. set the `ORT_DYLIB_PATH` to point to the location of downloaded onnxruntime + library (`onnxruntime.dll`/`libonnxruntime.so`/`libonnxruntime.dylib` + depending on the operating system). These can be downloaded from the + [release page](https://github.com/microsoft/onnxruntime/releases) of the + onnxruntime project -| Architecture | Encoder file | Decoder without past file | Decoder with past file | -|-----------------------------|---------------|---------------------------|-------------------------| -| Encoder (e.g. BERT) | required | not used | not used | -| Decoder (e.g. GPT2) | not used | required | optional | -| Encoder-decoder (e.g. BART) | required | required | optional | +Most architectures (including encoders, decoders and encoder-decoders) are +supported. the library aims at keeping compatibility with models exported using +the [optimum](https://github.com/huggingface/optimum) library. A detailed guide +on how to export a Transformer model to ONNX using optimum is available at +https://huggingface.co/docs/optimum/main/en/exporters/onnx/usage_guides/export_a_model +The resources used to create ONNX models are similar to those based on Pytorch, +replacing the pytorch by the ONNX model. Since ONNX models are less flexible +than their Pytorch counterparts in the handling of optional arguments, exporting +a decoder or encoder-decoder model to ONNX will usually result in multiple +files. These files are expected (but not all are necessary) for use in this +library as per the table below: -Note that the computational efficiency will drop when the `decoder with past` file is optional but not provided -since the model will not used cached past keys and values for the attention mechanism, leading to a high number of -redundant computations. The Optimum library offers export options to ensure such a `decoder with past` model file is created. -he base encoder and decoder model architecture are available (and exposed for convenience) in the `encoder` and `decoder` modules, respectively. +| Architecture | Encoder file | Decoder without past file | Decoder with past file | +| --------------------------- | ------------ | ------------------------- | ---------------------- | +| Encoder (e.g. BERT) | required | not used | not used | +| Decoder (e.g. GPT2) | not used | required | optional | +| Encoder-decoder (e.g. BART) | required | required | optional | -Generation models (pure decoder or encoder/decoder architectures) are available in the `models` module. -ost pipelines are available for ONNX model checkpoints, including sequence classification, zero-shot classification, -token classification (including named entity recognition and part-of-speech tagging), question answering, text generation, summarization and translation. -These models use the same configuration and tokenizer files as their Pytorch counterparts when used in a pipeline. Examples leveraging ONNX models are given in the `./examples` directory +Note that the computational efficiency will drop when the `decoder with past` +file is optional but not provided since the model will not used cached past keys +and values for the attention mechanism, leading to a high number of redundant +computations. The Optimum library offers export options to ensure such a +`decoder with past` model file is created. he base encoder and decoder model +architecture are available (and exposed for convenience) in the `encoder` and +`decoder` modules, respectively. + +Generation models (pure decoder or encoder/decoder architectures) are available +in the `models` module. ost pipelines are available for ONNX model checkpoints, +including sequence classification, zero-shot classification, token +classification (including named entity recognition and part-of-speech tagging), +question answering, text generation, summarization and translation. These models +use the same configuration and tokenizer files as their Pytorch counterparts +when used in a pipeline. Examples leveraging ONNX models are given in the +`./examples` directory ## Ready-to-use pipelines - -Based on Hugging Face's pipelines, ready to use end-to-end NLP pipelines are available as part of this crate. The following capabilities are currently available: -**Disclaimer** -The contributors of this repository are not responsible for any generation from the 3rd party utilization of the pretrained systems proposed herein. +Based on Hugging Face's pipelines, ready to use end-to-end NLP pipelines are +available as part of this crate. The following capabilities are currently +available: + +**Disclaimer** The contributors of this repository are not responsible for any +generation from the 3rd party utilization of the pretrained systems proposed +herein.
1. Question Answering -Extractive question answering from a given question and context. DistilBERT model fine-tuned on SQuAD (Stanford Question Answering Dataset) +Extractive question answering from a given question and context. DistilBERT +model fine-tuned on SQuAD (Stanford Question Answering Dataset) ```rust let qa_model = QuestionAnsweringModel::new(Default::default())?; @@ -167,20 +236,27 @@ Extractive question answering from a given question and context. DistilBERT mode ``` Output: + ``` [Answer { score: 0.9976, start: 13, end: 21, answer: "Amsterdam" }] ``` +
-  + 
2. Translation -Translation pipeline supporting a broad range of source and target languages. Leverages two main architectures for translation tasks: -- Marian-based models, for specific source/target combinations -- M2M100 models allowing for direct translation between 100 languages (at a higher computational cost and lower performance for some selected languages) +Translation pipeline supporting a broad range of source and target languages. +Leverages two main architectures for translation tasks: + +- Marian-based models, for specific source/target combinations +- M2M100 models allowing for direct translation between 100 languages (at a + higher computational cost and lower performance for some selected languages) + +Marian-based pretrained models for the following language pairs are readily +available in the library - but the user can import any Pytorch-based model for +predictions -Marian-based pretrained models for the following language pairs are readily available in the library - but the user can import any Pytorch-based -model for predictions - English <-> French - English <-> Spanish - English <-> Portuguese @@ -196,30 +272,36 @@ model for predictions - English <-> Hindi - French <-> German -For languages not supported by the proposed pretrained Marian models, the user can leverage a M2M100 model supporting direct translation between 100 languages (without intermediate English translation) -The full list of supported languages is available in the [crate documentation](https://docs.rs/rust-bert/latest/rust_bert/pipelines/translation/enum.Language.html) +For languages not supported by the proposed pretrained Marian models, the user +can leverage a M2M100 model supporting direct translation between 100 languages +(without intermediate English translation) The full list of supported languages +is available in the +[crate documentation](https://docs.rs/rust-bert/latest/rust_bert/pipelines/translation/enum.Language.html) + +```rust +use rust_bert::pipelines::translation::{Language, TranslationModelBuilder}; +fn main() -> anyhow::Result<()> { +let model = TranslationModelBuilder::new() + .with_source_languages(vec![Language::English]) + .with_target_languages(vec![Language::Spanish, Language::French, Language::Italian]) + .create_model()?; + let input_text = "This is a sentence to be translated"; + let output = model.translate(&[input_text], None, Language::Spanish)?; + for sentence in output { + println!("{}", sentence); + } + Ok(()) +} +``` - ```rust - use rust_bert::pipelines::translation::{Language, TranslationModelBuilder}; - fn main() -> anyhow::Result<()> { - let model = TranslationModelBuilder::new() - .with_source_languages(vec![Language::English]) - .with_target_languages(vec![Language::Spanish, Language::French, Language::Italian]) - .create_model()?; - let input_text = "This is a sentence to be translated"; - let output = model.translate(&[input_text], None, Language::Spanish)?; - for sentence in output { - println!("{}", sentence); - } - Ok(()) - } - ``` Output: + ``` Il s'agit d'une phrase à traduire ``` +
-  + 
3. Summarization @@ -252,26 +334,35 @@ about exoplanets like K2-18b."]; let output = summarization_model.summarize(&input); ``` -(example from: [WikiNews](https://en.wikinews.org/wiki/Astronomers_find_water_vapour_in_atmosphere_of_exoplanet_K2-18b)) + +(example from: +[WikiNews](https://en.wikinews.org/wiki/Astronomers_find_water_vapour_in_atmosphere_of_exoplanet_K2-18b)) Output: + ``` "Scientists have found water vapour on K2-18b, a planet 110 light-years from Earth. This is the first such discovery in a planet in its star's habitable zone. The planet is not too hot and not too cold for liquid water to exist." ``` +
-  + 
4. Dialogue Model -Conversation model based on Microsoft's [DialoGPT](https://github.com/microsoft/DialoGPT). -This pipeline allows the generation of single or multi-turn conversations between a human and a model. +Conversation model based on Microsoft's +[DialoGPT](https://github.com/microsoft/DialoGPT). This pipeline allows the +generation of single or multi-turn conversations between a human and a model. The DialoGPT's page states that -> The human evaluation results indicate that the response generated from DialoGPT is comparable to human response quality -> under a single-turn conversation Turing test. ([DialoGPT repository](https://github.com/microsoft/DialoGPT)) -The model uses a `ConversationManager` to keep track of active conversations and generate responses to them. +> The human evaluation results indicate that the response generated from +> DialoGPT is comparable to human response quality under a single-turn +> conversation Turing test. +> ([DialoGPT repository](https://github.com/microsoft/DialoGPT)) + +The model uses a `ConversationManager` to keep track of active conversations and +generate responses to them. ```rust use rust_bert::pipelines::conversation::{ConversationModel, ConversationManager}; @@ -282,19 +373,24 @@ let mut conversation_manager = ConversationManager::new(); let conversation_id = conversation_manager.create("Going to the movies tonight - any suggestions?"); let output = conversation_model.generate_responses(&mut conversation_manager); ``` + Example output: + ``` "The Big Lebowski." ``` +
-  + 
5. Natural Language Generation Generate language based on a prompt. GPT2 and GPT available as base models. -Include techniques such as beam search, top-k and nucleus sampling, temperature setting and repetition penalty. -Supports batch generation of sentences from several prompts. Sequences will be left-padded with the model's padding token if present, the unknown token otherwise. -This may impact the results, it is recommended to submit prompts of similar length for best results +Include techniques such as beam search, top-k and nucleus sampling, temperature +setting and repetition penalty. Supports batch generation of sentences from +several prompts. Sequences will be left-padded with the model's padding token if +present, the unknown token otherwise. This may impact the results, it is +recommended to submit prompts of similar length for best results ```rust let model = GPT2Generator::new(Default::default())?; @@ -309,7 +405,9 @@ This may impact the results, it is recommended to submit prompts of similar leng let output = model.generate(Some(&[input_context_1, input_context_2]), generate_options); ``` + Example output: + ``` [ "The dog's owners, however, did not want to be named. According to the lawsuit, the animal's owner, a 29-year" @@ -320,12 +418,15 @@ Example output: "The cat was attacked by two stray dogs and was taken to a hospital. Two other cats were also injured in the attack and are being treated." ] ``` +
-  + 
6. Zero-shot classification -Performs zero-shot classification on input sentences with provided labels using a model fine-tuned for Natural Language Inference. +Performs zero-shot classification on input sentences with provided labels using +a model fine-tuned for Natural Language Inference. + ```rust let sequence_classification_model = ZeroShotClassificationModel::new(Default::default())?; @@ -342,18 +443,22 @@ Performs zero-shot classification on input sentences with provided labels using ``` Output: + ``` [ [ Label { "politics", score: 0.972 }, Label { "public health", score: 0.032 }, Label {"economics", score: 0.006 }, Label {"sports", score: 0.004 } ], [ Label { "politics", score: 0.975 }, Label { "public health", score: 0.0818 }, Label {"economics", score: 0.852 }, Label {"sports", score: 0.001 } ], ] ``` +
-  + 
7. Sentiment analysis -Predicts the binary sentiment for a sentence. DistilBERT model fine-tuned on SST-2. +Predicts the binary sentiment for a sentence. DistilBERT model fine-tuned on +SST-2. + ```rust let sentiment_classifier = SentimentModel::new(Default::default())?; @@ -365,9 +470,11 @@ Predicts the binary sentiment for a sentence. DistilBERT model fine-tuned on SST let output = sentiment_classifier.predict(&input); ``` + (Example courtesy of [IMDb](http://www.imdb.com)) Output: + ``` [ Sentiment { polarity: Positive, score: 0.9981985493795946 }, @@ -375,13 +482,17 @@ Output: Sentiment { polarity: Positive, score: 0.9997248985164333 } ] ``` +
-  + 
8. Named Entity Recognition -Extracts entities (Person, Location, Organization, Miscellaneous) from text. BERT cased large model fine-tuned on CoNNL03, contributed by the [MDZ Digital Library team at the Bavarian State Library](https://github.com/dbmdz). +Extracts entities (Person, Location, Organization, Miscellaneous) from text. +BERT cased large model fine-tuned on CoNNL03, contributed by the +[MDZ Digital Library team at the Bavarian State Library](https://github.com/dbmdz). Models are currently available for English, German, Spanish and Dutch. + ```rust let ner_model = NERModel::new(default::default())?; @@ -392,7 +503,9 @@ Models are currently available for English, German, Spanish and Dutch. let output = ner_model.predict(&input); ``` + Output: + ``` [ [ @@ -405,8 +518,9 @@ Output: ] ] ``` +
-  + 
9. Keywords/keyphrases extraction @@ -427,7 +541,9 @@ fn main() -> anyhow::Result<()> { let output = keyword_extraction_model.predict(&[input])?; } ``` + Output: + ``` "rust" - 0.50910604 "programming" - 0.35731024 @@ -435,12 +551,14 @@ Output: "concurrent" - 0.31229728 "program" - 0.29115444 ``` +
-  + 
10. Part of Speech tagging Extracts Part of Speech tags (Noun, Verb, Adjective...) from text. + ```rust let pos_model = POSModel::new(default::default())?; @@ -448,7 +566,9 @@ Extracts Part of Speech tags (Noun, Verb, Adjective...) from text. let output = pos_model.predict(&input); ``` + Output: + ``` [ Entity { word: "My", score: 0.1560, label: "PRP" } @@ -457,12 +577,15 @@ Output: Entity { word: "Bob", score: 0.7460, label: "NNP" } ] ``` +
-  + 
11. Sentence embeddings -Generate sentence embeddings (vector representation). These can be used for applications including dense information retrieval. +Generate sentence embeddings (vector representation). These can be used for +applications including dense information retrieval. + ```rust let model = SentenceEmbeddingsBuilder::remote( SentenceEmbeddingsModelType::AllMiniLmL12V2 @@ -475,19 +598,23 @@ Generate sentence embeddings (vector representation). These can be used for appl let output = model.encode(&sentences)?; ``` + Output: + ``` [ [-0.000202666, 0.08148022, 0.03136178, 0.002920636 ...], [0.064757116, 0.048519745, -0.01786038, -0.0479775 ...] ] ``` +
-  + 
12. Masked Language Model Predict masked words in input sentences. + ```rust let model = MaskedLanguageModel::new(Default::default())?; @@ -498,7 +625,9 @@ Predict masked words in input sentences. let output = model.predict(&sentences); ``` + Output: + ``` [ [MaskedToken { text: "college", id: 2267, score: 8.091}], @@ -508,29 +637,61 @@ Output: ] ] ``` +
## Benchmarks -For simple pipelines (sequence classification, tokens classification, question answering) the performance between Python and Rust is expected to be comparable. This is because the most expensive part of these pipeline is the language model itself, sharing a common implementation in the Torch backend. The [End-to-end NLP Pipelines in Rust](https://www.aclweb.org/anthology/2020.nlposs-1.4/) provides a benchmarks section covering all pipelines. +For simple pipelines (sequence classification, tokens classification, question +answering) the performance between Python and Rust is expected to be comparable. +This is because the most expensive part of these pipeline is the language model +itself, sharing a common implementation in the Torch backend. The +[End-to-end NLP Pipelines in Rust](https://www.aclweb.org/anthology/2020.nlposs-1.4/) +provides a benchmarks section covering all pipelines. -For text generation tasks (summarization, translation, conversation, free text generation), significant benefits can be expected (up to 2 to 4 times faster processing depending on the input and application). The article [Accelerating text generation with Rust](https://guillaume-be.github.io/2020-11-21/generation_benchmarks) focuses on these text generation applications and provides more details on the performance comparison to Python. +For text generation tasks (summarization, translation, conversation, free text +generation), significant benefits can be expected (up to 2 to 4 times faster +processing depending on the input and application). The article +[Accelerating text generation with Rust](https://guillaume-be.github.io/2020-11-21/generation_benchmarks) +focuses on these text generation applications and provides more details on the +performance comparison to Python. ## Loading pretrained and custom model weights -The base model and task-specific heads are also available for users looking to expose their own transformer based models. -Examples on how to prepare the date using a native tokenizers Rust library are available in `./examples` for BERT, DistilBERT, RoBERTa, GPT, GPT2 and BART. -Note that when importing models from Pytorch, the convention for parameters naming needs to be aligned with the Rust schema. Loading of the pre-trained weights will fail if any of the model parameters weights cannot be found in the weight files. -If this quality check is to be skipped, an alternative method `load_partial` can be invoked from the variables store. +The base model and task-specific heads are also available for users looking to +expose their own transformer based models. Examples on how to prepare the date +using a native tokenizers Rust library are available in `./examples` for BERT, +DistilBERT, RoBERTa, GPT, GPT2 and BART. Note that when importing models from +Pytorch, the convention for parameters naming needs to be aligned with the Rust +schema. Loading of the pre-trained weights will fail if any of the model +parameters weights cannot be found in the weight files. If this quality check is +to be skipped, an alternative method `load_partial` can be invoked from the +variables store. -Pretrained models are available on Hugging face's [model hub](https://huggingface.co/models?filter=rust) and can be loaded using `RemoteResources` defined in this library. -A conversion utility script is included in `./utils` to convert Pytorch weights to a set of weights compatible with this library. This script requires Python and `torch` to be set-up, and can be used as follows: -`python ./utils/convert_model.py path/to/pytorch_model.bin` where `path/to/pytorch_model.bin` is the location of the original Pytorch weights. +Pretrained models are available on Hugging face's +[model hub](https://huggingface.co/models?filter=rust) and can be loaded using +`RemoteResources` defined in this library. +A conversion utility script is included in `./utils` to convert Pytorch weights +to a set of weights compatible with this library. This script requires Python +and `torch` to be set-up, and can be used as follows: +`python ./utils/convert_model.py path/to/pytorch_model.bin` where +`path/to/pytorch_model.bin` is the location of the original Pytorch weights. + +```bash +python3 -m venv .venv +source .venv/bin/activate + +pip install -r requirements.txt + +python utils/convert_model.py path/to/pytorch_model.bin +``` ## Citation -If you use `rust-bert` for your work, please cite [End-to-end NLP Pipelines in Rust](https://www.aclweb.org/anthology/2020.nlposs-1.4/): +If you use `rust-bert` for your work, please cite +[End-to-end NLP Pipelines in Rust](https://www.aclweb.org/anthology/2020.nlposs-1.4/): + ```bibtex @inproceedings{becquin-2020-end, title = "End-to-end {NLP} Pipelines in Rust", @@ -545,6 +706,7 @@ If you use `rust-bert` for your work, please cite [End-to-end NLP Pipelines in R ## Acknowledgements -Thank you to [Hugging Face](https://huggingface.co) for hosting a set of weights compatible with this Rust library. -The list of ready-to-use pretrained models is listed at [https://huggingface.co/models?filter=rust](https://huggingface.co/models?filter=rust). - +Thank you to [Hugging Face](https://huggingface.co) for hosting a set of weights +compatible with this Rust library. The list of ready-to-use pretrained models is +listed at +[https://huggingface.co/models?filter=rust](https://huggingface.co/models?filter=rust). diff --git a/requirements.txt b/requirements.txt deleted file mode 100644 index efd6030..0000000 --- a/requirements.txt +++ /dev/null @@ -1,3 +0,0 @@ -torch == 1.13.1 -requests == 2.32.0 -numpy == 1.23.4 \ No newline at end of file diff --git a/src/models/bart/decoder.rs b/src/models/bart/decoder.rs index 82efc61..7b86189 100644 --- a/src/models/bart/decoder.rs +++ b/src/models/bart/decoder.rs @@ -340,6 +340,7 @@ impl BartDecoder { } } +#[allow(dead_code)] ///Container holding a BART decoder output pub struct BartDecoderOutput { /// last decoder layer hidden state diff --git a/src/models/prophetnet/encoder.rs b/src/models/prophetnet/encoder.rs index 4152e5c..a1fa2a5 100644 --- a/src/models/prophetnet/encoder.rs +++ b/src/models/prophetnet/encoder.rs @@ -210,6 +210,7 @@ impl ProphetNetEncoder { } } +#[allow(dead_code)] /// Container for the ProphetNet encoder output. pub struct ProphetNetEncoderOutput { /// Last hidden states from the model diff --git a/src/models/reformer/encoder.rs b/src/models/reformer/encoder.rs index 75cd7f1..6e5030b 100644 --- a/src/models/reformer/encoder.rs +++ b/src/models/reformer/encoder.rs @@ -143,6 +143,7 @@ impl ChunkReformerFeedForward { } } +#[allow(dead_code)] pub struct ReformerLayerOutput { pub attention_output: Tensor, pub hidden_states: Tensor, diff --git a/src/models/reformer/reformer_model.rs b/src/models/reformer/reformer_model.rs index 4d4f529..4958a52 100644 --- a/src/models/reformer/reformer_model.rs +++ b/src/models/reformer/reformer_model.rs @@ -207,6 +207,7 @@ impl ReformerLMHead { } } +#[allow(dead_code)] pub struct PaddedReformerInput { pub input_ids: Option, pub input_embeds: Option, diff --git a/src/models/t5/encoder.rs b/src/models/t5/encoder.rs index 74a2c4d..404e4a6 100644 --- a/src/models/t5/encoder.rs +++ b/src/models/t5/encoder.rs @@ -541,6 +541,7 @@ impl T5Stack { } } +#[allow(dead_code)] pub struct T5BlockOutput { pub hidden_states: Tensor, pub self_attention_weights: Option, diff --git a/utils/convert_model.py b/utils/convert_model.py index b5e5f08..e4337c6 100644 --- a/utils/convert_model.py +++ b/utils/convert_model.py @@ -50,13 +50,20 @@ import sys import zipfile from pathlib import Path from typing import Dict - +import os import numpy as np import torch from numpy.lib.format import write_array -from numpy.lib.npyio import zipfile_factory +# from numpy.lib.npyio import zipfile_factory from torch import Tensor +def zipfile_factory(file, *args, **kwargs): + if not hasattr(file, 'read'): + file = os.fspath(file) + import zipfile + kwargs['allowZip64'] = True + kwargs['compresslevel'] = 4 + return zipfile.ZipFile(file, *args, **kwargs) def get_bf16_repr(input_tensor: torch.Tensor) -> np.ndarray: """Convert a bfloat16 tensor to an equivalent byte representation in Numpy. @@ -125,6 +132,12 @@ if __name__ == "__main__": help="Use this flag to enable automatic download of the libtorch library.", ) args = parser.parse_args() + + logger = logging.getLogger('convert_model') + logger.setLevel(logging.DEBUG) + fh = logging.FileHandler('convert_model.log') + fh.setLevel(logging.DEBUG) + logger.addHandler(fh) target_folder = Path(args.source_file[0]).parent with zipfile_factory( @@ -133,7 +146,7 @@ if __name__ == "__main__": for source_file_or_pattern in args.source_file: source_files = glob.glob(source_file_or_pattern) for source_file in source_files: - logging.info(f"Processing source file {source_file}...") + logger.info(f"Processing source file {source_file}") nps = {} source_file = Path(source_file) weights = torch.load(str(source_file), map_location="cpu") @@ -168,11 +181,11 @@ if __name__ == "__main__": ) else: nps[k] = np.ascontiguousarray(tensor) - logging.info( + logger.info( f"converted {k} - {str(sys.getsizeof(nps[k]))} bytes" ) else: - logging.info(f"skipped non-tensor object: {k}") + logger.info(f"skipped non-tensor object: {k}") append_to_zipf(nps, output_zipfile) source = str(target_folder / "model.npz") diff --git a/utils/requirements.txt b/utils/requirements.txt new file mode 100644 index 0000000..1ca9bcd --- /dev/null +++ b/utils/requirements.txt @@ -0,0 +1,11 @@ +filelock==3.15.3 +fsspec==2024.6.0 +Jinja2==3.1.4 +MarkupSafe==2.1.5 +mpmath==1.3.0 +networkx==3.3 +numpy==2.0.0 +sympy==1.12.1 +torch==2.3.1 +typing_extensions==4.12.2 +requests==2.32.0 \ No newline at end of file