Published May 14, 2020
| Version v2.9.1
Software
Open
huggingface/transformers: Marian
Creators
- Thomas Wolf1
- Lysandre Debut2
- Julien Chaumond2
- Victor SANH1
- Patrick von Platen
- Aymeric Augustin3
- Funtowicz Morgan4
- Rémi Louf
- Sam Shleifer5
- Stefan Schweter
- Manuel Romero
- Denis
- erenup
- Matt
- Piero Molino
- Grégory Châtel6
- Bram Vanroy7
- Tim Rault1
- Gunnlaugur Thor Briem8
- Anthony MOI2
- Malte Pietsch9
- Catalin Voss10
- Bilal Khan
- Fei Wang11
- Louis Martin
- Davide Fiocco
- Martin Malmsten
- Lorenzo Ampil12
- HUSEIN ZOLKEPLI
- Clement1
- 1. @huggingface
- 2. Hugging Face
- 3. @canalplus
- 4. HuggingFace
- 5. Huggingface
- 6. DisAItek & Intel AI Innovators
- 7. @UGent
- 8. Qlik
- 9. deepset
- 10. Stanford University
- 11. University of Southern California
- 12. @thinkingmachines
Description
Marian (@sshleifer)
- A new model architecture,
MarianMTModel
with 1,008+ pretrained weights is available for machine translation in PyTorch. - The corresponding
MarianTokenizer
uses aprepare_translation_batch
method to prepare model inputs. - All pretrained model names use the following format:
Helsinki-NLP/opus-mt-{src}-{tgt}
- See docs for information on pretrained model discovery and naming, or find your language here
A new model architecture has been added: AlbertForPreTraining
in both PyTorch and TensorFlow
Changes have been made to both the TensorFlow scripts and our internals so that we are compatible with TensorFlow 2.2
TFTrainer now supports new tasks- Multiple choice has been added to the TFTrainer (@ViktorAlm)
- Question Answering has been added to the TFTrainer (@jplu)
- Fixed a bug with the tf generation pipeline (@patrickvonplaten)
- Fixed the XLA spawn (@julien-c)
- The sentiment analysis pipeline tokenizer was cased while the model was uncased (@mfuntowicz)
- Albert was added to the conversion CLI (@fgaim)
- CamemBERT's token ID generation from tokenizer were removed like RoBERTa, as the model does not use them (@LysandreJik)
- Additional migration documentation was added (@guoquan)
- GPT-2 can now be exported to ONNX (@tianleiwu)
- Simplify cache vars and allow for TRANSFORMERS_CACHE env (@BramVanroy)
- Remove hard-coded pad token id in distilbert and albert (@monologg)
- BART tests were fixed on GPU (@julien-c)
- Better wandb integration (@vanpelt, @borisdayma, @julien-c)
Files
huggingface/transformers-v2.9.1.zip
Files
(3.7 MB)
Name | Size | Download all |
---|---|---|
md5:93e3080e5e9d50be32106758e15d9b99
|
3.7 MB | Preview Download |
Additional details
Related works
- Is supplement to
- https://github.com/huggingface/transformers/tree/v2.9.1 (URL)