There is a newer version of the record available.

Published December 20, 2019 | Version v2.3.0
Software Open

huggingface/transformers: Downstream NLP task API (feature extraction, text classification, NER, QA), Command-Line Interface and Serving – models: T5 – community-added models: Japanese & Finnish BERT, PPLM, XLM-R

Description

New class Pipeline (beta): easily run and use models on down-stream NLP tasks

We have added a new class called Pipeline to simply run and use models for several down-stream NLP tasks.

A Pipeline is just a tokenizer + model wrapped so they can take human-readable inputs and output human-readable results.

The Pipeline will take care of : tokenizing inputs strings => convert in tensors => run in the model => post-process output

Currently, we have added the following pipelines with a default model for each:

  • feature extraction (can be used with any pretrained and finetuned models) inputs: strings/list of strings – output: list of floats (last hidden-states of the model for each token)
  • sentiment classification (DistilBert model fine-tuned on SST-2) inputs: strings/list of strings – output: list of dict with label/score of the top class
  • Named Entity Recognition (XLM-R finetuned on CoNLL2003 by the awesome @stefan-it), and inputs: strings/list of strings – output: list of dict with label/entities/position of the named-entities
  • Question Answering (Bert Large whole-word version fine-tuned on SQuAD 1.0) inputs: dict of strings/list of dict of strings – output: list of dict with text/position of the answers

There are three ways to use pipelines:

  • in python: ```python from transformers import pipeline
Test the default model for QA (Bert large finetuned on SQuAD 1.0)

nlp = pipeline('question-answering') nlp(question= "Where does Amy live ?", context="Amy lives in Amsterdam.")

{'answer': 'Amsterdam', 'score': 0.9657156007786263, 'start': 13, 'end': 21}

Test a specific model for NER (XLM-R finetuned by @stefan-it on CoNLL03 English)

nlp = pipeline('ner', model='xlm-roberta-large-finetuned-conll03-english') nlp("My name is Amy. I live in Paris.")

[{'word': 'Amy', 'score': 0.9999586939811707, 'entity': 'I-PER'}, {'word': 'Paris', 'score': 0.9999983310699463, 'entity': 'I-LOC'}]

- in bash (using the command-line interface)
```bash
bash $ echo -e "Where does Amy live?\tAmy lives in Amsterdam" | transformers-cli run --task question-answering
{'score': 0.9657156007786263, 'start': 13, 'end': 22, 'answer': 'Amsterdam'}
  • as a REST API
    transformers-cli serve --task question-answering
    

This new feature is currently in beta and will evolve in the coming weeks.

CLI tool to upload community models @julien-c

Users can now create accounts on the huggingface.co website and then login using the transformers CLI. Doing so allows users to upload their models to our S3 in their respective directories, so that other users may download said models and use them in their tasks.

Users may upload files or directories.

It's been tested by @stefan-it for a German BERT and by @singletongue for a Japanese BERT.

New model architectures: T5, Japanese BERT, PPLM, XLM-RoBERTa, Finnish BERT Refactoring the SQuAD example

The run_squad script has been massively refactored. The reasons are the following:

  • it was made to work with only a few models (BERT, XLNet, XLM and DistilBERT), which had three different ways of encoding sequences. The script had to be individually modified in order to train different models, which would not scale as other models are added to the library.
  • the utilities did not rely on the QOL adjustments that were made to the encoding methods these past months.

It now leverages the full capacity of encode_plus, easing the addition of new models to the script. A new method squad_convert_examples_to_features encapsulates all of the tokenization. This method can handle tensorflow_datasets as well as squad v1 json files and squad v2 json files.

  • ALBERT was added to the SQuAD script
BertAbs summarization

A contribution by @rlouf building on the encoder-decoder mechanism to do abstractive summarization.

  • Utilities to load the CNN/DailyMail dataset
  • BertAbs now usable as a traditional library model (using from_pretrained())
  • ROUGE evaluation
New Models Additional architectures

@alexzubiaga added XLNetForTokenClassification and TFXLNetForTokenClassification

New model cards Community additions/bug-fixes/improvements
  • Added mish activation function @digantamisra98
  • run_bertology.py was updated with correct imports and the ability to overwrite the cache
  • Training can be exited and relaunched safely, while keeping the epochs, global steps, scheduler steps and other variables in run_lm_finetuning.py @bkkaggle
  • Tests now run on cuda @aaugustin @julien-c
  • Cleaned up the pytorch to tf conversion script @thomwolf
  • Progress indicator improvements when downloading pre-trained models @leopd
  • from_pretrained() can now load from urls directly.
  • New tests to check that all files are accessible on HuggingFace's S3 @rlouf
  • Updated tf.shape and tensor.shape to all use shape_list @thomwolf
  • Valohai integration @thomwolf
  • Always use SequentialSampler in run_squad.py @ethanjperez
  • Stop using GPU when importing transformers @ondewo
  • Fixed the XLNet attention output @roskoN
  • Several QOL adjustments: removed dead code, deep cleaned tests and removed pytest dependency @aaugustin
  • Fixed an issue with the Camembert tokenization @thomwolf
  • Correctly create an encoder attention mask from the shape of the hidden states @rlouf
  • Fixed a non-deterministic behavior when encoding and decoding empty strings @pglock
  • Fixing tensor creation in encode_plus @LysandreJik
  • Remove usage of tf.mean which does not exist in TF2 @LysandreJik
  • A segmentation fault error was fixed (due to scipy 1.4.0) @LysandreJik
  • Start sunsetting support of Python 2
  • An example usage of Model2Model was added to the quickstart.

Files

huggingface/transformers-v2.3.0.zip

Files (2.9 MB)

Name Size Download all
md5:ad69ac9eebca54c80b4dd6f34c6a52df
2.9 MB Preview Download

Additional details