Dataset Restricted Access

Kestemont, Mike; Tschuggnall, Michael; Stamatatos, Efstathios; Daelemans, Walter; Specht, Günther; Stein, Benno; Potthast, Martin

We provide a corpus which comprises a set of cross-domain authorship attribution problems in each of the following 5 languages: English, French, Italian, Polish, and Spanish. Note that we specifically avoid to use the term 'training corpus' because the sets of candidate authors of the development and the evaluation corpora are not overlapping. Therefore, your approach should not be designed to particularly handle the candidate authors of the development corpus.

Each problem consists of a set of known fanfics by each candidate author and a set of unknown fanfics located in separate folders. The file problem-info.json that can be found in the main folder of each problem, shows the name of folder of unknown documents and the list of names of candidate author folders.

The true author of each unknown document can be seen in the file ground-truth.json, also found in the main folder of each problem.

In addition, to handle a collection of such problems, the file collection-info.jsonincludes all relevant information. In more detail, for each problem it lists its main folder, the language (either "en", "fr", "it", "pl", or "sp") and encoding (always UTF-8) of its documents.

new version: removed passwords inside packages
##### Restricted Access

You may request access to the files in this upload, provided that you fulfil the conditions below. The decision whether to grant/deny access is solely under the responsibility of the record owner.

Please request access to the data with a short statement on how you want to use it. Thanks!
We would like to point out that you can register on pan.webis.de to be part of the PAN community.

• Mike Kestemont, Michael Tschuggnall, Efstathios Stamatatos, Walter Daelemans, Günther Specht, Benno Stein, and Martin Potthast. Overview of the Author Identification Task at PAN-2018: Cross-domain Authorship Attribution and Style Change Detection. In Linda Cappellato, Nicola Ferro, Jian-Yun Nie, and Laure Soulier, editors, Working Notes Papers of the CLEF 2018 Evaluation Labs volume 2125 of CEUR Workshop Proceedings, September 2018. CEUR-WS.org. ISSN 1613-0073.

177
10
views