Dataset Open Access

Dataset for generating TL;DR

Syed, Shahbaz; Voelske, Michael; Potthast, Martin; Stein, Benno


MARC21 XML Export

<?xml version='1.0' encoding='UTF-8'?>
<record xmlns="http://www.loc.gov/MARC21/slim">
  <leader>00000nmm##2200000uu#4500</leader>
  <datafield tag="041" ind1=" " ind2=" ">
    <subfield code="a">eng</subfield>
  </datafield>
  <datafield tag="653" ind1=" " ind2=" ">
    <subfield code="a">tl;dr challenge</subfield>
  </datafield>
  <datafield tag="653" ind1=" " ind2=" ">
    <subfield code="a">abstractive summarization</subfield>
  </datafield>
  <datafield tag="653" ind1=" " ind2=" ">
    <subfield code="a">social media</subfield>
  </datafield>
  <datafield tag="653" ind1=" " ind2=" ">
    <subfield code="a">user-generated content</subfield>
  </datafield>
  <controlfield tag="005">20190409133313.0</controlfield>
  <controlfield tag="001">1168855</controlfield>
  <datafield tag="700" ind1=" " ind2=" ">
    <subfield code="u">Bauhaus-Universität Weimar</subfield>
    <subfield code="a">Voelske, Michael</subfield>
  </datafield>
  <datafield tag="700" ind1=" " ind2=" ">
    <subfield code="u">Bauhaus-Universität Weimar</subfield>
    <subfield code="a">Potthast, Martin</subfield>
  </datafield>
  <datafield tag="700" ind1=" " ind2=" ">
    <subfield code="u">Bauhaus-Universität Weimar</subfield>
    <subfield code="a">Stein, Benno</subfield>
  </datafield>
  <datafield tag="856" ind1="4" ind2=" ">
    <subfield code="s">2157862847</subfield>
    <subfield code="z">md5:28951b6f3d5c6fd6f97e1f6314be3661</subfield>
    <subfield code="u">https://zenodo.org/record/1168855/files/tldr-challenge-dataset.zip</subfield>
  </datafield>
  <datafield tag="542" ind1=" " ind2=" ">
    <subfield code="l">open</subfield>
  </datafield>
  <datafield tag="260" ind1=" " ind2=" ">
    <subfield code="c">2018-02-08</subfield>
  </datafield>
  <datafield tag="909" ind1="C" ind2="O">
    <subfield code="p">openaire_data</subfield>
    <subfield code="o">oai:zenodo.org:1168855</subfield>
  </datafield>
  <datafield tag="100" ind1=" " ind2=" ">
    <subfield code="u">Bauhaus-Universität Weimar</subfield>
    <subfield code="a">Syed, Shahbaz</subfield>
  </datafield>
  <datafield tag="245" ind1=" " ind2=" ">
    <subfield code="a">Dataset for generating TL;DR</subfield>
  </datafield>
  <datafield tag="540" ind1=" " ind2=" ">
    <subfield code="u">http://creativecommons.org/licenses/by/4.0/legalcode</subfield>
    <subfield code="a">Creative Commons Attribution 4.0 International</subfield>
  </datafield>
  <datafield tag="650" ind1="1" ind2="7">
    <subfield code="a">cc-by</subfield>
    <subfield code="2">opendefinition.org</subfield>
  </datafield>
  <datafield tag="520" ind1=" " ind2=" ">
    <subfield code="a">&lt;p&gt;This is the dataset for the TL;DR challenge containing posts&amp;nbsp;from the Reddit corpus, suitable for abstractive summarization using deep learning. The format is a json file where each line is a JSON object representing a post. The schema of each post is shown below:&lt;/p&gt;

&lt;ul&gt;
	&lt;li&gt;author: string (nullable = true)&lt;/li&gt;
	&lt;li&gt;body: string (nullable = true)&lt;/li&gt;
	&lt;li&gt;normalizedBody: string (nullable = true)&lt;/li&gt;
	&lt;li&gt;content: string (nullable = true)&lt;/li&gt;
	&lt;li&gt;content_len: long (nullable = true)&lt;/li&gt;
	&lt;li&gt;summary: string (nullable = true)&lt;/li&gt;
	&lt;li&gt;summary_len: long (nullable = true)&lt;/li&gt;
	&lt;li&gt;id: string (nullable = true)&lt;/li&gt;
	&lt;li&gt;subreddit: string (nullable = true)&lt;/li&gt;
	&lt;li&gt;subreddit_id: string (nullable = true)&lt;/li&gt;
	&lt;li&gt;title: string (nullable = true)&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Specifically, the &lt;strong&gt;content&lt;/strong&gt; and &lt;strong&gt;summary&lt;/strong&gt; fields can be directly used as inputs to a deep learning model (e.g. Sequence to Sequence model ). The dataset consists of 3,084,410 posts with an average length of 211 words for content, and 25&amp;nbsp;words for the summary.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Note :&amp;nbsp;&lt;/strong&gt;As this is the complete dataset for the challenge, it is up to the participants to split it into training and validation sets accordingly.&lt;/p&gt;</subfield>
  </datafield>
  <datafield tag="773" ind1=" " ind2=" ">
    <subfield code="n">doi</subfield>
    <subfield code="i">cites</subfield>
    <subfield code="a">10.5281/zenodo.1043504</subfield>
  </datafield>
  <datafield tag="773" ind1=" " ind2=" ">
    <subfield code="n">doi</subfield>
    <subfield code="i">isVersionOf</subfield>
    <subfield code="a">10.5281/zenodo.1168854</subfield>
  </datafield>
  <datafield tag="024" ind1=" " ind2=" ">
    <subfield code="a">10.5281/zenodo.1168855</subfield>
    <subfield code="2">doi</subfield>
  </datafield>
  <datafield tag="980" ind1=" " ind2=" ">
    <subfield code="a">dataset</subfield>
  </datafield>
</record>
646
500
views
downloads
All versions This version
Views 646646
Downloads 500500
Data volume 1.1 TB1.1 TB
Unique views 592592
Unique downloads 411411

Share

Cite as