Dataset Open Access

# A meta analysis of Wikipedia's coronavirus sources during the COVID-19 pandemic

Sobel, Jonathan; Benjakob, Omer; Aviram, Rona

### DataCite XML Export

<?xml version='1.0' encoding='utf-8'?>
<identifier identifierType="DOI">10.5281/zenodo.3901741</identifier>
<creators>
<creator>
<creatorName>Sobel, Jonathan</creatorName>
<givenName>Jonathan</givenName>
<familyName>Sobel</familyName>
<nameIdentifier nameIdentifierScheme="ORCID" schemeURI="http://orcid.org/">0000-0002-5111-4070</nameIdentifier>
<affiliation>Weizmann Institute of Science, Rehovot, Israel</affiliation>
</creator>
<creator>
<creatorName>Benjakob, Omer</creatorName>
<givenName>Omer</givenName>
<familyName>Benjakob</familyName>
<nameIdentifier nameIdentifierScheme="ORCID" schemeURI="http://orcid.org/">0000-0002-7179-3509</nameIdentifier>
<affiliation>The Cohn Institute for the History and Philosophy of Science and Ideas, Tel Aviv University, Tel Aviv, Israel</affiliation>
</creator>
<creator>
<creatorName>Aviram, Rona</creatorName>
<givenName>Rona</givenName>
<familyName>Aviram</familyName>
<nameIdentifier nameIdentifierScheme="ORCID" schemeURI="http://orcid.org/">0000-0001-5544-3552</nameIdentifier>
<affiliation>Weizmann Institute of Science</affiliation>
</creator>
</creators>
<titles>
<title>A meta analysis of Wikipedia's coronavirus sources during the COVID-19 pandemic</title>
</titles>
<publisher>Zenodo</publisher>
<publicationYear>2021</publicationYear>
<subjects>
<subject>Wikipedia</subject>
<subject>COVID-19</subject>
<subject>Sources history</subject>
<subject>Infodemics</subject>
</subjects>
<dates>
<date dateType="Issued">2021-03-15</date>
</dates>
<language>en</language>
<resourceType resourceTypeGeneral="Dataset"/>
<alternateIdentifiers>
<alternateIdentifier alternateIdentifierType="url">https://zenodo.org/record/3901741</alternateIdentifier>
</alternateIdentifiers>
<relatedIdentifiers>
<relatedIdentifier relatedIdentifierType="DOI" relationType="IsVersionOf">10.5281/zenodo.3901740</relatedIdentifier>
<relatedIdentifier relatedIdentifierType="URL" relationType="IsPartOf">https://zenodo.org/communities/covid-19</relatedIdentifier>
<relatedIdentifier relatedIdentifierType="URL" relationType="IsPartOf">https://zenodo.org/communities/wikimedia</relatedIdentifier>
<relatedIdentifier relatedIdentifierType="URL" relationType="IsPartOf">https://zenodo.org/communities/wikipedia_data</relatedIdentifier>
</relatedIdentifiers>
<version>0.1</version>
<rightsList>
<rights rightsURI="info:eu-repo/semantics/openAccess">Open Access</rights>
</rightsList>
<descriptions>
<description descriptionType="Abstract">&lt;p&gt;At the height of the coronavirus pandemic, on the last day of March 2020, Wikipedia in all languages broke a record for most traffic in a single day. Since the breakout of the Covid-19 pandemic at the start of January, tens if not hundreds of millions of people have come to Wikipedia to read - and in some cases also contribute - knowledge, information and data about the virus to an ever-growing pool of articles. Our study focuses on the scientific backbone behind the content people across the world read: which sources informed Wikipedia&amp;rsquo;s coronavirus content, and how was the scientific research on this field represented on Wikipedia. Using citation as readout we try to map how COVID-19 related research was used in Wikipedia and analyse what happened to it before and during the pandemic. Understanding how scientific and medical information was integrated into Wikipedia, and what were the different sources that informed the Covid-19 content, is key to understanding the digital knowledge echosphere during the pandemic.&amp;nbsp;&lt;/p&gt;

&lt;p&gt;To delimitate the corpus of Wikipedia articles containing Digital Object Identifier (DOI), we applied two different strategies. First we scraped every Wikipedia pages form the COVID-19 Wikipedia project (about 3000 pages) and we filtered them to keep only page containing DOI citations. For our second strategy, we made a search with EuroPMC on Covid-19, SARS-CoV2, SARS-nCoV19 (30&amp;rsquo;000 sci papers, reviews and preprints) and a selection on scientific papers form 2019 onwards that we compared to the Wikipedia extracted citations from the english Wikipedia dump of &lt;strong&gt;May 2020&lt;/strong&gt; (2&amp;rsquo;000&amp;rsquo;000 DOIs). This search led to 231 Wikipedia articles containing at least one citation of the EuroPMC search or part of the wikipedia COVID-19 project pages containing DOIs. Next, from our 231 Wikipedia articles corpus we extracted DOIs, PMIDs, ISBNs, websites and URLs using a set of regular expressions. Subsequently, we computed several statistics for each wikipedia article&amp;nbsp; and we retrive Atmetics, CrossRef and EuroPMC infromations for each DOI. Finally, our method allowed to produce tables of citations annotated and extracted infromations in each wikipadia articles such as books, websites, newspapers.&lt;/p&gt;

&lt;p&gt;Files used as input and extracted information on Wikipedia&amp;#39;s COVID-19 sources are presented in this archive.&lt;/p&gt;

&lt;p&gt;See the &lt;a href="https://github.com/jsobel1/WikiCitationHistoRy"&gt;WikiCitationHistoRy&lt;/a&gt; Github repository for the R codes, and other bash/python scripts utilities related to this project.&lt;/p&gt;</description>
<description descriptionType="Other">Analysis on Wikipedia sources during the first wave of the COVID-19 pandemics (up to May 2020)</description>
</descriptions>
</resource>

809
731
views