Conference paper Open Access

Tri Automatique de la Littérature pour les Revues Systématiques

Norman, Christopher; Leeflang, Mariska; Zweigenbaum, Pierre; Névéol, Aurélie

Dublin Core Export

<?xml version='1.0' encoding='utf-8'?>
<oai_dc:dc xmlns:dc="" xmlns:oai_dc="" xmlns:xsi="" xsi:schemaLocation="">
  <dc:creator>Norman, Christopher</dc:creator>
  <dc:creator>Leeflang, Mariska</dc:creator>
  <dc:creator>Zweigenbaum, Pierre</dc:creator>
  <dc:creator>Névéol, Aurélie</dc:creator>
  <dc:description>Current approaches to document discovery for systematic reviews in biomedicine rely on exhaustive manual screening. We evaluate the performance of classifier based article discovery using different definitions of inclusion criteria. We test a logistic regressor on two datasets created from existing systematic reviews on clinical NLP and drug efficacy, using different criteria to generate positive and negative examples. The classification and ranking achieves an average AUC of 0.769 when relying on gold standard decisions based on title and abstracts of articles, and an AUC of 0.835 when relying on decisions based on full text. Results suggest that inclusion based on title and abstract generalizes to inclusion based on full text, so that references excluded in earlier stages are important for classification, and that common-off-the-shelves algorithms can partially automate the process.</dc:description>
  <dc:subject>Information Retrieval</dc:subject>
  <dc:subject>Supervised Classification</dc:subject>
  <dc:subject>Systematic Reviews</dc:subject>
  <dc:title>Tri Automatique de la Littérature pour les Revues Systématiques</dc:title>
All versions This version
Views 5151
Downloads 3131
Data volume 12.6 MB12.6 MB
Unique views 4343
Unique downloads 3131


Cite as