Planned intervention: On Wednesday April 3rd 05:30 UTC Zenodo will be unavailable for up to 2-10 minutes to perform a storage cluster upgrade.
Published July 24, 2021 | Version v1
Journal article Open

A Replication Package For The Paper "Improved Retrieval of Programming Solutions with Code Examples Using a Multi-featured Score"

  • 1. Faculty of Computing, Federal University of Uberlândia
  • 2. Faculty of Computer Science, Dalhousie University
  • 3. Department of Computer Science, University of Saskatchewan, Canada
  • 4. École Polytechnique de Montréal, Canada

Description

Developers often depend on code search engines to obtain solutions for their programming tasks. However, finding an expected solution containing code examples along with their explanations is challenging due to several issues. There is a vocabulary mismatch between the search keywords (the query) and the appropriate solutions. Semantic gap may increase for similar bag of words due to antonyms and negation. Moreover, documents retrieved by search engines might not contain solutions containing both code examples and their explanations. So, we propose CRAR (Crowd Answer Recommender) to circumvent those issues aiming at improving retrieval of relevant answers from Stack Overflow containing not only the expected code examples for the given task but also their explanations. Given a programming task, we investigate the effectiveness of  combining information retrieval techniques along with a set of features to enhance the ranking of important threads (i.e., the units containing questions along with their answers) for the given task and then selects relevant answers contained in those threads, including semantic features, like word embeddings and sentence embeddings, for instance, a Convolutional Neural Network (CNN). CRAR also leverages social aspects of Stack Overflow discussions like popularity to select relevant answers for the tasks. Our experimental evaluation shows that the combination of the different features performs better than each one individually. We also compare the retrieval performance with the state-of-art CROKAGE (Crowd Knowledge Answer Generator), which is also a system aimed at retrieving relevant answers from Stack Overflow. We show that CRAR outperforms CROKAGE  in Mean Reciprocal Rank and Mean Recall with small and medium effect sizes, respectively.

Files

data.zip

Files (11.8 GB)

Name Size Download all
md5:38045fe56bbb37d3ea0898fec4ea5462
11.1 GB Download
md5:ecdf4c1cec9b721bfbe51e170b3aa4d3
694.7 MB Download
md5:3c249fd0045fc775ef28104f547643c6
920.4 kB Preview Download
md5:63fe6529a8bb7a98a9624e4ac08d29fe
9.7 kB Download
md5:ecdee892bf63fd47f3139416788269e9
2.9 kB Preview Download

Additional details

Related works