Published March 31, 2026 | Version v1
Other Open

Replication Package for the Paper ``Can Large Language Models Decompose User Stories into Tasks? Exploring the Role of Prompting Strategies and Models''

Authors/Creators

Description

DELTA 

This repository contains supplementary materials for the research paper: "Can Large Language Models Decompose User Stories into Tasks? Exploring the Role of Prompting Strategies and Models". The paper was submitted to the Automated Software Engineering Conference 2026.

Structure

  • DELTA - Pipeline implementation for decomposing user stories into tasks and evaluating the quality of the resulting decompositions
  • Prompt iterations – Prompt engineering experiments and observations
  • Descriptive statistics – Task title length statistics
  • Evaluation session – Human evaluation session including informed consent form and evaluation session protocol
  • Statistics – Per research question (RQ) statistical results

Research Questions

RQ1
How do persona-based zero- and few-shot prompting strategies affect text similarity between LLM-generated and human-created decompositions?

RQ2
How do model family and size affect text similarity between LLM-generated and human-created decompositions?

RQ3
To what extent do text similarity metrics align with human expert judgments of decomposition quality?

RQ4
To what extent does the criteria-based automated evaluation module of DELTA agree with human expert judgments of decomposition quality?

Files

Replication Package DELTA - ASE 2026.zip

Files (588.5 kB)

Name Size Download all
md5:2d297067d509ba1d88aee1a1daece8a5
588.5 kB Preview Download