Measuring Similarity between Manual Course Concepts and ChatGPT-generated Course Concepts
Creators
Contributors
Editors:
- 1. WestEd, USA
- 2. EPFL, Switzerland
- 3. Google Research and Indian Institute of Science, India
Description
ChatGPT is a state-of-the-art language model that facilitates natural language interaction, enabling users to acquire textual responses to their inquiries. The model's ability to generate answers with a human-like quality has been reported. While the model's natural language responses can be evaluated by human experts in the field through thorough reading, assessing its structured responses, such as lists, can prove challenging even for experts. This study compares an openly accessible, manually validated list of "course concepts," or knowledge concepts taught in courses, to the concept lists generated by ChatGPT. Course concepts assist learners in deciding which courses to take by distinguishing what is taught in courses from what is considered prerequisites. Our experimental results indicate that only 22\% to 33\% of the concept lists produced by ChatGPT were included in the manually validated list of 4,096 concepts in computer science courses, suggesting that these concept lists require manual adjustments for practical use. Notably, when ChatGPT generates a concept list for non-native English speakers, the overlap increases, whereas the language used for querying the model has a minimal impact. Additionally, we conducted a qualitative analysis of the concepts generated but not present in the manual list.
Files
2023.EDM-posters.52.pdf
Files
(173.5 kB)
Name | Size | Download all |
---|---|---|
md5:5924c191ea69324d31f474a8f0513e4e
|
173.5 kB | Preview Download |