Analyzing Students’ Preferences for LLM-Generated Analogies

Seth Bernstein, Paul Denny, Juho Leinonen, Matt Littlefield, Arto Hellas, Stephen MacNeil

Research output: Contribution to conferencePosterScientificpeer-review

1 Citation (Scopus)

Abstract

Introducing students to new concepts in computer science can often be challenging, as these concepts may differ significantly from their existing knowledge and conceptual understanding. To address this, we employed analogies to help students connect new concepts to familiar ideas. Specifically, we generated analogies using large language models (LLMs), namely ChatGPT, and used them to help students make the necessary connections. In this poster, we present the results of our survey, in which students were provided with two analogies relating to different computing concepts, and were asked to describe the extent to which they were accurate, interesting, and useful. This data was used to determine how effective LLM-generated analogies can be for teaching computer science concepts, as well as how responsive students are to this approach.

Original languageEnglish
Pages812
Number of pages1
DOIs
Publication statusPublished - 8 Jul 2024
MoE publication typeNot Eligible
EventAnnual Conference on Innovation & Technology in Computer Science Education - Università degli Studi di Milano, Milan, Italy
Duration: 8 Jul 202410 Jul 2024
Conference number: 29
https://iticse.acm.org/2024/

Conference

ConferenceAnnual Conference on Innovation & Technology in Computer Science Education
Abbreviated titleITiCSE
Country/TerritoryItaly
CityMilan
Period08/07/202410/07/2024
Internet address

Keywords

  • analogies
  • computer science education
  • large language models

Fingerprint

Dive into the research topics of 'Analyzing Students’ Preferences for LLM-Generated Analogies'. Together they form a unique fingerprint.

Cite this