Abstract
This research has three prongs, with each comparing open- and closed-book exam questions across six years (2017-2023) in a final year undergraduate applied machine learning course. First, the authors evaluated the performance of numerous LLMs, compared to student performance, and comparing open and closed book exams. Second, at a micro level, the examination questions and categories for which LLMs were most and least effective were compared. This level of analysis is rarely if ever, discussed in the literature. The research finally investigates LLM detection techniques, specifically their efficacy in identifying replies created wholly by an LLM. It considers both raw LLM outputs and LLM outputs that have been tampered with by students, with an emphasis on academic integrity. This study is a staff-student research collaboration, featuring contributions from eight academic professionals and six students.
| Original language | English |
|---|---|
| Pages | 822 |
| Number of pages | 1 |
| DOIs | |
| Publication status | Published - 8 Jul 2024 |
| Event | 29th Conference Innovation and Technology in Computer Science Education, ITiCSE 2024 - Milan, Italy Duration: 8 Jul 2024 → 10 Jul 2024 |
Conference
| Conference | 29th Conference Innovation and Technology in Computer Science Education, ITiCSE 2024 |
|---|---|
| Country/Territory | Italy |
| City | Milan |
| Period | 8/07/24 → 10/07/24 |
Keywords
- AI
- Assessment
- Detection
- Large Language Models
- LLMs
- ML