The Death of Take-Home Assessment in the Era of GenAI, Here Is the Evidence
Discuss this preprint
Start a discussion What are Sciety discussions?Listed in
This article is not in any list yet, why not save it to one of your lists.Abstract
This study examined whether a mature, empirically validated generative artificial intelligence (GenAI) intervention framework can produce reliable process evidence when deployed in unsupervised take-home assessments. Twenty-five group submissions from two cybersecurity management cohorts were audited using a five-check protocol that tested primary evidence presence, traceability, internal data consistency, modification provenance, and reflection specificity. The assessments were designed using the Structured AI-Guided Education (SAGE) framework and incorporated base prompts, structured decision tables, mandatory AI interaction logs, and reflective commentary. Only 3 of 25 submissions (12%) produced evidence chains that were substantially auditable. Full traceability between documented AI outputs and human evaluation claims was not achieved in any of the 25 submissions. The remaining submissions exhibited logical checksum failures, compliance-pattern text in evaluation cells, procedural rather than functional reflection, and structural indicators consistent with audit trail simulation. These patterns were consistent across both cohorts. The paper identifies a compliance gradient in which conscientious students who follow the process in good faith incur a disproportionate documentation burden, while students who simulate compliance can produce comparable outputs with less effort. On the basis of this evidence, the paper argues that take-home assessments can no longer be relied upon as standalone assurance instruments in the GenAI era. SAGE remains a validated pedagogy for fostering AI orchestration competency through scaffolded tutorial practice. However, the burden of assurance must shift to secure, supervised tasks where process fidelity cannot be simulated.