{"id":86,"date":"2025-08-05T10:17:56","date_gmt":"2025-08-05T09:17:56","guid":{"rendered":"https:\/\/blogs.ncl.ac.uk\/sin\/?p=86"},"modified":"2025-08-05T13:10:34","modified_gmt":"2025-08-05T12:10:34","slug":"the-unfairness-of-ai-flagged-academic-misconduct-investigations-in-uk-universities","status":"publish","type":"post","link":"https:\/\/blogs.ncl.ac.uk\/sin\/2025\/08\/05\/the-unfairness-of-ai-flagged-academic-misconduct-investigations-in-uk-universities\/","title":{"rendered":"The Unfairness of AI-Flagged Academic Misconduct Investigations in UK Universities"},"content":{"rendered":"\n<p><em><strong>Dr David Grundy,<\/strong>  Director of Digital Education, Newcastle University Business School<\/em><\/p>\n\n\n\n<figure class=\"wp-block-image size-full\"><img loading=\"lazy\" decoding=\"async\" width=\"605\" height=\"605\" src=\"https:\/\/blogs.ncl.ac.uk\/sin\/files\/2025\/08\/David1.png\" alt=\"image illustrating the unfairness of academic misconduct allegations\" class=\"wp-image-81\" srcset=\"https:\/\/blogs.ncl.ac.uk\/sin\/files\/2025\/08\/David1.png 605w, https:\/\/blogs.ncl.ac.uk\/sin\/files\/2025\/08\/David1-300x300.png 300w, https:\/\/blogs.ncl.ac.uk\/sin\/files\/2025\/08\/David1-150x150.png 150w\" sizes=\"auto, (max-width: 605px) 100vw, 605px\" \/><\/figure>\n\n\n\n<h4 class=\"wp-block-heading\"><strong>Abstract<\/strong><\/h4>\n\n\n\n<p>The increasing reliance on AI text detectors in UK higher education to flag potential academic misconduct raises profound fairness concerns. When universities initiate investigations solely based on an AI \u201cred flag,\u201d they risk contaminating the entire process, violating principles of due process and natural justice. Using the Office of the Independent Adjudicator\u2019s (OIA) Good Practice Framework as a benchmark, this article examines how an initial AI suspicion can taint investigations (\u201cfruit of the poisoned tree\u201d), the high rate of false positives that undermines \u201cjust cause,\u201d cognitive biases introduced when investigators know about AI flags, and the legal and procedural standards demanding evidence-based, impartial processes. Policy recommendations are offered to ensure AI detection tools serve only as preliminary aids, not as sole grounds for disciplinary action.<\/p>\n\n\n\n<p><strong>Keywords:<\/strong> AI Detection, Academic Misconduct Investigations, Unfair Processes, False Positives, Confirmation Bias, Anchoring Bias, Good Practice Framework<\/p>\n\n\n\n<h4 class=\"wp-block-heading\"><strong>The \u201cFruit of the Poisoned Tree\u201d<\/strong><\/h4>\n\n\n\n<p>Originating in U.S. criminal law, the \u201cfruit of the poisoned tree\u201d doctrine holds that evidence derived from tainted sources is inadmissible. In UK academia, an AI detection flag functions as a \u201cpoisoned tree\u201d when it is inherently unreliable. Initiating a misconduct investigation based solely on such a flag means the process is contaminated at its root, predisposing investigators to find guilt irrespective of subsequent evidence (Mita, 2023). Although UK courts do not formally adopt this doctrine, its underlying principle is embedded in the OIA\u2019s mandate that disciplinary processes be fair and evidence-based (OIA, 2024). An AI-generated suspicion is akin to an unsubstantiated tip-off rather than concrete proof; using it as the trigger for formal inquiries unfairly prejudices the student, as any exculpatory evidence may be discounted through a lens of presumed guilt. To uphold justice, universities must treat AI flags as prompts for further verification, not as grounds for launching full-scale investigations.<\/p>\n\n\n\n<h4 class=\"wp-block-heading\"><strong>False Positives and Lack of \u201cJust Cause\u201d for Investigation<\/strong><\/h4>\n\n\n\n<p>A cornerstone of fairness in academic discipline is \u201cjust cause\u201d\u2014reasonable grounds to suspect misconduct. Generative AI detectors, however, exhibit significant error rates (Dalalah &amp; Dalalah, 2023; Giray, 2024). Turnitin\u2019s own data reveal a 4% sentence-level false-positive rate (Turnitin, 2023), implying that a sizeable fraction of human-written work could be mislabelled. Walters (2023) estimates that even a 10% false-positive rate could wrongfully implicate dozens of students per cohort, accumulating to hundreds over multiple years. Empirical assessments of 14 detectors report accuracies ranging from merely 33% to 81% (Weber-Wulff et al., 2023), and Temple University\u2019s evaluation found Turnitin\u2019s detector only 77% accurate at spotting AI text, with a 7% mis-flag rate for genuine writing (Asselta Law, 2025). The University of Pennsylvania concurs that many detectors have \u201cdangerously high\u201d false-positive defaults (Wood, 2024). Moreover, bias against non-native English writers exacerbates injustices: Liang et al. (2023) demonstrate that GPT detectors disproportionately target second-language students. An AI flag, therefore, falls far short of providing reasonable suspicion; it resembles unreliable hearsay rather than \u201chard evidence\u201d required to justify formal proceedings (OIA, 2024). Without corroborating indicators\u2014such as verbatim matches to external sources or inability to reproduce the work\u2014launching investigations on AI scores alone constitutes an unjustified witch-hunt, subjecting innocent students to undue stress and reputational harm (Gorichanaz, 2023).<\/p>\n\n\n\n<h4 class=\"wp-block-heading\"><strong>Cognitive Bias and Presumption of Guilt in AI-Flagged Cases<\/strong><\/h4>\n\n\n\n<p>Once a case is initiated on the basis of an AI flag, investigators become vulnerable to cognitive biases. Confirmation bias leads them to seek out evidence that confirms the initial AI suspicion while overlooking exculpatory signs (Rassin, 2022; Wallace, 2015). Anchoring bias further cements this effect: a reported \u201c85% AI-generated\u201d score becomes an immovable reference point, skewing all subsequent evaluations (Ly et al., 2023). Forensic research shows that contextual suggestions of guilt \u201ccannot be unseen,\u201d distorting experts\u2019 judgments even after the context is removed (Kunkler &amp; Roy, 2023). In academic settings, knowing an essay was flagged primes investigators to interpret well-crafted sections suspiciously (\u201ctoo good to be the student\u2019s own\u201d), rather than neutrally assessing content and process (OIA, 2024). Instructors may unconsciously scrutinize minor stylistic deviations or benign editing-tool usage as signs of cheating, while ignoring personal reflections or draft submissions that demonstrate authorship. This is particularly harmful for international students, whose strong writing skills can be misattributed to AI (Mathewson, 2023). By injecting a presumption of guilt and tainting the investigator\u2019s mindset, AI flags undermine the \u201cinnocent until proven guilty\u201d ethos and violate natural justice standards that forbid \u201cany reasonable perception of bias or pre-determination\u201d (OIA, 2024).<\/p>\n\n\n\n<h4 class=\"wp-block-heading\"><strong>Due Process, Fairness, and Evidence<\/strong><\/h4>\n\n\n\n<p>UK universities operate under internal regulations and public law requiring fairness, impartiality, and evidence-based decisions (OIA, 2024). Students are entitled to know precise allegations and the evidence against them, and to respond fully. AI detectors, however, provide opaque probability scores with no transparent rationale, denying students a meaningful opportunity to challenge the \u201cevidence\u201d (Turnitin, 2023). The typical \u201cbalance of probabilities\u201d standard demands substantive proof that misconduct more likely than not occurred\u2014an unsubstantiated AI score cannot meet this threshold. In appeals to the OIA, universities would struggle to justify decisions hinging on black-box algorithms rather than verifiable facts. Furthermore, evidence obtained through improper means\u2014such as uploading student work to unauthorized free detection tools\u2014may violate GDPR and intellectual property rights, rendering it inadmissible. Courts will intervene if procedural fairness or contract terms are breached; a case built chiefly on AI probabilities risks being overturned as procedurally unfair (OIA, 2024). To satisfy due process, any allegation of AI-assisted plagiarism should be substantiated with concrete examples\u2014verbatim matches, inability to reproduce work, or clear stylistic mismatches informed by multiple writing samples\u2014and accompanied by full disclosure of tool limitations to the student.<\/p>\n\n\n\n<h4 class=\"wp-block-heading\"><strong>Policy Recommendations<\/strong><\/h4>\n\n\n\n<p><em>Thresholds for \u201cJust Cause\u201d<\/em><\/p>\n\n\n\n<p>AI detection results must not be the sole trigger for investigations. Universities should require corroborating evidence\u2014verbatim text matches, stark deviations from the student\u2019s known writing style, or inability to demonstrate authorship\u2014before proceeding. Policies should explicitly state that AI flags serve only as preliminary guidance (JISC, 2023; Wargo &amp; Anderson, 2024).<\/p>\n\n\n\n<p><em>Human Oversight and Professional Scepticism<\/em><\/p>\n\n\n\n<p>Flagged submissions should prompt human review by trained subject-matter experts or integrity officers. Reviewers must consider benign explanations (talent, grammar tools, personal reflections) and treat AI outputs as invitations for inquiry, not as proof (Jones &amp; Newton, 2024).<\/p>\n\n\n\n<p><em>Process Design to Mitigate Bias<\/em><\/p>\n\n\n\n<p>Implement partial blinding: require markers to document independent concerns before viewing AI reports, or assign AI review to separate officers. Deliver training on cognitive biases in misconduct investigations, using case studies to highlight the risks of anchoring and confirmation bias (Born, 2024).<\/p>\n\n\n\n<p><em>Transparent, Fair Regulations<\/em><\/p>\n\n\n\n<p>Update academic integrity policies to affirm that no student may be penalized based solely on an AI detector. Incorporate explicit language: \u201cAutomated AI detection scores are preliminary aids only; findings must rest on verifiable evidence and academic judgment.\u201d Disclose tool limitations and provide students access to the AI report and its accuracy parameters.<\/p>\n\n\n\n<p><em>Assessment Design and Pedagogy<\/em><\/p>\n\n\n\n<p>Shift away from punitive, detector-centric approaches toward authentic, iterative assessments (vivas, personalized tasks, draft-based assignments) that inherently discourage misconduct. Emphasize trust-based evaluation and support systems, reducing the institutional reliance on unreliable detection software (OIA, 2024).<\/p>\n\n\n\n<h4 class=\"wp-block-heading\"><strong>Statement: Acknowledgement of Assistive Tool Usage<\/strong><\/h4>\n\n\n\n<p>This document is the authors original own work, however Microsoft Word and Grammarly drafting and editing tools were used in the creation of the document as the author is dyslexic and the work would be unreadable without their usage to refine the text. Both tools incorporate elements of machine learning and AI. The author takes full responsibility for the content of the published article.<\/p>\n\n\n\n<h4 class=\"wp-block-heading\"><strong>References<\/strong><\/h4>\n\n\n\n<p>Asselta Law. (2025, February 16). <em>The Hysteria of Professors Accusing Students of Using AI<\/em>. <a href=\"https:\/\/www.asseltalaw.com\/blog\/2025\/02\/the-hysteria-of-professors-accusing-students-of-using-ai\/\">https:\/\/www.asseltalaw.com\/blog\/2025\/02\/the-hysteria-of-professors-accusing-students-of-using-ai\/<\/a><\/p>\n\n\n\n<p>Baker, R. S., &amp; Hawn, A. (2022). Algorithmic Bias in Education. <em>International Journal of Artificial Intelligence in Education<\/em>, <em>32<\/em>(4), 1052\u20131092. <a href=\"https:\/\/doi.org\/10.1007\/s40593-021-00285-9\">https:\/\/doi.org\/10.1007\/s40593-021-00285-9<\/a><\/p>\n\n\n\n<p>Born, R. T. (2024). Stop Fooling Yourself! (Diagnosing and Treating Confirmation Bias). <em>eNeuro<\/em>, <em>11<\/em>(10). <a href=\"https:\/\/doi.org\/10.1523\/ENEURO.0415-24.2024\">https:\/\/doi.org\/10.1523\/ENEURO.0415-24.2024<\/a><\/p>\n\n\n\n<p>Chechitelli, A. (2023, June 14). <em>Understanding the false positive rate for sentences of our AI writing detection capability<\/em>. Turnitin Blog. <a href=\"https:\/\/www.turnitin.com\/blog\/understanding-the-false-positive-rate-for-sentences-of-our-ai-writing-detection-capability\">https:\/\/www.turnitin.com\/blog\/understanding-the-false-positive-rate-for-sentences-of-our-ai-writing-detection-capability<\/a><\/p>\n\n\n\n<p>Cambridge University (2023, October). <em>Investigating academic misconduct and mark checks<\/em> [Text]. <a href=\"https:\/\/www.studentcomplaints.admin.cam.ac.uk\/staff-support\/investigating-academic-misconduct-and-mark-checks\">https:\/\/www.studentcomplaints.admin.cam.ac.uk\/staff-support\/investigating-academic-misconduct-and-mark-checks<\/a><\/p>\n\n\n\n<p>Dalalah, D., &amp; Dalalah, O. M. A. (2023). The false positives and false negatives of generative AI detection tools in education and academic research: The case of ChatGPT. <em>The International Journal of Management Education<\/em>, <em>21<\/em>(2), 100822. <a href=\"https:\/\/doi.org\/10.1016\/j.ijme.2023.100822\">https:\/\/doi.org\/10.1016\/j.ijme.2023.100822<\/a><\/p>\n\n\n\n<p>Eaton, S. E. (2022, February 22). <em>Check your bias at the door<\/em>. University Affairs. <a href=\"https:\/\/universityaffairs.ca\/features\/check-your-bias-at-the-door\/\">https:\/\/universityaffairs.ca\/features\/check-your-bias-at-the-door\/<\/a><\/p>\n\n\n\n<p>Foltynek, T., Bjelobaba, S., Glendinning, I., Khan, Z. R., Santos, R., Pavletic, P., &amp; Kravjar, J. (2023). ENAI Recommendations on the ethical use of Artificial Intelligence in Education. <em>International Journal for Educational Integrity<\/em>, <em>19<\/em>(1), 12. <a href=\"https:\/\/doi.org\/10.1007\/s40979-023-00133-4\">https:\/\/doi.org\/10.1007\/s40979-023-00133-4<\/a><\/p>\n\n\n\n<p>Giray, L. (2024). The Problem with False Positives: AI Detection Unfairly Accuses Scholars of AI Plagiarism. <em>The Serials Librarian<\/em>, <em>85<\/em>(5\u20136), 181\u2013189. <a href=\"https:\/\/doi.org\/10.1080\/0361526X.2024.2433256\">https:\/\/doi.org\/10.1080\/0361526X.2024.2433256<\/a><\/p>\n\n\n\n<p><em>Good Practice Framework\u2014OIAHE<\/em> (Worldwide). (2024, June 3). <a href=\"https:\/\/www.oiahe.org.uk\/resources-and-publications\/good-practice-framework\/\">https:\/\/www.oiahe.org.uk\/resources-and-publications\/good-practice-framework\/<\/a><\/p>\n\n\n\n<p>Gorichanaz, T. (2023). Accused: How students respond to allegations of using ChatGPT on assessments. <em>Learning: Research and Practice<\/em>, <em>9<\/em>(2), 183\u2013196. <a href=\"https:\/\/doi.org\/10.1080\/23735082.2023.2254787\">https:\/\/doi.org\/10.1080\/23735082.2023.2254787<\/a><\/p>\n\n\n\n<p>Kunkler, K. S., &amp; Roy, T. (2023). Reducing the impact of cognitive bias in decision making: Practical actions for forensic science practitioners. <em>Forensic Science International: Synergy<\/em>, <em>7<\/em>, 100341. <a href=\"https:\/\/doi.org\/10.1016\/j.fsisyn.2023.100341\">https:\/\/doi.org\/10.1016\/j.fsisyn.2023.100341<\/a><\/p>\n\n\n\n<p>Liang, W., Yuksekgonul, M., Mao, Y., Wu, E., &amp; Zou, J. (2023). GPT detectors are biased against non-native English writers. <em>Patterns<\/em>, <em>4<\/em>(7). <a href=\"https:\/\/doi.org\/10.1016\/j.patter.2023.100779\">https:\/\/doi.org\/10.1016\/j.patter.2023.100779<\/a><\/p>\n\n\n\n<p>Ly, D. P., Shekelle, P. G., &amp; Song, Z. (2023). Evidence for Anchoring Bias During Physician Decision-Making. <em>JAMA Internal Medicine<\/em>, <em>183<\/em>(8), 818\u2013823. <a href=\"https:\/\/doi.org\/10.1001\/jamainternmed.2023.2366\">https:\/\/doi.org\/10.1001\/jamainternmed.2023.2366<\/a><\/p>\n\n\n\n<p>Mathewson, T. G. (2023, August 14). <em>AI Detection Tools Falsely Accuse International Students of Cheating \u2013 The Markup<\/em>. <a href=\"https:\/\/themarkup.org\/machine-learning\/2023\/08\/14\/ai-detection-tools-falsely-accuse-international-students-of-cheating\">https:\/\/themarkup.org\/machine-learning\/2023\/08\/14\/ai-detection-tools-falsely-accuse-international-students-of-cheating<\/a><\/p>\n\n\n\n<p>Mita, S. (2022). AI Proctoring: Academic Integrity vs. Student Rights Notes. <em>Hastings Law Journal<\/em>, <em>74<\/em>(5), [i]-1554.<\/p>\n\n\n\n<p>Newton, P. M., &amp; Jones, S. (2025). Education and Training Assessment and Artificial Intelligence. A Pragmatic Guide for Educators. <em>British Journal of Biomedical Science<\/em>, <em>81<\/em>, 14049. <a href=\"https:\/\/doi.org\/10.3389\/bjbs.2024.14049\">https:\/\/doi.org\/10.3389\/bjbs.2024.14049<\/a><\/p>\n\n\n\n<p>Rassin, E. (2022). \u2018Anyone who commits such a cruel crime, must be criminally irresponsible\u2019: Context effects in forensic psychological assessment. <em>Psychiatry, Psychology and Law<\/em>, <em>29<\/em>(4), 506\u2013515. <a href=\"https:\/\/doi.org\/10.1080\/13218719.2021.1938272\">https:\/\/doi.org\/10.1080\/13218719.2021.1938272<\/a><\/p>\n\n\n\n<p>Sadasivan, V. S., Kumar, A., Balasubramanian, S., Wang, W., &amp; Feizi, S. (2025). <em>Can AI-Generated Text be Reliably Detected?<\/em> (No. arXiv:2303.11156). arXiv. <a href=\"https:\/\/doi.org\/10.48550\/arXiv.2303.11156\">https:\/\/doi.org\/10.48550\/arXiv.2303.11156<\/a><\/p>\n\n\n\n<p>Wallace, W. A. (2015). The Effect of Confirmation Bias in Criminal Investigative Decision Making [Ph.D., Walden University]. In <em>ProQuest Dissertations and Theses<\/em>. <a href=\"https:\/\/www.proquest.com\/docview\/1668379477\/abstract\/576B938495004949PQ\/1\">https:\/\/www.proquest.com\/docview\/1668379477\/abstract\/576B938495004949PQ\/1<\/a><\/p>\n\n\n\n<p>Walters, W. H. (2023). The Effectiveness of Software Designed to Detect AI-Generated Writing: A Comparison of 16 AI Text Detectors. <em>Open Information Science<\/em>, <em>7<\/em>(1). <a href=\"https:\/\/doi.org\/10.1515\/opis-2022-0158\">https:\/\/doi.org\/10.1515\/opis-2022-0158<\/a><\/p>\n\n\n\n<p>Wargo, K., &amp; Anderson, B. (n.d.). <em>Striking a Balance: Navigating the Ethical Dilemmas of AI in Higher Education<\/em>. EDUCAUSE Review. Retrieved 28 February 2025, from <a href=\"https:\/\/er.educause.edu\/articles\/2024\/12\/striking-a-balance-navigating-the-ethical-dilemmas-of-ai-in-higher-education\">https:\/\/er.educause.edu\/articles\/2024\/12\/striking-a-balance-navigating-the-ethical-dilemmas-of-ai-in-higher-education<\/a><\/p>\n\n\n\n<p>Webb, M. (2023, September 18). AI Detection\u2014Latest Recommendations. <em>Artificial Intelligence<\/em>. <a href=\"https:\/\/nationalcentreforai.jiscinvolve.org\/wp\/2023\/09\/18\/ai-detection-latest-recommendations\/\">https:\/\/nationalcentreforai.jiscinvolve.org\/wp\/2023\/09\/18\/ai-detection-latest-recommendations\/<\/a><\/p>\n\n\n\n<p>Weber-Wulff, D., Anohina-Naumeca, A., Bjelobaba, S., Folt\u00fdnek, T., Guerrero-Dib, J., Popoola, O., \u0160igut, P., &amp; Waddington, L. (2023). Testing of detection tools for AI-generated text. <em>International Journal for Educational Integrity<\/em>, <em>19<\/em>(1), 26. <a href=\"https:\/\/doi.org\/10.1007\/s40979-023-00146-z\">https:\/\/doi.org\/10.1007\/s40979-023-00146-z<\/a><\/p>\n\n\n\n<p>Wood, C. (2024, September 10). AI detectors are easily fooled, researchers find. <em>EdScoop<\/em>. <a href=\"https:\/\/edscoop.com\/ai-detectors-are-easily-fooled-researchers-find\/\">https:\/\/edscoop.com\/ai-detectors-are-easily-fooled-researchers-find\/<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Dr David Grundy, Director of Digital Education, Newcastle University Business School Abstract The increasing reliance on AI text detectors in UK higher education to flag potential academic misconduct raises profound fairness concerns. When universities initiate investigations solely based on an AI \u201cred flag,\u201d they risk contaminating the entire process, violating principles of due process and [&hellip;]<\/p>\n","protected":false},"author":3172,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[1],"tags":[14,13,18,17,16,19,15],"class_list":["post-86","post","type-post","status-publish","format-standard","hentry","category-uncategorised","tag-academic-misconduct-investigations","tag-ai-detection","tag-anchoring-bias","tag-confirmation-bias","tag-false-positives","tag-good-practice-framework","tag-unfair-processes"],"_links":{"self":[{"href":"https:\/\/blogs.ncl.ac.uk\/sin\/wp-json\/wp\/v2\/posts\/86","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/blogs.ncl.ac.uk\/sin\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/blogs.ncl.ac.uk\/sin\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/blogs.ncl.ac.uk\/sin\/wp-json\/wp\/v2\/users\/3172"}],"replies":[{"embeddable":true,"href":"https:\/\/blogs.ncl.ac.uk\/sin\/wp-json\/wp\/v2\/comments?post=86"}],"version-history":[{"count":1,"href":"https:\/\/blogs.ncl.ac.uk\/sin\/wp-json\/wp\/v2\/posts\/86\/revisions"}],"predecessor-version":[{"id":87,"href":"https:\/\/blogs.ncl.ac.uk\/sin\/wp-json\/wp\/v2\/posts\/86\/revisions\/87"}],"wp:attachment":[{"href":"https:\/\/blogs.ncl.ac.uk\/sin\/wp-json\/wp\/v2\/media?parent=86"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/blogs.ncl.ac.uk\/sin\/wp-json\/wp\/v2\/categories?post=86"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/blogs.ncl.ac.uk\/sin\/wp-json\/wp\/v2\/tags?post=86"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}