The Use of AI-Detection Tools in the Assessment of Student Work

May 6, 2023

People have been asking if they should be using detection tools to identify text written by ChatGPT or other artificial intelligence writing apps. Just this week I was a panelist in a session on “AI and You: Ethics, Equity, and Accessibility”, part of ETMOOC 2.0. Alec Couros asked what I was seeing across Canada in terms of universities using artificial intelligence detection in misconduct cases.

The first thing I shared was the University of British Columbia web page stating that the university was not enabling Turnitin’s AI-detection feature. UBC is one of the few universities in Canada that subscribes to Turnitin.

The Univeristy of British Columbia declares the university is not enabling Turnitin’s AI-detection feature.

Turnitin’s rollout of AI detection earlier this year was widely contested and I won’t go into that here. What I will say is that whether AI detection is a new feature embedded into existing product lines or a standalone product, there is little actual scientific evidence to show that AI-generated text can be effectively detected (see Sadasivan et al., 2023). In a TechCrunch article, Open AI, the company that developed ChatGPT, talked about its own detection tool, noting that its success rate was around 26%

Key message: Tools to detect text written by artificial intelligence aren’t really reliable or effective. It would be wise to be skeptical of any marketing claims to the contrary.

There are news reports about students being falsely accused of misconduct when the results of AI writing detection tools were used as evidence. See news stories here and here, for example. 

There have been few studies done on the impact of a false accusation of student academic misconduct, but if we turn to the literature on false accusations in criminal offences, there is evidence showing that false accusations can result in reputation damage, self-stigma, depression, anxiety, PTSD, sleep problems, social isolation, and strained relationships, among other outcomes. Falsely accusing students of academic misconduct can be devastating, including dying by suicide as a result. You can read some stories about students dying by suicide after false allegations of academic cheating in the United States and in India. Of course, stories about student suicide are rarely discussed in the media, for a variety of reasons. The point here is that false accusations of students for academic cheating can have a negative impact on their mental and physical health.

Key message: False accusations of academic misconduct can be devastating for students.

Although reporting allegations of misconduct remains a responsibility of educators, having fully developed (and mandatory) case management and investigation systems is imperative. Decisions about whether misconduct has occurred should be made carefully and thoughtfully, using due process that follows established policies.

It is worth noting that AI-generated text can be revised and edited such that the end product is neither fully written by AI, nor fully written by a human. At our university, the use of technology to detect possible misconduct may not be used deceptively or covertly. For example, we do not have an institutional license to any text-matching software. Individual professors can get a subscription if they wish, but the use of detection tools should be declared in the course syllabus. If detection tools are used post facto, it can be considered a deception on the part of the professor because the students were not made aware of the technology prior to handing in their assessment. 

Key message: Students can appeal any misconduct case brought forward with the use of deceptive or undisclosed assessment tools or technology (and quite frankly, they would probably win the appeal).

If we expect students to be transparent about their use of tools, then it is up to educators and administrators also to be transparent about their use of technology prior to assessment and not afterwards. A technology arms race in the name of integrity is antithetical to teaching and learning ethically and can perpetuate antagonistic and adversarial relationships between educators and students.

Ethical Principles for Detecting AI-Generated Text in Student Work

Let me be perfectly clear: I am not at all a fan of using detection tools to identify possible cases of academic misconduct. But, if you insist on using detection tools, for heaven’s sake, be transparent and open about your use of them.

Here is an infographic you are welcome to use and share: Infographic: “Ethical Principles for Detecting AI-Generated Text in Student Work” (Creative Commons License: Attribution-NonCommercial-ShareAlike 4.0 International). The text inside the infographic is written out in full with some additional details below.

Here is some basic guidance:

Check your Institutional Policies First

Before you use any detection tools on student work, ensure that the use of such tools is permitted according to your school’s academic integrity policy. If your school does not have such a policy or if the use of detection tools is not mentioned in the policy, that does not automatically mean that you have the right to use such tools covertly. Checking the institutional policies and regulations is a first step, but it is not the only step in applying the use of technology ethically in assessment of student work.

Check with Your Department Head

Whether the person’s title is department head, chair, headmaster/headmistress, principal, or something else, there is likely someone in your department, faculty or school whose job it is to oversee the curriculum and/or matters relating to student conduct. Before you go rogue using detection tools to catch students cheating, ask the person to whom you report if they object to the use of such tools. If they object, then do not go behind their back and use detection tools anyway. Even if they agree, then it is still important to use such tools in a transparent and open way, as outlined in the next two recommendations.

Include a Statement about the Use of Detection Tools in Your Course Syllabus

Include a clear written statement in your course syllabus that outlines in plain language exactly which tools will be used in the assessment of student work. A failure to inform students in writing about the use of detection tools before they are used could constitute unethical assessment or even entrapment. Detection tools should not be used covertly. Their use should be openly and transparently declared to students in writing before any assessment or grading begins.

Of course, having a written statement in a course syllabus does not absolve educators of their responsibility to have open and honest conversations with students, which is why the next point is included.

Talk to Students about Your Use of Tools or Apps You will Use as Part of Your Assessment 

Have open and honest conversations with students about how you plan to use detection tools. Point out that there is a written statement in the course outline and that you have the support of your department head and the institution to use these tools. Be upfront and clear with students.

It is also important to engage students in evidence-based conversations about the limitations tools to detect artificial intelligence writing, including the current lack of empirical evidence about how well they work.

Conclusion

Again, I emphasize that I am not at all promoting the use of any AI detection technology whatsoever. In fact, I am opposed to the use of surveillance and detection technology that is used punitively against students, especially when it is done in the name of teaching and learning. However, if you are going to insist on using technology to detect possible breaches of academic integrity, then at least do so in an open and transparent way — and acknowledge that the tools themselves are imperfect.

Key message: Under no circumstances should the results from an AI-writing detection tool be used as the only evidence in a student academic misconduct allegation.

I am fully anticipating some backlash to this post. There will be some of you who will object to the use detection tools on principle and counter that any blog post talking about how they can be used is in itself unethical. You might be right, but the reality remains that thousands of educators are currently using detection tools for the sole purpose of catching cheating students. As much as I rally against a “search and destroy” approach, there will be some people who insist on taking this position. This blog post is to offer some guidelines to avoid deceptive assessment and covert use of technology in student assessment.

Key message: Deceptive assessment is a breach of academic integrity on the part of the educator. If we want students to act with integrity, then it is up to educators to model ethical behaviour themselves.

References

Sadasivan, V. S., Kumar, A., Balasubramanian, S., Wang, W., & Feizi, S. (2023). Can AI-Generated Text be Reliably Detected? ArXiv. https://doi.org/10.48550/arXiv.2303.11156

Fowler, G. A. (2023, April 3). We tested a new ChatGPT-detector for teachers. It flagged an innocent student. Washington Post. https://www.washingtonpost.com/technology/2023/04/01/chatgpt-cheating-detection-turnitin/

Jimenez, K. (2023, April 13). Professors are using ChatGPT detector tools to accuse students of cheating. But what if the software is wrong? USA Today. https://www.usatoday.com/story/news/education/2023/04/12/how-ai-detection-tool-spawned-false-cheating-case-uc-davis/11600777002/

_________________________________

Share this post: The Use of AI-Detection Tools in the Assessment of Student Work https://drsaraheaton.wordpress.com/2023/05/06/the-use-of-ai-detection-tools-in-the-assessment-of-student-work/

This blog has had over 3 million views thanks to readers like you. If you enjoyed this post, please “like” it or share it on social media. Thanks! Sarah Elaine Eaton, PhD, is a faculty member in the Werklund School of Education, and the Educational Leader in Residence, Academic Integrity, University of Calgary, Canada. Opinions are my own and do not represent those of the University of Calgary.


How to Talk to Your Students about ChatGPT: A Lesson Plan for High School and College Students

April 7, 2023
bionic hand and human hand finger pointing

Photo by cottonbro studio on Pexels.com

This article by Ben Edwards in ArtsTechnica (April 6, 2023) is worth a read, “Why ChatGPT and Bing Chat are so good at making things up”.

Edwards explains in clear language, with lots of details and examples, how and why large language models (LLMs) such as ChatGPT make up content. As I read this article, it occurred to me that it could serve as a really great way to have pro-active and generative conversations with students about the impact of artificial intelligence for teaching, learning, assessment, and academic integrity. So, here is a quick lesson plan about how to use this article in class:

Education level

Secondary school and post-secondary (e.g., community college, polytechnic, undergraduate or graduate university courses)

Lesson Plan Title: Understanding ChatGPT: Benefits and Limitations

Learning Objectives

By the end of this lesson students will be able to:

  • Understand how and why AI-writing apps make up content.
  • Explain the term “confabulation”.
  • Discuss the implications of fabricated content on academic integrity
  • Generate ideas about how to fact-check AI-generated content to ensure its accuracy

Lesson Preparation

Prior to the class, students should read this article: “Why ChatGPT and Bing Chat are so good at making things up by Ben Edwards, published in ArtsTechnica (April 6, 2023)

Come to class prepared to discuss the article.

Learning Activity

Class discussion (large group format if the class is small or small group format with a large group debrief at the end):

Possible guiding questions:

  • What is your experience with ChatGPT and other AI writing apps?
  • What were the main points in this article? (Alternate phrasing: What were your key takeaways from this article?)
  • What are some of the risks when AI apps engage in confabulation (i.e., fabrication)?
  • Discuss this quotation from the article, “ChatGPT as it is currently designed, is not a reliable source of factual information and cannot be trusted as such.”
  • Fabrication and falsification are commonly included in academic misconduct policies. What do you think the implications are for students and researchers when they write with AI apps?
  • What are some strategies or tips we can use to fact-check text generated by AI apps?
  • What is the importance of prompt-writing when working with AI writing apps?

Duration

The time commitment for the pre-reading will vary from one student to the next. The duration of the learning activity can be adjusted to suit the needs of your class.
  • Students’ pre-reading of the article: 60-minutes or less
  • Learning activity: 45-60 minutes

Lesson closure

Thank students for engaging actively in the discussion and sharing their ideas.

Possible Follow-up Activities

  • Tips for fact-checking. Have students in the class generate their own list of tips to fact-check AI-generated content (e.g., in a shared Google doc or by sharing ideas orally in class that one person inputs into a document on behalf of the class.)
  • Prompt-writing activity. Have students use different prompts to generate content from AI writing apps. Ask them to document each prompt and write down their observations about what worked and what didn’t. Discuss the results as a class.
  • Academic Integrity Policy Treasure Hunt and Discussion. Have students locate the school’s academic misconduct / academic integrity policy. Compare the definitions and categories for academic misconduct in the school’s policies with concepts presented in this article such as confabulation. Have students generate their own ideas about how to uphold the school’s academic integrity policies when using AI apps.

Creative Commons License

This lesson plan is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License (CC BY-NC 4.0). This license applies only to the lesson plan, not to the original article by Ben Edwards.

Additional Notes

This is a generic (and imperfect) lesson plan. It can (and probably should) be adapted or personalized depending on the needs of the learners.

Acknowledgements

Thanks to Dr. Rahul Kumar, Brock University for providing an open peer review of this lesson plan.

 _________________________________

Share or Tweet this: How to Talk to Your Students about ChatGPT: A Lesson Plan for High School and College Students – https://drsaraheaton.wordpress.com/2023/04/07/how-to-talk-to-your-students-about-chatgpt-a-lesson-plan-for-high-school-and-college-students This blog has had over 3 million views thanks to readers like you. If you enjoyed this post, please “like” it or share it on social media. Thanks! Sarah Elaine Eaton, PhD, is a faculty member in the Werklund School of Education, and the Educational Leader in Residence, Academic Integrity, University of Calgary, Canada. Opinions are my own and do not represent those of the University of Calgary.

Now Published: Contract Cheating in Higher Education

August 11, 2022
Eaton, Curtis, Clare, Stoesz, Seeland & Rundle cover

Now published: Contract Cheating in Higher Education: Global Perspectives on Theory, Practice, and Policy (Eaton, Curtis, Stoesz, Clare, Rundle, & Seeland, 2022). This volume, published by Palgrave MacMillan, includes twenty chapters from contributors across Australia, North America, and Europe:

  1. Introduction: Contract cheating and introduction to the problem. – Curtis, G. J., Clare, J., Rundle, K., Eaton, S. E., Stoesz, B. M., & Seeland, J.
  2. What can we learn from measuring crime when looking to quantify the prevalence and incidence of contract cheating? – Clare, J., & Rundle, K.
  3. Limitations of contract cheating research. –  Krásničan, V., Foltýnek, T., & Dlabolová, D. H.
  4. Essay mills and contract cheating from a legal point of view. – Draper, M.
  5. Leveraging college copyright ownership against file-sharing and contract cheating websites. – Seeland, J., Eaton, S. E., & Stoesz, B. M.
  6. The encouragement of file sharing behaviours through technology and social media: Impacts on student cheating behaviours and academic piracy. – Rogerson, A. M.
  7. Higher education assessment design. – Sutherland-Smith, W., & Dawson, P.
  8. Critical thinking as an antidote to contract cheating. – Stoesz, B. M., Eaton, S. E., & Seeland, J.
  9. Contract cheating and the Dark Triad traits. – Baran, L., & Jonason, P. K.
  10. Contract cheating: The influence of attitudes and emotions. – Curtis, G. J., & Tindall, I. K.
  11. Applying situational crime prevention techniques to contract cheating. – Clare, J.
  12. Presentation, Properties and Provenance: the three Ps of identifying evidence of contract-cheating in student assignments. – Crockett, R.
  13. “(Im)possible to prove”: Formalising academic judgement evidence in contract cheating cases using bibliographic forensics. – Ellis, C., Rogerson, A. M., House, D., & Murdoch, K.
  14. Aligning academic quality and standards with academic integrity – Glendinning, I.
  15. Addressing contract cheating through staff-student partnerships. – Lancaster, T.
  16. The extortionate cost of contract cheating. – Veeran-Colton, T., Sefcik, L., & Yorke, J.
  17. The rise of contract cheating in graduate education. –  Parnther, C.
  18. Listening to ghosts: A qualitative study of narratives from contract cheating writers from the 1930s onwards. – Eaton, S. E., Stoesz, B. M., & Seeland, J.
  19. Assessment brokering and collaboration: Ghostwriter and student academic literacies. – Thacker, E. J.
  20. Conclusion – Eaton, S. E., Stoesz, B. M., Seeland, J. Curtis, G. J., Clare, J., & Rundle, K. 

Updates:

October 4, 2022 – The final proofs have been submitted to the publisher. The book now moves into production and should be available very soon!

October 28, 2022 – The book has now been published!

_______________________________

Share or Tweet this: New Book Coming Soon: Contract Cheating in Higher Education – https://drsaraheaton.wordpress.com/2022/08/11/new-book-coming-soon-contract-cheating-in-higher-education/

This blog has had over 3 million views thanks to readers like you. If you enjoyed this post, please “like” it or share it on social media. Thanks!

Sarah Elaine Eaton, PhD, is a faculty member in the Werklund School of Education, and the Educational Leader in Residence, Academic Integrity, University of Calgary, Canada. Opinions are my own and do not represent those of the University of Calgary.


Upcoming Webinar: None of the above: Integrity concerns of standardized English proficiency tests with Soroush Sabbaghan and Ismaeil Fazel

January 4, 2021

Sabbaghan Fazel webinarThis session will bring to fore (or highlight) the oft-neglected discord between equity and integrity in high-stakes standardized language tests. The equity issues surrounding these so-called standardized language tests can potentially precipitate and predispose academic dishonesty. This presentation will discuss the ramifications of inherent inequities in high stakes language proficiency tests for academic integrity and will call for a more critical consideration of commercialized high stakes language tests. Redressing equity issues in language assessment can serve to promote academic integrity and reduce academic dishonesty.

Learning outcomes

  • Learn more about challenges to equity in high stakes language testing.
  • Recognize discords between equity and integrity in commercial standardized language tests
  • Review principles and best practices for equitable language assessment

About the presenters

Ismaeil Fazel - web sizeIsmaeil Fazel is a Lecturer at the Faculty of Education of Simon Fraser University. He has a PhD in TESL and a sub-specialization in Measurement and Evaluation from the University of British Columbia. His main research interests include English for Academic and Professional Purposes, academic discourse socialization, and language assessment. His publications have appeared in the Journal of English for Academic Purposes, English for Specific Purposes Journal, and TESL Canada Journal, among others.

Soroush Sabbaghan - web sizeSoroush Sabbaghan is a Senior Instructor at the Werklund School of Education at the University of Calgary. He holds two PhDs, one in TESL and the other in Curriculum and Learning with a focus in Mathematics Education. His main research interests include language and mathematics education of bilingual and multilinguals, Language Learning and Technology, and ecological complexity discourses. He has publications in both Language and mathematics education journals and books.

Friday, 08 January 2021

10:00 a.m. – 11:30 a.m. Calgary (Mountain) time

This session is open to the public and everyone is welcome. Registration required. Deadline to register is 07 Jan 2021.

More information and registration:

https://taylorinstitute.ucalgary.ca/series-and-events/academic-integrity-urgent-emerging-topics

This webinar is part of our series, Academic Integrity: Urgent and Emerging Topics. This series addresses timely and emergent topics that are cutting edge, provocative or high profile in nature. The series is hosted by the Taylor Institute for Teaching and Learning, University of Calgary.

_____

Share or Tweet this: Upcoming Webinar: None of the above: Integrity concerns of standardized English proficiency tests with Soroush Sabbaghan and Ismaeil Fazel

This blog has had over 2 million views thanks to readers like you. If you enjoyed this post, please “like” it or share it on social media. Thanks!

Sarah Elaine Eaton, PhD, is a faculty member in the Werklund School of Education, and the Educational Leader in Residence, Academic Integrity, University of Calgary, Canada. Opinions are my own and do not represent those of the University of Calgary.


Comparing E-Proctoring Software to Hydroxychloroquine: An Apt Analogy

November 4, 2020

Image courtesy of patrisyu at FreeDigitalPhotos.net

Image courtesy of patrisyu at FreeDigitalPhotos.net

To help educators and administrators understand why I urge caution, and even skepticism about the use of e-proctoring software and other surveillance technologies such as those that lockdown students’ Internet browsers, here’s an analogy I have been using that seems to resonate:

In my opinion, e-proctoring software is to higher education what Hydroxycloroquine has been to the COVID-19 virus.

It’s not that e-proctoring software is bad, it is that it was never designed to be used under the current conditions. There are colleagues who would disagree with me about this kind of software being bad in principle. I accept their position. Let’s look at this through the eyes of scholar who is trained to reserve judgement on an issue without evidence to back it up. If we assume the software was designed for a specific purpose – to invigilate exams taken via a computer, then it fulfills that purpose. So, in that sense, it does what it is supposed to do. However, that is not the whole story.

We can turn to Hydroxychloroquine as an analogy to help us understand why we should be skeptical.

Hydroxychloroquine is an anti-malaria drug, also used to treat arthritis. It was never designed to be used against the SARS-CoV-2 (COVID-19) virus. Hasty attempts to do research on the coronavirus, including studies on Hydroxychloroquine, have resulted in numerous papers now being retracted from scientific journals. People ran to this drug as a possible antidote the coronavirus, just as schools are running to e-proctoring software as an antidote for exam cheating. Neither e-proctoring software nor Hydroxychloroquine were designed to be used during the current pandemic. People flocked to them both as if they were some kind of magic pill that would solve a massively complex problem, without sufficient evidence that either would actually do what they so desperately wanted it to do.

The reality is that there is scant scientific data to show that e-proctoring actually works in the way that people want it to, that is, to provide a way of addressing academic misconduct during the pandemic. By “scientific data” I do not mean sales pitches. I am talking about independent scholarly studies undertaken by qualified academic researchers employed at reputable universities. By “independent scholarly studies” I mean research that has not been funded in any way by the companies that produce the products. That kind of research is terrifyingly lacking.

We need to back up for a minute and look about why we invigilate exams in the first place. To invigilate means “to keep watch over”. Keeping watch over students while they write an exam is part of ensuring that testing conditions are fair and objective.

The point of a test, in scientific terms, involves controlling all variables except one. In traditional testing, all other factors are controlled, including the conditions under which the test was administered such as the exam hall with desks separated, same lighting and environment for all test-takers, length of time permitted to take the test, how it is invigilated, and so on. All variables are presumably controlled except one: the student’s knowledge of the subject matter. That’s what’s being tested, the student’s knowledge.

Exams are administered in what could be termed, academically sterile environments. In an ideal situation, academic hygiene is the starting point for administering a test. Invigilation is just one aspect of ensuring academic hygiene during testing, but it is not the only factor that contributes to this kind of educational hygiene that we need to ensure testing conditions control for all possible variables except a student’s knowledge of the subject matter.

During the pandemic, with the shift to remote learning, we cannot control all the variables. We simply cannot assure an academically hygienic environment for testing. Students may have absolutely no control over who else is present in their living/studying quarters. They may have no control over a family member (including their own children) who might enter a room unannounced during a test. The conditions under which students are being tested during the pandemic are not academically hygienic. And that’s not their fault.

E-proctoring may address one aspect of exam administration: invigilation. It cannot, however, ensure that all variables are controlled.

As an academic integrity scholar, I am distressed by the lack of objective, peer-reviewed data about e-proctoring software. Schools have turned to e-proctoring software as if it were some kind of magic pill that will make academic cheating go away. We have insufficient evidence to substantiate that e-proctoring software, or any technology for that matter, can serve as a substitute for an in-person academically hygienic testing environment.

Schools that were using e-proctoring before the pandemic, such as Thompson Rivers University or Athabasca University in Canada, offered students a choice about whether students preferred to take their exams online, at home, using an e-proctoring service, or whether they preferred to drive to an in-person exam centre. During the pandemic, students’ choice has been taken away.

We all want an antidote to academic misconduct during remote learning, but I urge you educators and administrators to think like scholars and scientists. In other words, approach this “solution” with caution, and even skepticism. At present, we lack sufficient evidence to make informed decisions. Educators need to be just as skeptical about this technology and how it works during pandemic conditions as physicians and the FDA have been about using Hydroxychloroquine as a treatment for the coronavirus. Its use as being effective against the coronavirus is a myth. The use of e-proctoring software as being an effective replacement for in-person exams is also a myth, one perpetuated by the companies that sell the product.

Forcing surveillance technology on students against their will during a pandemic is tantamount to forcing an untested treatment on a patient; it is unethical to the extreme.

______

Share or Tweet this: Comparing E-Proctoring Software to Hydroxychloroquine: An Apt Analogy – https://drsaraheaton.wordpress.com/2020/11/04/comparing-e-proctoring-software-to-hydroxychloroquine-an-apt-analogy/(opens in a new tab)

This blog has had over 2 million views thanks to readers like you. If you enjoyed this post, please “like” it or share it on social media. Thanks!

Sarah Elaine Eaton, PhD, is a faculty member in the Werklund School of Education, and the Educational Leader in Residence, Academic Integrity, University of Calgary, Canada. Opinions are my own and do not represent those of the University of Calgary.