Did your students use ChatGPT or other LLM tools to cheat on their assignments? If you taught a philosophy course this term, please respond to the one-question poll, below.
It would be useful to hear from professors about their experiences with this. Here are some further questions for discussion:
- If you did not detect any of your students cheating using ChatGPT or the like, was that because of something you said to or did with them, or because you intentionally crafted your assignments in a way resistant to their use, or did you just get lucky, or did you not bother trying to detect their use, or what?
- If you did try to craft ChatGPT-resistant assignments, what were they?
- Did you use AI-detection software to find instances of cheating, and if so, which did you use, and how well did it work, in your opinion?
- Did you spot the AI-written work without the assistance of detection software, and if so, what did you look for?
- Did you report students who cheated using this technology to official academic misconduct offices at your university or college, and if so, how receptive to (and prepared for) this kind of cheating were they?
- As more and more students learn of these tools and how to use them, do you think your methods for handling the issue this term will work next term?
- What would you do differently?
Thanks for sharing your thoughts on this.