Via [**Rebecca Bultsma:**](https://www.linkedin.com/posts/rebecca-bultsma_educationtechnology-ai-edtech-activity-7214583758954381313-4I_n?utm_source=share&utm_medium=member_ios)
> A top student at Baltimore School for the Arts was accused of cheating based on AI detection software and I want to talk about it.
Quick landscape scan:
▶ Impact Research revealed this month that teachers, students, and parents say their schools have no AI policy and not enough support and training.
▶ Detection software isn’t foolproof – false positives happen all the time, and in this case, the GPTZero website clearly states, “results should not be used to directly punish students.”
▶ Students (and parents) are confused about what tools are allowed – in this instance, the students used Grammarly.
The questions to be asking in your school district:
▶ How do we balance academic integrity with the benefits of AI in learning? Where is the line?
▶ Are we punishing students for using tools we haven’t properly regulated?
▶ What’s the right way to implement AI policies in schools? Is anyone doing it WELL yet?
👉 My recommendations
▶ Develop clear AI use policies and OVERcommunicate them to staff, students, and parents.
▶ Educate students AND teachers on ethical AI use. Spell it out.
▶ If you use detection software, consider using it for discussion, not punishment
▶ It’s time to explore new assessment methods for the AI age
What would you add?
Discover more from Another Think Coming
Subscribe to get the latest posts sent to your email.