But that helps the student get better at overseeing the AI, not at writing or critical thinking. I don’t even think it would help them get better at analyzing writing. Most students would just turn in the first result, unless the teacher requires them turn in the whole transcript of their session and then the teacher is just grading them on AI overseeing. And that’s one case. Every bit of homework I ever did (before higher education) has been shown by ChatGPT: analyzing literature, writing in various styles, physics problems, foreign language translation, etc.

I doubt that AI will increase the percentage of students that want to cheat, and it is easy to cheat.

What? Cheating right now requires a human somewhere to do the work. A student can steal their work from online or pay someone they know to do it, but basic work will have available answers. A teacher may be able to come up with a unique format for their specific questions in their subject, but that takes extra work for the teacher. If a student can literally just type the question on the paper into ChatGPT and get an answer, I can’t imagine many teenage students not doing that, at least some of the time.

It sounds like you’re suggesting they rewrite their curriculums around ChatGPT (or similar AIs). That would require the teacher themselves to have a good understanding of the AI. So they have to gain proficiency in a brand new technology and then design a way of teaching around that. This is a ridiculous ask of a group of people who are already under-resourced and not keeping up with their current goals (at least in the U.S where I am).

That’s the main point of my previous post. It’s irresponsible and immoral to develop and release a technology for your own profit and just say everybody else needs to adapt to it.

Create a post

Subscribe to see more stories about technology on your homepage

  • 1 user online
  • 79 users / day
  • 111 users / week
  • 220 users / month
  • 531 users / 6 months
  • 14 subscribers
  • 1.25K Posts
  • Modlog