• ganymede@lemmy.ml
    link
    fedilink
    arrow-up
    2
    ·
    2 years ago

    wonder what it’s false positive rate is, and how that will be handled for sensitive issues like university degree work etc

    • d-RLY?@lemmy.ml
      link
      fedilink
      arrow-up
      2
      ·
      2 years ago

      I get the feeling that we will see a bunch of snake oil that claim to do this stuff. Especially in the education sectors, like the ones you are talking about. And given how much money higher education throws around for at least giving the image of protections and other theatre. It will be like all the PC “tune-up” programs that claim to be doing stuff to help, but just run things the OS already has or slows things down.

      That all being said. As long as the tools for AI are made to be open and auditable. Then it could be helpful in giving a starting point for actual professors to double-check. But I also worry that many professors (and other folks) will only go with the AI answer and not bother to look any deeper. Since the hype-people for stuff like AI tend to do the same things that hype-people for other industries. They will constantly play up everything and make it out to be so much more capable that it actually is at the time.

      I also worry that some false positives will come from students learning to write things in similar ways as the AI. People do often (IMO at least) seem to emulate stuff they interact with often. They see examples of stuff that is well written/done and try to copy the styles (because they want to get a high grade). Even if they aren’t students, folks that are really focused on “vibes” also try to copy things that they see getting results. Which worries me given how much style over substance is focused on in school and in the business worlds. AI could be a “fake it till you make it” person’s absolute best friend.

      This stuff is going to be frustrating and difficult to figure out no matter what side you are on for sure.