• dack@lemmy.world
    link
    fedilink
    English
    arrow-up
    7
    ·
    1 year ago

    At a very high level, training is something like:

    • generate some output
    • give the output a score based on how much it looks like real human text
    • adjust the parameters slightly to improve the score
    • repeat

    Step #2 is also exactly what an “AI detector” does. If someone is able to write code that reliably distinguishes between AI and human text, then AI developers would plug it in to that training step in order to improve their AI.

    In other words, if some theoretical machine perfectly “knows” the difference between generated and human text, then the same machine can also be used to make text that is indistinguishable from human text.

    • stevedidWHAT@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      ·
      edit-2
      1 year ago

      Exactly right, I mentioned this in a comment elsewhere but basically we can’t have our cake and eat it too.

      We can’t have a perfect NL impersonator that can also be detected as not NL. (Best case, obviously things arent perfect for any AI model so technically detecting those mistakes could be used to help identify perhaps, but who’s to say what the FP rate would look like!)

      Ultimately the cat is out of the bag and I’m not quite sure there is anything we can do now. Ultimately some smart fingerprinting solution would be ideal but I just don’t know how feasible that would remain.

      Edit: source: I took a few 600 level ai classes in college and have made several of my own of varying types and what not