• celsiustimeline@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    12
    arrow-down
    5
    ·
    edit-2
    2 months ago

    If mainstream blogs are writing about it, what would make someone think that AI companies haven’t thoroughly dissected the problem and are already working on filtering out AI fingerprints from the training data set? If they can make a sophisticated LLM, chances are they can find methods to XOR out generated content.

    • aesthelete@lemmy.world
      link
      fedilink
      English
      arrow-up
      10
      arrow-down
      1
      ·
      2 months ago

      What would make me think that they haven’t “thoroughly dissected” it yet is that I’m a skeptic, and since I’m a skeptic I don’t immediately and without evidence believe that every industry is capable of identifying, dissecting, and solving every problem with its products.

      • Schadrach@lemmy.sdf.org
        link
        fedilink
        English
        arrow-up
        2
        ·
        2 months ago

        Ironically given their skillset, training an ML model on known and properly tagged AI generated and non-AI-generated stuff might actually work.