• Admiral Patrick@dubvee.org
    link
    fedilink
    English
    arrow-up
    100
    ·
    edit-2
    2 months ago

    My takeaway from this is:

    1. Get a bunch of AI-generated slop and put it in a bunch of individual .htm files on my webserver.
    2. When my bot user agent filter is invoked in Nginx, instead of returning 444 and closing the connection, return a random .htm of AI-generated slop (instead of serving the real content)
    3. Laugh as the LLMs eat their own shit
    4. ???
    5. Profit
    • mesamune@lemmy.world
      link
      fedilink
      English
      arrow-up
      38
      ·
      edit-2
      2 months ago

      I might just do this. It would be fun to write a quick python script to automate this so that it keeps going forever. Just have a link that regens junk then have it go to another junk html file forever more.

      • capital@lemmy.world
        link
        fedilink
        English
        arrow-up
        13
        ·
        2 months ago

        Also send this junk to Reddit comments to poison that data too because fuck Spez?

          • capital@lemmy.world
            link
            fedilink
            English
            arrow-up
            6
            ·
            2 months ago

            That’s a little different than what I mean.

            I mean to run a single bot from a script which interacts a normal human amount during normal human times within a configurable time zone which is acting as a real person just to poison their dataset.

  • paddirn@lemmy.world
    link
    fedilink
    English
    arrow-up
    12
    ·
    2 months ago

    So kinda like the human centipede, but with LLMs? The LLMillipede? The AI Centipede? The Enshittipede?

    • reinei@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      2 months ago

      No don’t listen to them!

      Keikaku means cake! (Muffin to be precise, because we got the muffin button!)

  • Chickenstalker@lemmy.world
    link
    fedilink
    English
    arrow-up
    4
    arrow-down
    3
    ·
    2 months ago

    Well then, here’s an idea for all those starving artists: start a business that makes AND sells human-made art/data to AI companies. Video yourself drawing the rare Pepe or Wojak from scratch as proof.

  • BetaDoggo_@lemmy.world
    link
    fedilink
    English
    arrow-up
    28
    arrow-down
    6
    ·
    edit-2
    2 months ago

    How many times is this same article going to be written? Model collapse from synthetic data is not a concern at any scale when human data is in the mix. We have entire series of models now trained with mostly synthetic data: https://huggingface.co/docs/transformers/main/model_doc/phi3. When using entirely unassisted outputs error accumulates with each generation but this isn’t a concern in any real scenarios.

    • Something Burger 🍔@jlai.lu
      link
      fedilink
      English
      arrow-up
      34
      ·
      2 months ago

      As the number of articles about this exact subject increases, so does the likelihood of AI only being able to write about this very subject.

  • leftzero@lemmynsfw.com
    link
    fedilink
    English
    arrow-up
    19
    ·
    2 months ago

    Anyone old enough to have played with a photocopier as a kid could have told you this was going to happen.

  • gravitas_deficiency@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    39
    arrow-down
    1
    ·
    2 months ago

    Imo this is not a bad thing.

    All the big LLM players are staunchly against regulation; this is one of the outcomes of that. So, by all means, please continue building an ouroboros of nonsense. It’ll only make the regulations that eventually get applied to ML stricter and more incisive.

  • kowcop@aussie.zone
    link
    fedilink
    English
    arrow-up
    4
    ·
    2 months ago

    I always thought this is why the Facebooks and Googles of the world are hoovering up the data now

  • jet@hackertalks.com
    link
    fedilink
    English
    arrow-up
    18
    arrow-down
    2
    ·
    2 months ago

    The best analogy I can think of:

    Imagine you speak English, and your dropped off in the middle of the Siberian forest. No internet, old days. Nobody around you knows English. Nobody you can talk to knows English. English for all intents purposes only exists in your head.

    How long do you think you could still speak English correctly? 10 years? 20 years? 30 years? Will your children be able to speak English? What about your grandchildren? At what point will your island of English diverge sufficiently from your home English that they’re unintelligible to each other.

    I think this is the training problem in a nutshell.