• Kyrgizion@lemmy.world
    link
    fedilink
    English
    arrow-up
    17
    ·
    4 days ago

    That might be a while. AI cannibalizing itself is a real problem right now and it’s only going to get worse.

      • Neshura@bookwormstory.social
        link
        fedilink
        English
        arrow-up
        15
        ·
        4 days ago

        pretty much, AI (LLMs specifically) are just fancy statistical models which means that when they ingest data without reasoning behind it (think the many hallucinations of AI our brains manage to catch and filter out) it corrupts the entire training process. The problem is that AI can not distinguish other AI text from human text anymore so it just ingests more and more “garbage” which leads to worse results. There’s a reason why progress in the AI models has almost completely stalled compared to when this craze first started: the companies have an increasingly hard time actually improving the models because there is more and more garbage in the training data.

        • oce 🐆@jlai.lu
          link
          fedilink
          English
          arrow-up
          14
          ·
          edit-2
          4 days ago

          There’s actually a lot of human intervention in the mix. Data labelers for source data, also domain experts who will rectify answers after a first layer of training, some layers of prompts to improve common answers. Without those domain experts, the LLM would never have the nice looking answers we are getting. I think the human intervention is going to increase to counter the AI pollution in the data sources. But it may not be economically viable anymore eventually.

          This is a nice deep dive of the different steps to make today’s LLMs: https://youtube.com/watch?v=7xTGNNLPyMI

    • Kokesh@lemmy.world
      link
      fedilink
      English
      arrow-up
      8
      ·
      4 days ago

      Let’s hope the current ai chokes on the crap it produces and eats afterwards.

        • Kokesh@lemmy.world
          link
          fedilink
          English
          arrow-up
          6
          ·
          4 days ago

          Yup. Me on my quite big karma account on Stack Overflow: I gave up on it when they decided to sell my answers/questions for AI training. First I wanted to delete my account, but my data would stay. So I started editing my answers to say “fuck ai” (in a nutshell). I got suspended for a couple months " to think about what I did". So I dag deep into my consciousness and came up with a better plan. I went through my answers (and questions) and poisoned them little by little every day bit by bit with errors. After that I haven’t visited that crap network anymore. Before all this I was there all the time, had lots of karma (or whatever it was called there). Couldn’t care less after the AI crap. I honestly hope, that I helped make the AI, that was and probably still is trained on data that the users didn’t consent to be sold, little bit more shitty.