Just a trillion more, bro!

  • peoplebeproblems@midwest.social
    link
    fedilink
    English
    arrow-up
    25
    ·
    5 days ago

    It’s a fucking chatbot that used modern ML training methods on enormous datasets so it’s slightly fancier than the ones that already existed.

    They just fed it so much data that it almost appears like it knows anything, when all it does is respond to the words you give it.

    • Knock_Knock_Lemmy_In@lemmy.world
      link
      fedilink
      arrow-up
      7
      ·
      5 days ago

      They just fed it so much data that it almost appears like it knows anything,

      The strawberry test shows this. Ask directly and it will give the correct number of letters. Ask in a more indirect fashion (in a way unlikely to be in the training set) and it falls over like before.

    • npdean@lemmy.todayOP
      link
      fedilink
      arrow-up
      4
      arrow-down
      1
      ·
      5 days ago

      People need to realise this point. Difference between previous models and new ones are so much dependent on the amount of data it has eaten.

      • vrighter@discuss.tchncs.de
        link
        fedilink
        arrow-up
        5
        ·
        5 days ago

        and the seed lottery. You can see this if you try training a simple network with two inputs to learn xor. It can converge in multiple ways, and sometimes it converges to a really bad approximation. And sometimes it doesn’t converge at all (or it converges so slowly that it might as well be considered not to converge). And even then it might still converge to an approximation that’s more accurate on one side of the input space than the other. Tons of ways to get an undesirable result. For a simple 2-input network.

        Imagine how unlikely it is for txese models to actually converge to the optimal thing. And how often the training is for nothing.

          • vrighter@discuss.tchncs.de
            link
            fedilink
            arrow-up
            6
            ·
            5 days ago

            yes, but the only way to weed out a bad seed is to “play the lottery”

            By the time you discover a seed is bad, you’ve already spent a shitton on training. Money down the drain, you gotta start over

            • npdean@lemmy.todayOP
              link
              fedilink
              arrow-up
              3
              ·
              5 days ago

              Just train another AI to train AI, then train another and another, and another. Imagine the stock rally