• zeropointone@lemmy.world
    link
    fedilink
    English
    arrow-up
    16
    arrow-down
    5
    ·
    3 days ago

    As if a stochastic parrot could ever be able to count or calculate…

    LLM work really well. You get something out of it that resembles human language. This is what it has been designed for and nothing else. Stop trying to make a screwdriver shoot laser beams, it’s not going to happen.

    • theunknownmuncher@lemmy.world
      link
      fedilink
      English
      arrow-up
      7
      arrow-down
      9
      ·
      edit-2
      3 days ago

      As if a stochastic parrot could ever be able to count or calculate…

      https://www.anthropic.com/research/tracing-thoughts-language-model

      well that’s provably false as the neural network can be traced to understand the exact process and calculations that it follows when doing math. See the above research under heading “Mental Math”.

      The reason LLMs struggle to count letters is because of tokenization. The text is first converted to tokens, numeric vectors which represents whole words or parts of words, before being given to the LLM, and the LLM has no concept of or access to the underlying letters that make up the words. The LLM outputs only tokens, which are converted back into text.

      EDIT: you can downvote me, but facts are facts lol.

      • jacksilver@lemmy.world
        link
        fedilink
        English
        arrow-up
        5
        ·
        3 days ago

        There is definitely more going on with LLMs than just direct parroting.

        However, there is also an upper limit to what an LLM can logic/calculate. Since LLMs basically boil down to a series of linear operations, there is an upper bound on all of them as to how accurately they can calculate anything.

        Most chat systems use python behind the scene for any actual math, but if you run a raw LLM you can see the errors grow faster as you look at higher orders of growth (addition, multiplication, powers, etc.).

        • theunknownmuncher@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          2
          ·
          edit-2
          3 days ago

          Yes, exactly. It can do basic math and also doesn’t really matter because it is really good at interfacing with tools/functions for calculation anyway

          However, there is also an upper limit to what an LLM can logic/calculate. Since LLMs basically boil down to a series of linear operations, there is an upper bound on all of them as to how accurately they can calculate anything.

          Also this is only true when LLMs are not using Chain of Thought.

        • theunknownmuncher@lemmy.world
          link
          fedilink
          English
          arrow-up
          4
          arrow-down
          2
          ·
          edit-2
          3 days ago

          Reading comprehension. The only thing it cannot count are specifically letters of tokenized words. If you separate the letters out into separate tokens (“B L U E B E R R Y”, etc) it will get it correct 100% of the time.

      • zeropointone@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        2
        ·
        3 days ago

        What makes you think that using single letters as tokens instead could teach a stochastic parrot to count or calculate? Both are abilities. You can’t create an ability only from a set of data no matter how much data you have. You can only make a model seem to have that ability. Again: All you can ever get out of it is something that resembles human language. There is nothing beyond/behind that, by design. Not even hallucinations. Whenever a LLM gives you the advice to eat a rock per day it still works. Because it outputs a correct sounding sentence purely and entirely based on probability. But counting and calculating are not based on probability which is something everyone who ever had a math class knows very well. No math teacher will let students guess the result of an equation.

        • theunknownmuncher@lemmy.world
          link
          fedilink
          English
          arrow-up
          3
          arrow-down
          1
          ·
          3 days ago

          What makes you think that using single letters as tokens instead could teach a stochastic parrot to count or calculate? Both are abilities. You can’t create an ability only from a set of data no matter how much data you have. You can only make a model seem to have that ability.

          Yeah, that’s just not how neural networks work…

          No math teacher will let students guess the result of an equation.

          https://en.wikipedia.org/wiki/Regression_analysis

          You tried.

            • theunknownmuncher@lemmy.world
              link
              fedilink
              English
              arrow-up
              5
              arrow-down
              1
              ·
              edit-2
              3 days ago

              https://en.wikipedia.org/wiki/Universal_approximation_theorem Actual math directly contradicts your beliefs. I know its trendy and you want to feel smarter than people who have spent literally decades researching NNs and staked billions of dollars developing it, but you’re wrong.

              Your claim is like claiming that boolean circuits cannot do math because all they do is “true” or “false”.

              And also, https://en.wikipedia.org/wiki/Recurrent_neural_network#Neural_Turing_machines recurrent neural networks are Turing complete when paired with memory, and therefore can be used to any calculations or computations that a conventional computer can do.

              • zeropointone@lemmy.world
                link
                fedilink
                English
                arrow-up
                1
                arrow-down
                2
                ·
                3 days ago

                And you don’t know what a circular argument is either…

                No, 2+2 is never “about 4” nor is it 4 in most cases. It’s always exactly 4. And no LLM can ever come to this conclusion. LLMs fail at math in a truly spectacular way. Just like no LLM will ever be able to understand what complementary colors are. Which is one of my favorite tests because it has a 100 % error rate.

                • theunknownmuncher@lemmy.world
                  link
                  fedilink
                  English
                  arrow-up
                  3
                  arrow-down
                  1
                  ·
                  edit-2
                  3 days ago

                  Just like no LLM will ever be able to understand what complementary colors are. Which is one of my favorite tests because it has a 100 % error rate.

                  LOL

                  The funniest part of this is not the fact that an LLM just got 3 for 3 correct, and therefore has a 100% success rate, thus proving you wrong again, but the fact that your “favorite test” would be one that you incorrectly believe “no LLM will ever be able to” do because…

                  Stop trying to make a screwdriver shoot laser beams, it’s not going to happen.

                  ^ this you??? “My favorite test is to see if the screwdriver shoots laser beams” 🙃

                  • zeropointone@lemmy.world
                    link
                    fedilink
                    English
                    arrow-up
                    1
                    arrow-down
                    2
                    ·
                    3 days ago

                    And why didn’t you include the name of the model in your test? Looks like you don’t want me to try it myself. It would be interesting to do so. Of course with values which don’t fit perfectly into 8 bit. What if I define the range from 0 to 47204 for each color channel instead? What if I would use CMY(K) instead of RGB? A good “great” AI must be able to handle all of that. And of course correctly explain what complementary colors are (which you didn’t include either). So yeah - what you provided does not go beyond the output from htmlcolorcodes.com - a very simple website with very simple code. I doubt it requires much power either.

                  • zeropointone@lemmy.world
                    link
                    fedilink
                    English
                    arrow-up
                    1
                    arrow-down
                    4
                    ·
                    3 days ago

                    Funny. Each time I ask any LLM what the complementary color to red is. Then I always get green as answer instead of cyan (With cyan being the only correct answer). And a completely wrong explanation about what complementary colors are based on digital screens. So yeah - LLMs still fail miserably at language-based tests. And rearranging complex equations doesn’t work either.