• @[email protected]
        link
        fedilink
        English
        2
        edit-2
        23 days ago

        No the chances of being wrong 10x in a row are 2%. So the chances of being right at least once are 98%.

        • Log in | Sign up
          link
          fedilink
          English
          224 days ago

          Ah, my bad, you’re right, for being consistently correct, I should have done 0.3^10=0.0000059049

          so the chances of it being right ten times in a row are less than one thousandth of a percent.

          No wonder I couldn’t get it to summarise my list of data right and it was always lying by the 7th row.

          • @[email protected]
            link
            fedilink
            English
            123 days ago

            That looks better. Even with a fair coin, 10 heads in a row is almost impossible.

            And if you are feeding the output back into a new instance of a model then the quality is highly likely to degrade.

            • Log in | Sign up
              link
              fedilink
              English
              123 days ago

              Whereas if you ask a human to do the same thing ten times, the probability that they get all ten right is astronomically higher than 0.0000059049.

              • @[email protected]
                link
                fedilink
                English
                123 days ago

                Dunno. Asking 10 humans at random to do a task and probably one will do it better than AI. Just not as fast.

                • Log in | Sign up
                  link
                  fedilink
                  English
                  1
                  edit-2
                  23 days ago

                  You’re better off asking one human to do the same task ten times. Humans get better and faster at things as they go along. Always slower than an LLM, but LLMs get more and more likely to veer off on some flight of fancy, further and further from reality, the more it says to you. The chances of it staying factual in the long term are really low.

                  It’s a born bullshitter. It knows a little about a lot, but it has no clue what’s real and what’s made up, or it doesn’t care.

                  If you want some text quickly, that sounds right, but you genuinely don’t care whether it is right at all, go for it, use an LLM. It’ll be great at that.

        • 𝕛𝕨𝕞-𝕕𝕖𝕧
          link
          fedilink
          English
          124 days ago

          don’t you dare understand the explicitly obvious reasons this technology can be useful and the essential differences between P and NP problems. why won’t you be angry >:(