• Veraxus@kbin.social
    link
    fedilink
    arrow-up
    32
    arrow-down
    3
    ·
    edit-2
    1 year ago

    Well yeah - because that’s not how LLMs work. They generate sentences that conform to the word-relationship statistics that were generated during the training (e.g. making comparisons between all the data the model was trained on). It does not have any kind of logic and it does not know things. It literally just navigates a complex web of relationships between words using the prompt as a guide, creating sentences that look statistically similar to the average of all trained sentences.

    TL;DR; It’s an illusion. You don’t need to run experiments to realize this, you just need to understand how AI/ML works.

    • Chetzemoka@startrek.website
      link
      fedilink
      arrow-up
      5
      arrow-down
      2
      ·
      1 year ago

      Tell that to all the tech bros on the internet are convinced that ChatGPT means AGI is just around the corner…

  • Buffalox@lemmy.world
    link
    fedilink
    arrow-up
    13
    ·
    edit-2
    1 year ago

    Very interesting, this seems so incredibly stupid it’s hard to believe it’s true.

    It’s amazing how far AI has come recently, but also kind of amazing how far away we still are from a truly general AI.

    • snooggums@kbin.social
      link
      fedilink
      arrow-up
      7
      arrow-down
      1
      ·
      1 year ago

      That is because the I in AI as currently used is as literal as the word hover in Hoverboard. You know, those things that don’t hover, just catch on fire.

      There is no intelligence in AI.

      • Buffalox@lemmy.world
        link
        fedilink
        arrow-up
        4
        ·
        1 year ago

        I strongly disagree, remember intelligence does not require consciousness, when we have that, it’s called strong AI or (AGI) artificial general intelligence.

        AI really has been making huge progress the past 10 years, probably equivalent to all the time that goes before.

  • phario@lemmy.ca
    link
    fedilink
    arrow-up
    11
    arrow-down
    1
    ·
    edit-2
    1 year ago

    These things are interesting for two reasons (to me).

    The first is that it seems utterly unsurprising that these inconsistencies exist. These are language models. People seem to fall easily into the trap in believing them to have any kind of “programming” on logic.

    The second is just how unscientific NN or ML is. This is why it’s hard to study ML as a science. The original paper referenced doesn’t really explain the issue or explain how to fix it because there’s not much you can do to explain ML(see their second paragraph in the discussion). It’s not like the derivation of a formula where you point to one component of the formula as say “this is where you go wrong”.

    • OhNoMoreLemmy@lemmy.ml
      link
      fedilink
      arrow-up
      1
      ·
      1 year ago

      It’s actually getting more scientific. Think of it like biology. We do a big study of an ml model or an organism and confirm a property of it.

      It used to be it was just maths, you could spot an error in your code and fix it. Then it was a bag of hacks and you could keep just patching your model with more and more tweaks that didn’t have a solid theoretical basis but that improved performance.

      Now it’s too big and too complex and we have to do science to understand the model limitations.