How an artificial intelligence (as in large language model based generative AI) could be better for information access and retrieval than an encyclopedia with a clean classification model and a search engine?

If we add a step of processing – where a genAI “digests” perfectly structured data and tries, as bad as it can, to regurgitate things it doesn’t understand – aren’t we just adding noise?

I’m talking about the specific use-case of “draw me a picture explaining how a pressure regulator works”, or “can you explain to me how to code a recursive pattern matching algorithm, please”.

I also understand how it can help people who do not want or cannot make the effort to learn an encyclopedia’s classification plan, or how a search engine’s syntax work.

But on a fundamental level, aren’t we just adding an incontrolable step of noise injection in a decent time-tested information flow?

  • fxdave@lemmy.ml
    link
    fedilink
    arrow-up
    2
    ·
    edit-2
    5 days ago

    But LLMs are not simply probabilistic machines. They are neural nets. For sure, they haven’t seen the world. They didn’t learn the way we learn. What they mean by a caterpillar is just a vector. For humans, that’s a 3D, colorful, soft object with some traits.

    You can’t expect that a being that sees chars and produces chars knows what we mean by a caterpillar. Their job is to figure out the next char. But you could expect them to understand some grammar rules. Although, we can’t expect them to explain the grammar.

    For another example, I wrote a simple neural net, and with 6 neurons it could learn XOR. I think we can say that it understands XOR. Can’t we? Or would you say then that an XOR gate understands XOR better? I would not use the word understand for something that cannot learn. But why wouldn’t we use it for a NN?

    • Solumbran@lemmy.world
      link
      fedilink
      arrow-up
      1
      ·
      5 days ago

      Your whole logic is based on the idea that being able to do something means understanding that thing. This is simply wrong.

      Humans feel emotions, yet they don’t understand them. A calculator makes calculations, but no one would say that it understands math. People blink and breathe and hear, without any understanding of it.

      The concept of understanding implies some form of meta-knowledge about the subject. Understanding math is more than using math, it’s about understanding what you’re doing and doing it out of intention. All of those things are absent in an AI, neural net or not. They cannot “see the world” because they need to be programmed specifically for a task to be able to do it; they are unable to actually grow out of their programming, which is what understanding would ultimately cause. They simply absorb data and spit it back out after doing some processing, and the fact that an AI can be made to produce completely incompatible results shows that there is nothing behind it.

      • fxdave@lemmy.ml
        link
        fedilink
        arrow-up
        1
        ·
        edit-2
        2 days ago

        The concept of understanding implies some form of meta-knowledge about the subject.

        That can be solved if you teach it the meta-knowledge with intermediary steps, for example:

        prompt: 34*3=
        
        step1: 4*3 + 30*3 = 
        step2: 12 + 10*3*3 = 
        step3: 12 + 10*9=
        step4: 12 + 90 =
        step5: 100 + 2 =
        step6: 102
        
        result: 102
        

        It’s hard to find such learning data though, but e.g. claude already uses intermediary steps. It preprocesses your input multiple times. It writes code, runs code to process your input, and that’s still not the final response. Unfortunately, it’s already smarter than some junior developers, and its consequence is worrying.