This is a cache of https://slashdot.org/story/24/11/10/1911204/generative-ai-doesnt-have-a-coherent-understanding-of-the-world-mit-researchers-find. It is a snapshot of the page at 2024-11-11T01:09:52.563+0000.
Generative AI Doesn't Have a Coherent Understanding of the World, MIT Researchers Find - Slashdot

Become a fan of Slashdot on Facebook

 



Forgot your password?
typodupeerror
×
AI

Generative AI Doesn't Have a Coherent Understanding of the World, MIT Researchers Find (mit.edu) 33

Long-time Slashdot reader Geoffrey.landis writes: Despite its impressive output, a recent study from MIT suggests generative AI doesn't have a coherent understanding of the world. While the best-performing large language models have surprising capabilities that make it seem like the models are implicitly learning some general truths about the world, that isn't necessarily the case. The recent paper showed that Large Language Models and game-playing AI implicitly model the world, but the models are flawed and incomplete.

An example study showed that a popular type of generative AI model accurately provided turn-by-turn driving directions in New York City, without having formed an accurate internal map of the city. Though the model can still navigate effectively, when the researchers closed some streets and added detours, its performance plummeted. And when they dug deeper, the researchers found that the New York maps the model implicitly generated had many nonexistent streets curving between the grid and connecting far away intersections.

Generative AI Doesn't Have a Coherent Understanding of the World, MIT Researchers Find

Comments Filter:
  • No kidding (Score:5, Insightful)

    by alvinrod ( 889928 ) on Sunday November 10, 2024 @05:41PM (#64935477)
    LLMs don't have an understanding of anything. They can only regurgitate derivations of what they've been trained on and can't apply that to something new in the same ways that humans or even other animals can. The models are just so large that the illusion is impressive.
  • by ls671 ( 1122017 ) on Sunday November 10, 2024 @05:41PM (#64935479) Homepage

    Seriously, did we need a MIT study to know that?

    • And if it was MIT, why didn't they first try it out on a map of Boston [app.goo.gl]?
      • And if it was MIT, why didn't they first try it out on a map of Boston [app.goo.gl]?

        Probably because the city of NY (Manhattan island especially) is much simpler road-wise than Boston.

    • by narcc ( 412956 )

      Apparently. There are a surprising number of people, even in the field, who have what I can only describe as religious beliefs about emergence. It's disturbing.

  • understanding? (Score:5, Insightful)

    by dfghjk ( 711126 ) on Sunday November 10, 2024 @05:41PM (#64935483)

    More anthropomorphizing neural networks. They don't have "understanding" at all, much less "coherent" understanding.

  • ...CaptObviousGPT

    Very few experts ever claimed it had common sense-like reasoning, and those who did usually added caveats to their claims.

  • I have met lots of people who don't have a coherent understanding of the world. This week I watched them ... oh, never mind.

    • The expectation we have that LLMs should somehow be perfect is an expectation we don't tend to apply to humans.
      If you ask a human a complex question it is likely the human will have a less than 100% understanding of the issue, and will give a less than 100% correct answer. But that's ok. We can continue the conversation, challenging the dubious parts, and working towards the truth.
      Perhaps as this sort of AI progresses, systems like ChatGPT might become better at recognising their own weak areas of knowledge

  • I'm wondering if it would be possible to hook it up to the likes of Cyc, a logic engine and common-sense-rules-of-life database. The engine could find the best match between the language model (text) and Cyc models, weighting to favor shorter candidates (smallest logic graph). Generating candidate Cyc models from language models may first require a big training session itself.

    I just smell value in Cyc's knowledge-base, there's nothing on Earth comparable (except smaller clones). Wish I could by stock in it

    • Corrections:

      "weighing to favor shorter candidates" [No JD jokes, please]

      "Wish I could buy stock in it"

      (Bumped the damned Submit too early)

    • It's amazing how many startups are out there just repeating the same LLM approach with more data, but none (afaik) are trying something like joining it with Cyc. If I were raising billions for an AI startup, I would consider at least trying that as a side project.
      • by Tablizer ( 95088 )

        Indeed. With all the investing going into increasingly questionable AI projects you'd think somebody with money would zig when everyone else is zagging to try bagging a missed solution branch/category.

        Reminds me of the Shuji Nakamura story on the invention of a practical blue LED. Red and green LED's were already commercial viable. Blue was the missing "primary light color" in order to mix to get the full rainbow. Many big co's spent a lot of R&D on blue, but kept failing. Their blue LED's were just way

      • Aside from being novelty-addled herd animals; I think that there's a much stronger cultural affinity for the technology that is all about the fact that you can sometimes get surprisingly plausible outputs from nescience so profound that it would be anthropomorphizing to call it ignorance; than for the technology founded on the hope that if you systematically plug away at knowing enough you might eventually be rewarded by competent outputs.
    • I had a similar thought. Expert systems are good at some things; LLMs are good at others. We need to combine them. LLMs are superb at converting unstructured input to structured input. There has to be something there.
  • Cool approach (Score:4, Interesting)

    by phantomfive ( 622387 ) on Sunday November 10, 2024 @05:58PM (#64935525) Journal
    Of course, everyone knows these models hallucinate. The question is, what is going on inside the model to make it hallucinate? (Or alternately, what is it doing to be right so often?). Once you can figure out what's going on inside the model, then you can improve it. Actually a lot of work has been done in this area, so they are just adding to it. From the article:

    'These results show that transformers can perform surprisingly well at certain tasks without understanding the rules. If scientists want to build LLMs that can capture accurate world models, they need to take a different approach, the researchers say.'

    The key thing here is they don't understand the rules. For example, an AI model might make legal chess moves every time, but if you modified the chess board [wikipedia.org] then it would suddenly make illegal moves with the knight. With current AI technology, you would try to "fix" this by including as many possible different chess boards as possible, but that's not how humans think. We know the rules of the knight and recognize that in a new situation, changing the board doesn't change the way the knight moves (but it might). And if you wanted to clarify,, you could ask someone, "Do all the pieces still move the same on this new board?", but that is what these researchers did (modified the map of NY with a detour), and it really confused the model.

    It is of course obvious that current LLMs do not have human intelligence because they are not Turing complete, but to understand what that means you'd need to have an internal understanding and mental model of what Turing machines are, and LLMs don't have that. :)

    • Re:Cool approach (Score:4, Interesting)

      by iAmWaySmarterThanYou ( 10095012 ) on Sunday November 10, 2024 @06:04PM (#64935545)

      These things are really good at sussing out patterns in (seemingly) random data. It's an extremely useful feature/ability.

      But of course they're lost when the rules and patterns change because they mastered the initial patterns already. They have no ability to think and realize, "oh, this is a new thing" because they don't ever "realize" anything. Change the chess board, the fake-I fails.

      I assume the hallucinations come from finding patterns that really weren't there but only appeared to be based on previous training.

      All I got from this silly MIT study was some poor bastards stuck in the world of publish or perish who were drinking and partying too much right up to their publish deadline then put out some bullshit to keep their careers alive for another season.

      • But of course they're lost when the rules and patterns change because they mastered the initial patterns already.

        That's part of it, but also they have no ability to recognize what changes are "important" and what are not. Change the color of the chess piece from white to ivory and it won't recognize it (unless it has ivory in its training set).

      • You are probably right about the source of the hallucinations. The world is full of meaningless statistical correlations that happen by pure chance. If you randomly look for correlations among large sets of data, you will inevitably find many false ones. These models cannot distinguish between true correlations and false ones. Of course, people are not very good at this either, or Las Vegas would not be making so much money.
    • It is of course obvious that current LLMs do not have human intelligence because they are not Turing complete, but to understand what that means you'd need to have an internal understanding and mental model of what Turing machines are, and LLMs don't have that. :)

      You are correct they don't have human intelligence yet wrong about the reason why.

      "Memory Augmented Large Language Models are Computationally Universal"
      https://arxiv.org/pdf/2301.045... [arxiv.org]

      The key thing here is they don't understand the rules. For example, an AI model might make legal chess moves every time, but if you modified the chess board then it would suddenly make illegal moves with the knight. With current AI technology, you would try to "fix" this by including as many possible different chess boards as possible,

      Have you tried asking the AI to convert to a normal chess board before moving?

      but that's not how humans think.

      LLMs obviously don't work like humans.

    • by narcc ( 412956 )

      The question is, what is going on inside the model to make it hallucinate?

      So-called 'hallucinations' are not errors or mistakes, they are a natural and expected result of the how these models function.

      The key thing here is they don't understand the rules.

      They don't understand anything. That's not how they work. They don't operate on facts and concepts, they operate on statistical relationships between tokens.

  • They are tools. Very useful tools, if you understand how to use them and what to use them for.
  • Anyone who thought this was even possible doesn't have a coherent understanding of the world.

  • as humans that grow up in the world, we hardly have a coherent understanding of the world, let alone an LLM that is trained on huge data sets and forms patterns to mimic a form of intelligence.

    Sounds like the people that wasted their time on this don't have a coherent understanding of the world either. Marketing wank is just that, were they expecting actual Intelligence?

  • This is all rather interesting. People create systems inspired by how brains work then they turn around and get all upset it isn't perfect and criticize the system for its failure to magically compile and execute some kind of robust model of how the world works that would enable it to always generate infallible predictions.

    On one hand we have people who either hate with a passion or dismiss LLMs outright as cut and paste machines which don't even deserve to be called AI. On the other hand we have people r

  • LLM's will have achieved true intelligence when they answer the question with <Maine_accent>Ya' can't get thea' from hea'</Maine_accent>.

"You must have an IQ of at least half a million." -- Popeye

Working...