Zubnet AI学习Wiki › World Model
Models

World Model

Internal World Model, 学习ed Simulator
一个建立世界如何运作的内部表示的模型 — 不只是统计相关,还有因果关系、物理定律、空间推理。LLM 是否有 world model 的辩论是 AI 中最具争议的之一:它们是真的理解物体被放下会落下,还是只知道“落下”在文本中经常跟在“放下”之后?

为什么重要

World model 是 AI 中最重要问题的核心:理解是否需要超越模式匹配?如果 LLM 建立真正的 world model,它们比我们以为的更接近理解。如果没有,有一个单靠 scaling 不会弥合的根本能力差距。答案对 AI 安全、能力、通向更一般智能的路径都有巨大含义。

Deep Dive

Evidence that LLMs may build world models: they can play chess (requiring spatial reasoning), solve novel physics problems, generate working code for described algorithms (requiring causal reasoning about program execution), and navigate text-based worlds consistently. Research by Li et al. (2023) showed that a model trained only on Othello game transcripts developed an internal representation of the board state — a literal world model emerging from sequence prediction.

Evidence Against

LLMs make errors that suggest pattern matching rather than understanding: they struggle with spatial reasoning ("I walk north, then east, then south — where am I relative to the start?"), fail at novel physical reasoning (situations not in training data), and can be tripped up by simple modifications to familiar problems (changing numbers in a math problem they solved correctly in standard form). These failures suggest the model learned surface patterns, not underlying mechanisms.

The Middle Ground

The emerging view: LLMs build partial, approximate world models that work well for common situations but break down at the edges. They learn useful representations of how the world works — good enough for most text generation tasks — but these representations are incomplete, inconsistent, and not grounded in actual physical experience. Whether this constitutes "understanding" depends on your definition. What's practical: LLM world models are useful but shouldn't be trusted for safety-critical physical reasoning without verification.

相关概念

← 所有术语
← Word Embedding xAI →