Zubnet AI學習Wiki › World Model
Models

World Model

Internal World Model, 學習ed Simulator
一個建立世界如何運作的內部表示的模型 — 不只是統計相關,還有因果關係、物理定律、空間推理。LLM 是否有 world model 的辯論是 AI 中最具爭議的之一:它們是真的理解物體被放下會落下,還是只知道「落下」在文字中經常跟在「放下」之後?

為什麼重要

World model 是 AI 中最重要問題的核心:理解是否需要超越模式匹配?如果 LLM 建立真正的 world model,它們比我們以為的更接近理解。如果沒有,有一個單靠 scaling 不會彌合的根本能力差距。答案對 AI 安全、能力、通向更一般智能的路徑都有巨大含義。

Deep Dive

Evidence that LLMs may build world models: they can play chess (requiring spatial reasoning), solve novel physics problems, generate working code for described algorithms (requiring causal reasoning about program execution), and navigate text-based worlds consistently. Research by Li et al. (2023) showed that a model trained only on Othello game transcripts developed an internal representation of the board state — a literal world model emerging from sequence prediction.

Evidence Against

LLMs make errors that suggest pattern matching rather than understanding: they struggle with spatial reasoning ("I walk north, then east, then south — where am I relative to the start?"), fail at novel physical reasoning (situations not in training data), and can be tripped up by simple modifications to familiar problems (changing numbers in a math problem they solved correctly in standard form). These failures suggest the model learned surface patterns, not underlying mechanisms.

The Middle Ground

The emerging view: LLMs build partial, approximate world models that work well for common situations but break down at the edges. They learn useful representations of how the world works — good enough for most text generation tasks — but these representations are incomplete, inconsistent, and not grounded in actual physical experience. Whether this constitutes "understanding" depends on your definition. What's practical: LLM world models are useful but shouldn't be trusted for safety-critical physical reasoning without verification.

相關概念

← 所有術語
← Word Embedding xAI →