Zubnet AI學習Wiki › Feature
基礎

Feature

學習ed Representation, Activation
神經網路學習在輸入中偵測的模式或概念。視覺中,早期層特徵是邊緣和紋理;後期層特徵是物體部分和整體物體。語言模型中,特徵範圍從簡單(字母「a」、特定句法模式)到抽象(諷刺的概念、特定推理策略)。特徵被表示為跨神經元的激活模式。

為什麼重要

特徵是模型實際學到的東西 — 不是個別事實而是能泛化的模式。模型不記憶「貓有毛」;它學到一個對皮毛紋理的特徵偵測器,對貓、狗、泰迪熊都激活。理解特徵幫助解釋模型行為:為什麼它泛化(特徵遷移)、為什麼失敗(錯誤特徵激活)、以及如何改進它(讓它接觸更多多樣特徵)。

Deep Dive

The term "feature" has different meanings depending on context. In classical ML, features are hand-engineered input variables (height, weight, age). In deep learning, features are learned representations in hidden layers — the model discovers useful patterns on its own. This shift from hand-engineered to learned features is the core innovation of deep learning and why it outperforms classical ML on complex tasks like vision and language.

Hierarchical Features

Deep networks learn hierarchical features: each layer builds on the previous one. In a vision model: layer 1 detects edges, layer 2 combines edges into textures and corners, layer 3 combines textures into object parts (eyes, wheels), layer 4 combines parts into objects (faces, cars). This hierarchy emerges automatically from training — no one programs it. The same hierarchical feature learning happens in language models, from character patterns to syntax to semantics to reasoning.

Feature Visualization

Researchers visualize features to understand what models learn. For vision models, you can generate images that maximally activate a specific neuron or direction, revealing what pattern it detects. For language models, you can find the text examples that most activate a specific feature direction. Anthropic's research has visualized features in Claude, finding interpretable concepts like "Golden Gate Bridge," "code bugs," "deception," and "French language" encoded as specific directions in the model's activation space.

相關概念

← 所有術語
← Existential Risk Federated 學習ing →