Zubnet AI学习Wiki › Feature
基础

Feature

学习ed Representation, Activation
神经网络学习在输入中检测的模式或概念。视觉中,早期层特征是边缘和纹理;后期层特征是物体部分和整体物体。语言模型中,特征范围从简单(字母“a”、特定句法模式)到抽象(讽刺的概念、特定推理策略)。特征被表示为跨神经元的激活模式。

为什么重要

特征是模型实际学到的东西 — 不是个别事实而是能泛化的模式。模型不记忆“猫有毛”;它学到一个对皮毛纹理的特征检测器,对猫、狗、泰迪熊都激活。理解特征帮助解释模型行为:为什么它泛化(特征迁移)、为什么失败(错误特征激活)、以及如何改进它(让它接触更多多样特征)。

Deep Dive

The term "feature" has different meanings depending on context. In classical ML, features are hand-engineered input variables (height, weight, age). In deep learning, features are learned representations in hidden layers — the model discovers useful patterns on its own. This shift from hand-engineered to learned features is the core innovation of deep learning and why it outperforms classical ML on complex tasks like vision and language.

Hierarchical Features

Deep networks learn hierarchical features: each layer builds on the previous one. In a vision model: layer 1 detects edges, layer 2 combines edges into textures and corners, layer 3 combines textures into object parts (eyes, wheels), layer 4 combines parts into objects (faces, cars). This hierarchy emerges automatically from training — no one programs it. The same hierarchical feature learning happens in language models, from character patterns to syntax to semantics to reasoning.

Feature Visualization

Researchers visualize features to understand what models learn. For vision models, you can generate images that maximally activate a specific neuron or direction, revealing what pattern it detects. For language models, you can find the text examples that most activate a specific feature direction. Anthropic's research has visualized features in Claude, finding interpretable concepts like "Golden Gate Bridge," "code bugs," "deception," and "French language" encoded as specific directions in the model's activation space.

相关概念

← 所有术语
← Existential Risk Federated 学习ing →