Zubnet AI学习Wiki › Neuron
基础

Neuron

Artificial Neuron, Perceptron, Node
神经网络最基本的计算单元。一个人工神经元接收输入,把每个输入乘以一个权重,相加,加上一个偏置,然后把结果通过一个激活函数产生输出。数千到数十亿这样的神经元,按层组织并通过学到的权重连接,构成了驱动所有现代 AI 的神经网络。

为什么重要

神经元是深度学习的原子。理解一个神经元 — 加权求和加激活 — 就让神经网络架构的其他部分变得直观。一层是一组神经元。一个网络是一堆层。训练就是调整权重。其他一切都是细节(重要的细节,但终究是细节)。

Deep Dive

The artificial neuron is loosely inspired by biological neurons but shouldn't be taken as a literal analogy. A biological neuron receives electrical signals through dendrites, integrates them in the cell body, and fires (or doesn't) through the axon. An artificial neuron computes: output = activation(w1·x1 + w2·x2 + ... + wn·xn + bias). The weights (w) determine how much each input matters. The bias shifts the activation threshold. The activation function (ReLU, GELU) introduces non-linearity.

From Perceptron to Deep 学习ing

The perceptron (Rosenblatt, 1958) was the first artificial neuron — a single unit that could learn to classify linearly separable data. Minsky and Papert showed in 1969 that a single perceptron couldn't learn XOR (a simple non-linear function), contributing to the first AI winter. The solution: stack multiple layers of neurons (multi-layer perceptrons / MLPs), which can learn any function given enough neurons. This is the universal approximation theorem — the theoretical foundation of deep learning.

Neurons in Modern LLMs

A model like Llama-70B has roughly 70 billion parameters (weights and biases connecting neurons). Each feedforward layer has thousands of neurons. But modern research shows that individual neurons often don't correspond to single concepts — instead, concepts are encoded as directions in activation space across many neurons (superposition). A single neuron might participate in encoding dozens of different features, making interpretation challenging.

相关概念

← 所有术语
← Neural Network Normalization →