Zubnet AI學習Wiki › Neuron
基礎

Neuron

Artificial Neuron, Perceptron, Node
神經網路最基本的計算單元。一個人工神經元接收輸入,把每個輸入乘以一個權重,相加,加上一個偏置,然後把結果通過一個激活函數產生輸出。數千到數十億這樣的神經元,按層組織並通過學到的權重連接,構成了驅動所有現代 AI 的神經網路。

為什麼重要

神經元是深度學習的原子。理解一個神經元 — 加權求和加激活 — 就讓神經網路架構的其他部分變得直觀。一層是一組神經元。一個網路是一堆層。訓練就是調整權重。其他一切都是細節(重要的細節,但終究是細節)。

Deep Dive

The artificial neuron is loosely inspired by biological neurons but shouldn't be taken as a literal analogy. A biological neuron receives electrical signals through dendrites, integrates them in the cell body, and fires (or doesn't) through the axon. An artificial neuron computes: output = activation(w1·x1 + w2·x2 + ... + wn·xn + bias). The weights (w) determine how much each input matters. The bias shifts the activation threshold. The activation function (ReLU, GELU) introduces non-linearity.

From Perceptron to Deep 學習ing

The perceptron (Rosenblatt, 1958) was the first artificial neuron — a single unit that could learn to classify linearly separable data. Minsky and Papert showed in 1969 that a single perceptron couldn't learn XOR (a simple non-linear function), contributing to the first AI winter. The solution: stack multiple layers of neurons (multi-layer perceptrons / MLPs), which can learn any function given enough neurons. This is the universal approximation theorem — the theoretical foundation of deep learning.

Neurons in Modern LLMs

A model like Llama-70B has roughly 70 billion parameters (weights and biases connecting neurons). Each feedforward layer has thousands of neurons. But modern research shows that individual neurons often don't correspond to single concepts — instead, concepts are encoded as directions in activation space across many neurons (superposition). A single neuron might participate in encoding dozens of different features, making interpretation challenging.

相關概念

← 所有術語
← Neural Network Normalization →