Zubnet AI學習Wiki › Machine Translation
Using AI

Machine Translation

MT, Neural Machine Translation, NMT
自動把文字從一種語言翻譯成另一種。現代神經機器翻譯(NMT)用在平行語料(文字及其翻譯)上訓練的 encoder-decoder Transformer。Google Translate、DeepL 和基於 LLM 的翻譯都用這種方法的變體。品質已經顯著提高 — 對常見語言對,MT 在常規內容上接近專業人工翻譯。

為什麼重要

機器翻譯規模化地打破語言障礙。它實現全球商業、跨語言搜尋、即時溝通、以及跨語言的資訊獲取。對 AI 來說,MT 是主要用英語訓練的模型如何服務 100+ 種語言使用者的方法 — 也是多語言 tokenizer 效率對成本重要的原因。

Deep Dive

Modern NMT uses the encoder-decoder Transformer architecture: the encoder processes the source sentence, and the decoder generates the target sentence token by token, attending to the encoded source through cross-attention. Training requires parallel corpora — millions of sentence pairs in both languages. Data quality and domain match are critical: a model trained on EU Parliament proceedings translates legal text well but informal chat poorly.

LLMs as Translators

Large language models have become competitive translators, sometimes exceeding dedicated MT systems for high-resource language pairs. Their advantage: they understand context, idioms, and cultural nuances better because they've seen language used in diverse contexts. Their disadvantage: they're much slower and more expensive per sentence than dedicated MT models. For real-time translation of millions of sentences, dedicated models (like those behind Google Translate) are necessary. For quality-critical translation of smaller volumes, LLMs often produce more natural results.

The Long Tail of Languages

MT quality varies enormously across language pairs. English-French, English-Spanish, and English-Chinese are well-served (abundant training data). But for the world's 7,000+ languages, most pairs have little or no parallel training data. Low-resource translation remains an active research area, with approaches including: zero-shot translation through multilingual models, back-translation (using the MT system itself to generate synthetic training data), and transfer learning from related languages.

相關概念

← 所有術語
ESC