Zubnet AIसीखेंWiki › Knowledge Editing
Training

Knowledge Editing

Model Editing, Fact Editing
एक trained model में specific facts को बिना retrain किए modify करने की techniques। अगर एक model एक नए election के बाद incorrectly कहता है “The president of France is Macron”, knowledge editing targeted weights modify करके इस specific fact को update कर सकती है, model के दूसरे knowledge या capabilities को affect किए बिना। Goal है surgical precision: एक fact change करो, बाकी सब intact छोड़ो।

यह क्यों matter करता है

Knowledge editing एक practical problem address करती है: models outdated हो जाते हैं, और retraining expensive है। अगर आप specific facts cheaply update कर सकें, models major training runs के बीच current रह सकें। इसके safety implications भी हैं: क्या आप dangerous knowledge edit कर सकते हैं? Field promising है लेकिन immature — edits अक्सर related knowledge पर unintended side effects रखते हैं।

Deep Dive

The dominant approach (ROME/MEMIT): identify which feedforward network weights encode a specific fact by tracing the causal effect of neurons on the model's prediction, then modify those weights to change the stored association. For example, to update "The Eiffel Tower is in Paris" to "The Eiffel Tower is in London," you find the weights that map "Eiffel Tower" → "Paris" in the FFN layers and redirect them to "London."

The Ripple Effect Problem

Editing "The Eiffel Tower is in London" should also change answers to "What country is the Eiffel Tower in?" (UK, not France) and "What landmarks are in Paris?" (no longer the Eiffel Tower). Current editing methods often fail at this: they change the direct fact but leave related inferences inconsistent. This "ripple effect" problem suggests that knowledge in LLMs is more interconnected than the surgical editing metaphor implies.

Scaling Challenges

A few edits work reasonably well. Hundreds of edits start to degrade model quality — the edited weights accumulate changes that interfere with each other and with unedited knowledge. This limits knowledge editing's practical use: it's fine for a few corrections but can't serve as a general model update mechanism. For staying current, RAG (providing updated information at inference time) remains more practical than editing the model's weights.

संबंधित अवधारणाएँ

← सभी Terms
← Knowledge Cutoff Knowledge Graph →