File Download

There are no files associated with this item.

  • Find it @ UNIST can give you direct access to the published full text of this article. (UNISTARs only)
Related Researcher

나승훈

Na, Seung-Hoon
Natural Language Processing Lab
Read More

Views & Downloads

Detailed Information

Cited time in webofscience Cited time in scopus
Metadata Downloads

MECA: Modular editing via customized expert networks and adaptors in large language models

Author(s)
Nyange, RoselineQiao, ShanbaoNa, Seung-Hoon
Issued Date
2026-03
DOI
10.1016/j.eswa.2025.129675
URI
https://scholarworks.unist.ac.kr/handle/201301/88548
Citation
EXPERT SYSTEMS WITH APPLICATIONS, v.298, no.C, pp.129675
Abstract
Updating language models with new information through targeted edits without resorting to expensive full model retraining remains a critical challenge, particularly when aiming to preserve pre-existing capabilities. In this work, we introduce Modular Editing via Customized expert networks and Adaptors) (MECA), a unified framework that selectively integrates new knowledge into language models. MECA employs a module-level deferral router to evaluate whether incoming queries fall within the scope of existing edit requests. Queries are then dynamically routed to either customized editing experts or key-value adaptors. This modular strategy ensures that updates are localized, thereby mitigating risks of unintended alterations on unrelated outputs. We validate our approach on sequential editing tasks using Llama2-7B, Llama2-13B and Falcon 11B, benchmarked across two diverse datasets ZsRE and Hallucination. Experimental results show that MECA consistently outperforms several stateof-the-art knowledge editing techniques, achieving improved integration of new information while preserving the model's original performance. Our analysis further demonstrates that the deferral routing mechanism for selecting modules effectively balances editing precision with overall model stability.
Publisher
PERGAMON-ELSEVIER SCIENCE LTD
ISSN
0957-4174
Keyword (Author)
Large language models (LLM)Mixture of experts (MoE)Knowledge editingContinual learning

qrcode

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.