Circuit-aware Knowledge Editing Improves Reasoning in Large Language Models

Top post
Ending Outdated Knowledge: Circuit-aware Editing for Learnable AI Models
Large language models (LLMs) have revolutionized the way we interact with information. However, updating these models with new knowledge poses a challenge. Changes to isolated facts do not necessarily lead to the correct application of this knowledge in more complex reasoning. A promising approach to solving this problem is "Knowledge Editing" (KE). This article highlights the challenges of KE and introduces an innovative method called CaKE (Circuit-aware Knowledge Editing).
The Challenge of Knowledge Updates in LLMs
Existing KE methods encounter difficulties in generalizing updates. While they can efficiently change individual facts, they show weaknesses when it comes to applying these changes to multi-step reasoning processes. The reason lies in the way LLMs function: they use so-called "Reasoning Circuits" – neural pathways responsible for knowledge-based reasoning. Previous KE approaches, which often only edit individual or a few layers of the model, do not intervene deeply enough in these reasoning circuits. As a result, the models fail to integrate updated knowledge into their complex thought processes.
CaKE: A New Approach for Effective Knowledge Editing
CaKE offers an innovative approach to effectively integrate updated knowledge into LLMs. In contrast to previous methods that focus on individual layers, CaKE considers the complex reasoning circuits of the models. By using strategically selected data compiled based on an analysis of these circuits, the model is encouraged to actively use the modified knowledge. This targeted training process stimulates the development of appropriate reasoning circuits for the newly integrated knowledge. This ensures that the updates are not only stored in isolation but are also correctly applied in complex thought processes.
Promising Results and Future Potential
Initial test results demonstrate the potential of CaKE. Compared to existing KE methods, CaKE was able to increase the accuracy of multi-step reasoning by an average of 20%. These results were determined using the MQuAKE dataset, which was specifically developed for the evaluation of KE methods. The improved integration of updated knowledge through CaKE opens up new possibilities for the use of LLMs in areas that require high accuracy and reliability. The research results on CaKE are publicly available, including the code and the data used, to promote further development and application of this promising technology.
The Significance of CaKE for the Future of AI
The development of CaKE is an important step towards more learnable and adaptable AI systems. The ability to update knowledge efficiently and accurately is crucial for the use of LLMs in dynamic environments where information is constantly changing. CaKE helps to close the gap between static knowledge and the constantly evolving reality and paves the way for AI systems that can continuously adapt to new information and integrate it into their complex thought processes. This opens up new perspectives for the application of AI in areas such as research, education, and decision-making.
Bibliographie: - https://www.ml.informatik.tu-darmstadt.de/ - https://iclr.cc/virtual/2025/papers.html - https://www.researchgate.net/publication/379718517_ProxEdit_Improving_Tuning-Free_Real_Image_Editing_with_Proximal_Guidance - https://nips.cc/virtual/2024/papers.html - https://johnguerra.co/viz/neurips2024Papers/ - https://www2025.thewebconf.org/poster-session - https://aaai.org/wp-content/uploads/2025/01/AAAI-25-Poster-Schedule.pdf - https://de.linkedin.com/in/ioan-albert-florea-b468a51a2 - https://iclr.cc/virtual/2025/calendar - https://aaai.org/wp-content/uploads/2025/01/AAAI-25-Oral-Talks-Schedule.pdf - Yao, Yunzhi, et al. "CaKE: Circuit-aware Editing Enables Generalizable Knowledge Learners." arXiv preprint arXiv:2503.16356 (2025).