Latest AI Papers August 06 2025 Knowledge Model Editing GUI Agents Steering And Efficient LLMs

by StackCamp Team 95 views

Hey everyone! Check out the latest and greatest AI papers from August 6, 2025! This week, we're diving into some fascinating topics, including knowledge editing, model editing, GUI agents, steering vectors, and efficient LLMs. For a better reading experience and more papers, don't forget to visit the Github page. Let's get started!

Knowledge Editing

Knowledge editing is a crucial area in AI, especially for ensuring that our models are up-to-date, accurate, and reliable. This involves the ability to modify a model's existing knowledge without retraining it from scratch. Several interesting papers have surfaced in this domain, each tackling different aspects of knowledge manipulation in AI models. These advancements are super important as we strive to build AI systems that can learn and adapt dynamically. Imagine being able to correct a factual error in a large language model (LLM) without needing to retrain the entire thing – that's the power of knowledge editing! This field is not just about fixing mistakes; it's also about enhancing a model's ability to reason and make informed decisions. This week, we've seen a surge in research that pushes the boundaries of what's possible. For example, one standout paper, ADS-Edit: A Multimodal Knowledge Editing Dataset for Autonomous Driving Systems, addresses the crucial need for knowledge editing in self-driving cars. Imagine an autonomous vehicle needing to update its knowledge about road closures or new traffic patterns – knowledge editing makes this real-time adaptation possible. Then there's FPEdit: Robust LLM Fingerprinting through Localized Knowledge Editing, which focuses on safeguarding LLMs by identifying and rectifying vulnerabilities through precise edits. This is essential for maintaining the integrity and security of these models. KEDAS: Knowledge Editing Alignment with Diverse Augmentation and Self-adaptive Inference introduces a novel approach to aligning knowledge editing across diverse datasets, improving the overall adaptability of AI systems. This paper highlights the importance of ensuring that edits made to a model's knowledge base are consistent and effective across various contexts. Knowledge Editing for Multi-Hop Question Answering Using Semantic Analysis explores how semantic analysis can enhance knowledge editing in models designed for complex question answering. This is particularly relevant for AI systems that need to understand and reason through multiple pieces of information to provide accurate answers. SAKE: Steering Activations for Knowledge Editing presents a method for directly influencing a model's activations to achieve specific knowledge editing goals, offering a more controlled approach to modifying a model's behavior. This is akin to fine-tuning the model's internal mechanisms to align with desired outputs. DocTER: Evaluating Document-based Knowledge Editing focuses on assessing the effectiveness of knowledge editing techniques when applied to document-based knowledge. This is vital for applications such as information retrieval and document summarization, where models need to accurately process and update information from textual sources. Decoupling Knowledge and Reasoning in LLMs: An Exploration Using Cognitive Dual-System Theory delves into the theoretical underpinnings of knowledge editing, drawing parallels with human cognition to better understand how knowledge is processed and updated in LLMs. This paper provides a fascinating perspective on the cognitive architecture of AI models. NeuralDB: Scaling Knowledge Editing in LLMs to 100,000 Facts with Neural KV Database demonstrates the scalability of knowledge editing by handling massive amounts of factual information, a critical step towards building truly comprehensive AI systems. This work addresses the challenge of efficiently managing and updating large knowledge bases within LLMs. Why Does New Knowledge Create Messy Ripple Effects in LLMs? investigates the potential unintended consequences of knowledge editing, such as the spread of errors or inconsistencies, emphasizing the need for careful evaluation and mitigation strategies. This paper highlights the importance of understanding the ripple effects of edits to ensure the model's overall integrity. How Well Can Knowledge Edit Methods Edit Perplexing Knowledge? challenges the current state of knowledge editing by testing its limits with complex and ambiguous information, pushing researchers to develop more robust and nuanced editing techniques. This work encourages the development of methods that can handle the complexities of real-world knowledge. Understanding Language Model Circuits through Knowledge Editing provides a deep dive into the inner workings of language models by analyzing how knowledge editing affects their internal circuits, offering valuable insights into model behavior. This paper contributes to the broader goal of making AI models more transparent and interpretable. An Exploration of Knowledge Editing for Arabic extends the application of knowledge editing to different languages, highlighting the importance of adapting these techniques to diverse linguistic contexts. This work recognizes the need for knowledge editing methods that are effective across various languages and cultural settings. ChainEdit: Propagating Ripple Effects in LLM Knowledge Editing through Logical Rule-Guided Chains presents a method for managing the ripple effects of knowledge editing by guiding the updates through logical rules, ensuring consistency and accuracy. This approach aims to create a more systematic and reliable knowledge editing process. Towards a Principled Evaluation of Knowledge Editors proposes a framework for evaluating the effectiveness of knowledge editing methods, addressing the need for standardized metrics and benchmarks in this field. This paper helps to establish a foundation for comparing and improving knowledge editing techniques. Lastly, Uncovering Overfitting in Large Language Model Editing sheds light on the potential for knowledge editing to lead to overfitting, emphasizing the need for regularization techniques and careful validation. This work underscores the importance of balancing the benefits of knowledge editing with the risk of compromising the model's generalization ability.

Title Date Comment
ADS-Edit: A Multimodal Knowledge Editing Dataset for Autonomous Driving Systems 2025-08-05 ACM MM 2025
FPEdit: Robust LLM Fingerprinting through Localized Knowledge Editing 2025-08-04
KEDAS: Knowledge Editing Alignment with Diverse Augmentation and Self-adaptive Inference 2025-08-02 Preprint
Knowledge Editing for Multi-Hop Question Answering Using Semantic Analysis 2025-07-29
14 pa...

14 pages, 15 figures, pre-print of paper accepted to IJCAI 2025

SAKE: Steering Activations for Knowledge Editing 2025-07-29
DocTER: Evaluating Document-based Knowledge Editing 2025-07-24
Infor...

Information processing & management

Decoupling Knowledge and Reasoning in LLMs: An Exploration Using Cognitive Dual-System Theory 2025-07-24
NeuralDB: Scaling Knowledge Editing in LLMs to 100,000 Facts with Neural KV Database 2025-07-24
Why Does New Knowledge Create Messy Ripple Effects in LLMs? 2025-07-20
How Well Can Knowledge Edit Methods Edit Perplexing Knowledge? 2025-07-15
A pre...

A previous version of this document contained a hidden prompt entered by Z Zhu without knowledge of -- or consent by -- his co-authors. This version does not contain the prompt

Understanding Language Model Circuits through Knowledge Editing 2025-07-15
A pre...

A previous version of this document contained a hidden prompt entered by Z Zhu without knowledge of -- or consent by -- his co-authors. This version does not contain the prompt

An Exploration of Knowledge Editing for Arabic 2025-07-13
ChainEdit: Propagating Ripple Effects in LLM Knowledge Editing through Logical Rule-Guided Chains 2025-07-11
Accep...

Accepted to ACL 2025 (main)

Towards a Principled Evaluation of Knowledge Editors 2025-07-08
Accep...

Accepted at L2M2 workshop at ACL 2025

Uncovering Overfitting in Large Language Model Editing 2025-06-17 ICLR 2025

Model Editing

Model editing is another exciting area, focusing on how we can modify the behavior of AI models after they've been trained. This is particularly useful for addressing issues like bias, safety, and the incorporation of new information. The ability to edit models without extensive retraining is a game-changer, offering a more efficient and targeted approach to model improvement. Think of it as AI surgery – making precise changes to improve performance without disrupting the entire system. This week's papers showcase the breadth and depth of research in this domain. The advancements in model editing are not just about correcting errors; they're also about aligning AI behavior with ethical standards and societal values. The paper Latent Adversarial Training Improves Robustness to Persistent Harmful Behaviors in LLMs introduces a technique to make LLMs more resilient against generating harmful content by training them with adversarial examples in the latent space. This approach is vital for ensuring that AI systems are safe and responsible. Then we have A Survey on Generative Model Unlearning: Fundamentals, Taxonomy, Evaluation, and Future Direction, which offers a comprehensive overview of the field of generative model unlearning. Unlearning is crucial for removing sensitive information from models while maintaining their performance, making this survey a valuable resource for researchers. Improving Code LLM Robustness to Prompt Perturbations via Layer-Aware Model Editing explores methods to enhance the stability of code-generating LLMs against slight variations in input prompts, a practical concern in real-world applications. This is especially important for software development tools where consistent and reliable outputs are essential. Cross-Encoder Rediscovers a Semantic Variant of BM25 delves into how cross-encoders can semantically reframe the traditional BM25 ranking algorithm, showcasing how model editing can bridge the gap between classical methods and modern neural approaches. This research highlights the potential for combining the strengths of different AI paradigms. The paper How Well Can Knowledge Edit Methods Edit Perplexing Knowledge? (also listed in the knowledge editing section) is relevant here as well, emphasizing the challenges of editing models with complex information. Retention analysis of edited knowledge after fine-tuning investigates how well edits made to a model's knowledge are retained after subsequent fine-tuning, providing insights into the durability of model editing techniques. This is critical for ensuring that edits are not lost during further training. Towards a Principled Evaluation of Knowledge Editors (also in knowledge editing) is important for model editing as well, as it advocates for standardized evaluation metrics to assess the effectiveness of different editing approaches. SeaLion: Semantic Part-Aware Latent Point Diffusion Models for 3D Generation introduces a novel approach to 3D generation that leverages semantic part awareness, showcasing the versatility of model editing in creative applications. This work opens up possibilities for more detailed and controllable 3D content creation. Attributing Data for Sharpness-Aware Minimization explores how data attribution can be used in conjunction with sharpness-aware minimization to improve model generalization, highlighting a connection between model editing and training techniques. Eigenvoice Synthesis based on Model Editing for Speaker Generation presents a method for generating diverse speaker voices through model editing, demonstrating its utility in speech synthesis applications. This research could lead to more natural and personalized voice interfaces. Model Editing as a Double-Edged Sword: Steering Agent Ethical Behavior Toward Beneficence or Harm raises critical ethical considerations about model editing, highlighting its potential for both good and bad and emphasizing the need for responsible development. This paper is a crucial reminder of the ethical dimensions of AI research. How Robust is Model Editing after Fine-Tuning? An Empirical Study on Text-to-Image Diffusion Models assesses the robustness of model editing in text-to-image diffusion models after fine-tuning, providing empirical evidence of the stability of these edits. This research contributes to the reliability of model editing in generative models. QueueEDIT: Structural Self-Correction for Sequential Model Editing in LLMs introduces a method for self-correcting edits in LLMs, improving the accuracy and consistency of model modifications. This approach is particularly relevant for maintaining the integrity of complex models. The OCR Quest for Generalization: Learning to recognize low-resource alphabets with model editing explores how model editing can enable OCR systems to recognize alphabets with limited training data, expanding the applicability of OCR technology. This work addresses the challenge of adapting AI systems to diverse linguistic contexts. Finally, Resolving UnderEdit & OverEdit with Iterative & Neighbor-Assisted Model Editing presents a strategy for addressing common issues in model editing, such as insufficient edits or excessive changes, through iterative refinement and neighbor-assisted techniques. This paper offers practical solutions for improving the precision of model editing.

Title Date Comment
Latent Adversarial Training Improves Robustness to Persistent Harmful Behaviors in LLMs 2025-07-29
Code ...

Code at https://github.com/aengusl/latent-adversarial-training. Models at https://huggingface.co/LLM-LAT

A Survey on Generative Model Unlearning: Fundamentals, Taxonomy, Evaluation, and Future Direction 2025-07-26
Improving Code LLM Robustness to Prompt Perturbations via Layer-Aware Model Editing 2025-07-22
Cross-Encoder Rediscovers a Semantic Variant of BM25 2025-07-22
How Well Can Knowledge Edit Methods Edit Perplexing Knowledge? 2025-07-15
A pre...

A previous version of this document contained a hidden prompt entered by Z Zhu without knowledge of -- or consent by -- his co-authors. This version does not contain the prompt

Retention analysis of edited knowledge after fine-tuning 2025-07-14
Towards a Principled Evaluation of Knowledge Editors 2025-07-08
Accep...

Accepted at L2M2 workshop at ACL 2025

SeaLion: Semantic Part-Aware Latent Point Diffusion Models for 3D Generation 2025-07-07
Accep...

Accepted by CVPR 2025

Attributing Data for Sharpness-Aware Minimization 2025-07-05 25 pages
Eigenvoice Synthesis based on Model Editing for Speaker Generation 2025-07-04
Accep...

Accepted by INTERSPEECH 2025

Model Editing as a Double-Edged Sword: Steering Agent Ethical Behavior Toward Beneficence or Harm 2025-06-25
Main ...

Main paper: 9 pages; total: 18 pages (including appendix). Code, data, results, and additional resources are available at: https://model-editing.github.io

How Robust is Model Editing after Fine-Tuning? An Empirical Study on Text-to-Image Diffusion Models 2025-06-23
QueueEDIT: Structural Self-Correction for Sequential Model Editing in LLMs 2025-06-22
The OCR Quest for Generalization: Learning to recognize low-resource alphabets with model editing 2025-06-18
Prepr...

Preprint (under review) For Journal

Resolving UnderEdit & OverEdit with Iterative & Neighbor-Assisted Model Editing 2025-06-17 Under Review

GUI Agent

GUI Agents, or Graphical User Interface Agents, are AI systems designed to interact with software applications through their graphical interfaces, mimicking human interaction. This area is rapidly evolving, with agents becoming more sophisticated in their ability to navigate and operate complex software environments. Imagine having an AI assistant that can not only understand your requests but also execute them by clicking buttons and filling forms within an application – that's the promise of GUI Agents. This week's papers highlight significant advancements in this field, from improving agent robustness to creating new benchmarks and datasets. The development of GUI Agents is crucial for automating tasks, enhancing accessibility, and creating more intuitive user experiences. The ability of these agents to learn and adapt to different interfaces is a key factor in their widespread adoption. GEM: Gaussian Embedding Modeling for Out-of-Distribution Detection in GUI Agents introduces a method for detecting unfamiliar GUI elements, enhancing the robustness of GUI Agents in real-world scenarios. This capability is vital for ensuring that agents can handle unexpected situations gracefully. NatureGAIA: Pushing the Frontiers of GUI Agents with a Challenging Benchmark and High-Quality Trajectory Dataset presents a new benchmark and dataset designed to challenge and improve the capabilities of GUI Agents. Such resources are essential for driving progress in this field. GUIOdyssey: A Comprehensive Dataset for Cross-App GUI Navigation on Mobile Devices offers a dataset specifically tailored for training agents to navigate across multiple mobile applications, addressing a practical need in mobile automation. This dataset supports the development of agents that can seamlessly integrate different apps to achieve user goals. UI-E2I-Synth: Advancing GUI Grounding with Large-Scale Instruction Synthesis explores the use of synthesized instructions for training GUI Agents, a technique that can significantly reduce the need for manually labeled data. This approach is particularly useful for scaling up the training of agents. UI-AGILE: Advancing GUI Agents with Effective Reinforcement Learning and Precise Inference-Time Grounding focuses on improving the learning and grounding abilities of GUI Agents through reinforcement learning, leading to more accurate and efficient interactions. This research contributes to the development of agents that can adapt and learn in dynamic environments. MMBench-GUI: Hierarchical Multi-Platform Evaluation Framework for GUI Agents proposes a framework for evaluating GUI Agents across different platforms, ensuring that agents are versatile and perform well in diverse settings. This framework helps to standardize the evaluation of GUI Agents, making it easier to compare different approaches. MobileUse: A GUI Agent with Hierarchical Reflection for Autonomous Mobile Operation introduces a GUI Agent designed for autonomous mobile operation, leveraging hierarchical reflection to improve decision-making and task execution. This agent demonstrates the potential for AI to automate tasks on mobile devices. macOSWorld: A Multilingual Interactive Benchmark for GUI Agents provides a benchmark for evaluating GUI Agents in a multilingual context, addressing the need for agents that can operate in diverse linguistic environments. This benchmark supports the development of agents that can cater to a global user base. Visual Test-time Scaling for GUI Agent Grounding explores techniques for improving the visual grounding abilities of GUI Agents at test time, enhancing their ability to accurately identify and interact with GUI elements. OS-Kairos: Adaptive Interaction for MLLM-Powered GUI Agents presents a GUI Agent powered by Multimodal Large Language Models (MLLMs), capable of adaptive interaction based on user context. This agent represents a significant step towards more intelligent and responsive interfaces. LaSM: Layer-wise Scaling Mechanism for Defending Pop-up Attack on GUI Agents addresses the security aspect of GUI Agents, introducing a mechanism to defend against pop-up attacks, ensuring the robustness of agents in adversarial environments. This research is vital for maintaining the security of AI systems interacting with sensitive applications. GTA1: GUI Test-time Scaling Agent focuses on scaling the performance of GUI Agents at test time, improving their efficiency and effectiveness in real-world applications. VisualTrap: A Stealthy Backdoor Attack on GUI Agents via Visual Grounding Manipulation highlights potential vulnerabilities in GUI Agents, demonstrating a stealthy backdoor attack that manipulates visual grounding. This research underscores the importance of security considerations in the design of GUI Agents. MobileGUI-RL: Advancing Mobile GUI Agent through Reinforcement Learning in Online Environment explores the use of reinforcement learning to train GUI Agents in online mobile environments, allowing agents to learn and adapt continuously. Finally, R-VLM: Region-Aware Vision Language Model for Precise GUI Grounding introduces a region-aware vision language model for precise GUI grounding, improving the accuracy with which agents can identify and interact with specific GUI elements. This model contributes to the overall precision and reliability of GUI Agents.

Title Date Comment
GEM: Gaussian Embedding Modeling for Out-of-Distribution Detection in GUI Agents 2025-08-04
NatureGAIA: Pushing the Frontiers of GUI Agents with a Challenging Benchmark and High-Quality Trajectory Dataset 2025-08-02
GUIOdyssey: A Comprehensive Dataset for Cross-App GUI Navigation on Mobile Devices 2025-08-01
22 pa...

22 pages, 14 figures, ICCV 2025, a cross-app GUI navigation dataset

UI-E2I-Synth: Advancing GUI Grounding with Large-Scale Instruction Synthesis 2025-07-30
UI-AGILE: Advancing GUI Agents with Effective Reinforcement Learning and Precise Inference-Time Grounding 2025-07-30
MMBench-GUI: Hierarchical Multi-Platform Evaluation Framework for GUI Agents 2025-07-25 in progress
MobileUse: A GUI Agent with Hierarchical Reflection for Autonomous Mobile Operation 2025-07-21
A tec...

A technical report on a GUI agent based on multi-agent systems

macOSWorld: A Multilingual Interactive Benchmark for GUI Agents 2025-07-16
Visual Test-time Scaling for GUI Agent Grounding 2025-07-14
ICCV2...

ICCV2025, https://github.com/tiangeluo/RegionFocus

OS-Kairos: Adaptive Interaction for MLLM-Powered GUI Agents 2025-07-14
25 pa...

25 pages, 24 figures, 11 tables (ACL 2025, Findings)

LaSM: Layer-wise Scaling Mechanism for Defending Pop-up Attack on GUI Agents 2025-07-13 10 pages, 9 figures
GTA1: GUI Test-time Scaling Agent 2025-07-10
VisualTrap: A Stealthy Backdoor Attack on GUI Agents via Visual Grounding Manipulation 2025-07-09
MobileGUI-RL: Advancing Mobile GUI Agent through Reinforcement Learning in Online Environment 2025-07-08 17 pages, 4 figures
R-VLM: Region-Aware Vision Language Model for Precise GUI Grounding 2025-07-08 ACL 2025; 17 pages

Steering Vector

Steering vectors are a fascinating concept in AI, allowing us to control and guide the behavior of large language models by manipulating their internal representations. Think of it as having a steering wheel for your AI – you can influence its outputs and responses without retraining the entire model. This is a powerful tool for various applications, from generating specific types of text to mitigating biases. This week's papers delve into various aspects of steering vectors, including their application in emotion control, topic modeling, and reasoning. Steering vectors are not just about directing the model's output; they also provide insights into how these models work internally. This research is crucial for making AI more controllable, interpretable, and aligned with human values. The paper EmoSteer-TTS: Fine-Grained and Training-Free Emotion-Controllable Text-to-Speech via Activation Steering introduces a method for generating speech with specific emotions by manipulating the model's activations through steering vectors, a training-free approach to emotion control in text-to-speech systems. This is a significant step towards more expressive and natural-sounding AI voices. Model Directions, Not Words: Mechanistic Topic Models Using Sparse Autoencoders explores the use of sparse autoencoders to identify meaningful directions in a model's embedding space, which can then be used as steering vectors for topic modeling. This work offers a novel approach to understanding and manipulating the topical content of text generated by AI models. GrAInS: Gradient-based Attribution for Inference-Time Steering of LLMs and VLMs presents a gradient-based method for identifying steering vectors that can influence the behavior of both LLMs and vision-language models (VLMs) at inference time. This approach enables fine-grained control over model outputs. Understanding Reasoning in Thinking Language Models via Steering Vectors investigates how steering vectors can be used to understand and influence the reasoning processes of language models, providing insights into the inner workings of these systems. This research contributes to making AI reasoning more transparent and controllable. Reasoning-Finetuning Repurposes Latent Representations in Base Models explores how fine-tuning for reasoning tasks can repurpose the latent representations in base models, which can then be leveraged through steering vectors. This work highlights the potential for transferring reasoning abilities across different models. Simple Mechanistic Explanations for Out-Of-Context Reasoning proposes simple mechanisms for explaining how models perform out-of-context reasoning, potentially paving the way for more interpretable steering vectors. Shared Global and Local Geometry of Language Model Embeddings examines the geometric properties of language model embeddings, providing a foundation for designing more effective steering vectors that leverage the underlying structure of the embedding space. Beyond Multiple Choice: Evaluating Steering Vectors for Adaptive Free-Form Summarization presents a method for evaluating steering vectors in the context of adaptive summarization, showcasing their ability to generate summaries tailored to specific needs. KV Cache Steering for Inducing Reasoning in Small Language Models introduces a technique for inducing reasoning in small language models by manipulating their key-value (KV) cache using steering vectors. This approach demonstrates the potential for enhancing the reasoning capabilities of smaller models. Multi-Attribute Steering of Language Models via Targeted Intervention presents a method for steering language models along multiple attributes simultaneously using targeted interventions, allowing for more nuanced control over model outputs. This research enables the generation of text that satisfies multiple constraints or preferences. Activation Steering for Chain-of-Thought Compression explores the use of activation steering to compress the chain-of-thought reasoning process in language models, improving efficiency without sacrificing performance. EasyEdit2: An Easy-to-use Steering Framework for Editing Large Language Models introduces a user-friendly framework for steering and editing large language models, making these techniques more accessible to a wider audience. FairSteer: Inference Time Debiasing for LLMs with Dynamic Activation Steering presents a method for mitigating biases in LLMs at inference time using dynamic activation steering, ensuring fairer and more equitable AI systems. This research contributes to the development of AI that is aligned with ethical principles. Controlling Thinking Speed in Reasoning Models explores how steering vectors can be used to control the