Manipulate User LLM Chat History (AML.T0092)

Tactic: Defense Evasion

Tactics
Defense Evasion
Maturity
demonstrated
Reference
atlas.mitre.org/techniques/AML.T0092

Description

Adversaries may manipulate a user’s large language model (LLM) chat history to cover the tracks of their malicious behavior. They may hide persistent changes they have made to the LLM’s behavior, or obscure their attempts at discovering private information about the user.

To do so, adversaries may delete or edit existing messages or create new threads as part of their coverup. This is feasible if the adversary has the victim’s authentication tokens for the backend LLM service or if they have direct access to the victim’s chat interface.

Chat interfaces (especially desktop interfaces) often do not show the injected prompt for any ongoing chat, as they update chat history only once when initially opening it. This can help the adversary’s manipulations go unnoticed by the victim.

How GTK Cyber trains on this

GTK Cyber's hands-on AI security courses cover adversarial-AI techniques across the MITRE ATLAS framework, including the Defense Evasion tactic this technique falls under. Our practitioner-led training is taught by Charles Givre and other field-tested SMEs and focuses on real adversarial scenarios, not slide decks.

View AI security courses →

Related techniques

Train your team on real adversarial-AI attacks.

GTK Cyber's AI red teaming courses are taught by practitioners who break models for a living.

View AI Security Courses