Researchers propose safeguards to curb AI trading errors

Researchers propose safeguards to limit financial losses when AI trading agents make errors.

A team of researchers from Microsoft, Google DeepMind, Columbia University, and startups Virtuals Protocol and t54.ai in a new paper proposed technical and operational safeguards to limit financial losses when autonomous AI trading agents make errors. The recommendations target firms, exchanges and brokerages that deploy machine-learning trading systems.

The researchers say the growing use of machine-learning systems, including reinforcement-learning agents that learn strategies from market data, raises the risk that software bugs, mis-specified objectives or unexpected market conditions could produce rapid, large losses.

At the agent level the paper recommends hard loss caps and position limits so an algorithm cannot keep increasing exposure after early losses. The authors propose adding risk-penalty terms to objective functions so agents favor stable trades over high-variance strategies and using safe-exploration techniques to restrict novel actions while the agent is learning. They also call for formal verification and stress testing of trading policies in simulated markets that include adversarial scenarios.

On the operational side the researchers recommend real-time monitoring that flags anomalous behavior and automatically throttles or pauses an agent when abnormal patterns appear. They advise maintaining tamper-evident logs of decision inputs and outputs to allow operators and auditors to reconstruct events. Exchange-level defenses mentioned include tighter circuit breakers and automated volume or price limits that trigger earlier when an automated trader is suspected of malfunctioning.

The paper emphasizes human oversight and clear escalation paths. The authors recommend human-in-the-loop controls enabling traders, risk managers and exchange operators to intervene quickly, and a ‘kill switch’ or vetted emergency shutdown procedure that can isolate a misbehaving agent without disrupting unrelated market participants. The researchers describe layered defenses in which an agent with built-in limits runs in a sandbox, is promoted to live trading with continuous monitoring, and has its trading speed stepped down or exchange limits invoked if thresholds are breached.

For governance the team proposes standardized audit requirements for firms deploying autonomous agents, routine red-teaming exercises to probe models for unexpected responses, a registry of model change events, and reporting requirements for incidents so lessons can be shared across the industry.

The researchers cite past incidents such as the 2010 U.S. “Flash Crash” and a 2012 software failure at a major market maker as examples of how algorithmic failures can cascade. They recommend immediate pilot programs that implement the proposed safeguards at willing firms and exchanges to evaluate performance in live conditions.

The paper states, “Combining agent-level constraints, continuous monitoring, and rapid human intervention creates multiple opportunities to stop a fault from becoming a systemic loss,” and notes the safeguards are easier to apply when they are simple to implement and test.

The material on GNcrypto is intended solely for informational use and must not be regarded as financial advice. We make every effort to keep the content accurate and current, but we cannot warrant its precision, completeness, or reliability. GNcrypto does not take responsibility for any mistakes, omissions, or financial losses resulting from reliance on this information. Any actions you take based on this content are done at your own risk. Always conduct independent research and seek guidance from a qualified specialist. For further details, please review our Terms, Privacy Policy and Disclaimers.

Articles by this author