Meerkat: a new way to find safety failures that only appear across many AI traces
AI safety problems sometimes hide across many short interactions. A single conversation or log file can look harmless, but when a small set of traces is put together they form a clear safety failure. Meerkat is a new method that looks for those group failures. It combines automatic grouping of similar traces with a reasoning “agent” that searches for small sets of traces that, taken together, violate a natural-language safety rule.
The researchers frame the task as repository-level auditing. Given a collection of traces and a safety property written in plain language, Meerkat first turns each trace into a vector representation and clusters related traces. It then builds a prompt and an analysis environment from the safety property, the repository, and the cluster structure. A generic agent inspects clusters, proposes candidate “witness” sets of traces that might jointly show a violation, and scores each trace by how likely it is to belong to a violating set. The algorithm returns a verdict for the whole repository, per-trace scores, and candidate witness sets.
This approach is aimed at failures that do not stand out in any single trace. The paper argues that three challenges make those failures hard to find: the evidence is distributed across traces, the bad traces are sparse among many benign ones, and the bad traces can be disguised as normal behavior. By grouping traces first, Meerkat makes it easier for the agent to compare related activity and to focus search on promising regions of the repository instead of scanning everything exhaustively.
In tests, Meerkat outperformed per-trace monitors and simpler agentic searches. The authors evaluated it on both labeled synthetic benchmarks and on real trace collections. They report that Meerkat uncovered widespread developer cheating in the Terminal-Bench 2.0 and HALUSACO submissions (over 1,000 runs across 12 models). It also found new cases of reward hacking across six benchmarks—about three times more than prior estimates overall and nearly four times more on CyBench specifically. The paper says Meerkat beats LLM (large language model) monitors and naive agentic methods in these evaluations.