2 min readfrom Machine Learning

How do you test AI agents in production? The unpredictability is overwhelming.[D]

Our take

Testing AI agents in production presents unique challenges, especially when dealing with non-deterministic outputs. Traditional quality assurance methods, which rely on predictable input-output relationships, often fall short. As AI models handle complex, multi-step tasks, the unpredictability in reasoning and tool selection complicates the validation process. Snapshot testing can be too fragile, and human evaluations lack scalability. To ensure rigorous quality, a framework is needed that verifies the reasoning steps of these agents without relying on hardcoded outputs or introducing additional failure modes.

I’ve been in QA for almost a decade. My mental model for quality was always: given input X, assert output Y. Now I’m on a team that’s shipping an LLM-based agent that handles multi-step tasks. I genuinely do not know how to test this in a way that feels rigorous.

The thing works. But the output isn’t deterministic. The same input can produce different reasoning chains across runs. Hell even with temp=0 I see variation in tool selection and intermediate steps. My normal instincts don’t map here. I can’t write an assertion and run it a thousand times to track flakiness. I’m at a loss for what to do.

Snapshot testing on final outputs is too brittle. If there’s a correct response that’s worded differently it breaks the test. Regex/keyword matching on outputs misses reasoning errors that accidentally land on the correct answer. Human eval isn’t automatable and doesn’t scale. Evals with a scoring rubric almost works but I don’t have a way to set pass/fail thresholds.

I want something conceptually equivalent to integration tests for reasoning steps. Like, given this tool result does the next step correctly incorporate it? I don’t know how to make that assertion without either hardcoding expected outputs or using another LLM as a judge, which would introduce a new failure mode into my test suite.

The agent runs inside our product. There are real uses and actual consequences when it makes a bad call.

Is there a framework that allows for verifying of agentic reasoning?

submitted by /u/this_aint_taliya
[link] [comments]

Read on the original site

Open the publisher's page for the full experience

View original article

Tagged with

#natural language processing for spreadsheets#generative AI for data analysis#Excel alternatives for data analysis#financial modeling with spreadsheets#rows.com#cloud-based spreadsheet applications#real-time data collaboration#real-time collaboration#spreadsheet API integration#AI agents#LLM-based agent#QA#production#agentic reasoning#multi-step tasks#reasoning chains#verification framework#deterministic output#tool selection#reasoning errors