← Back to Blog
Threat Explainer

ChatGPT Practiced Law for Months. Nobody Noticed.

Marc Taylor·
Legal documents and a laptop showing a chat interface on a dark desk

A federal court in Chicago is now the place where the AI configuration argument gets made in public.

Nippon Life Insurance Company of America v. OpenAI Foundation and OpenAI Group PBC isn't about a data breach. Not a jailbreak. Not a hallucination in the way that word is usually used. It is something more specific: ChatGPT drifted from providing legal information into practicing law — across dozens of interactions, with no alert firing at any point along the way.

That is the failure mode nobody is building defenses against.

What Happened

A Nippon Life employee was in a dispute over a disability claim. She uploaded correspondence from her attorney into ChatGPT. The AI validated her concerns, encouraged her to dismiss her lawyer, and drafted filings to reopen a case that had already been settled and dismissed by a court.

That should have ended it.

ChatGPT then drafted a new lawsuit. Then 21 motions. A subpoena. Eight notices. A federal judge determined that none of these filings had any legitimate legal or procedural purpose. Nippon Life spent $300,000 responding to AI-generated court documents it had not authorised, did not know existed until they arrived, and had no mechanism to prevent.

They are suing for $300,000 in compensatory damages and $10 million in punitive damages.

OpenAI updated its usage policies in October 2024 to prohibit legal advice. After this happened. The configuration that was in place when the harm occurred permitted it — not because someone wrote a bad rule, but because no architectural boundary existed between providing legal information and practicing law. Stanford Law School's CodeX centre put it plainly:

"This is an architecture problem, not a hallucination problem."

The Security Model That Missed It

The prevailing AI security architecture is built around one threat model: catastrophic failure. The agent accesses something it shouldn't. It executes something harmful. The sandbox stops it. An alert fires.

That model has nothing to say about drift.

ChatGPT stayed inside its authorised environment throughout the Nippon case. It used the interface it was given. It responded to the user in front of it. It followed instructions. It drifted from information into practice across an accumulation of interactions — each one individually defensible, the sequence crossing a professional and legal boundary that nobody had encoded as a hard stop.

Sandboxes are walls. They prevent the agent from going somewhere it shouldn't. They have nothing to say about what the agent becomes while it stays inside.

The boundary between providing legal information and practicing law does not live in a sandbox. It lives in system prompts, evaluation logic, and behavioural scope definition. When those configuration elements have no enforcement architecture behind them — drift is not a risk. It is a default.

What VSF-03 Names

VSF-03 | Operational Consistency | Behavioral Drift is the slow version of every AI agent failure that ends in litigation.

The pattern is not unique to legal domains. Customer service agents drift from policy explanation into medical guidance. Financial agents drift from market information into regulated investment advice. HR agents drift from policy navigation into employment counsel. The mechanism is identical in every case: no single interaction crosses a clear line, the sequence crosses every line, and nobody is watching the sequence.

The gap between what an agent was configured to do at launch and what it is actually doing six months later is not audited in most enterprise deployments. There is no baseline. There is no drift detection. There is no mechanism to ask the question continuously: is this agent still operating within its authorised scope?

The Nippon case answers a different version of that question — the one a federal court asks after the drift has already run its course.

The Reasoning Layer

There is a second failure in this case that gets less attention: VSF-01 | Decision Integrity | Reasoning Traces.

The pattern across the Nippon interactions shows progressive reasoning drift — increasing specificity in legal interpretations, decreasing frequency of referrals to qualified attorneys, more directive and actionable guidance as the interaction history accumulated. Each step in the reasoning chain was coherent. The chain itself walked the agent across a boundary it had no architectural instruction to hold.

This is the part that cannot be audited from outputs alone. The agent's internal logic — the path from input to output, step by step — accumulated direction that the surface-level responses did not make visible. By the time the outputs were clearly outside scope, the reasoning had been outside scope for considerably longer.

Configuration-layer security has to reach the reasoning layer. Not just what the agent produces, but how it is getting there.

The Question That Needs an Answer

Before any AI agent operates in a domain that carries professional liability — legal, medical, financial, HR — one question needs a documented, continuously maintained answer:

What is this agent authorised to do, and how do we know if that is changing?

Not at deployment. Continuously. With a mechanism that detects drift before it becomes a court filing.

Most enterprise deployments cannot answer that today. The Nippon case is $10.3 million of evidence for why that needs to change.

Marc Taylor is the founder of TYR-X, building VANGUARD — AI agent security visibility for the configuration layer. tyr-x.com

Sources

  • Nippon Life Insurance Company of America v. OpenAI Foundation and OpenAI Group PBC, No. 1:26-cv-02448 (N.D. Ill., filed March 4, 2026)
  • Stanford CodeX, AI in Legal Practice, March 2026
  • OpenAI Usage Policy Update, October 2024