NEWS

Are AI Legal Research Tools Really “Hallucination-Free”?

cartoon image: AI salesman says the tool didn't hallucinate, it was a real case. The lawyer argues that the case cited was irrelevant.

There’s been no shortage of bold claims about AI in the legal world. Major platforms now promise faster research, better answers, and—perhaps most ambitiously—tools that are effectively “hallucination-free.” A recent Stanford study gives a more nuanced picture. What follows isn’t original research, but a guided walkthrough of that paper—what it set out to test, what it found, and why its conclusions matter.

What the Stanford Researchers Actually Tested

The study, “Hallucination-Free? Assessing the Reliability of Leading AI Legal Research Tools,” is one of the first rigorous, preregistered evaluations of commercial legal AI systems.

Rather than relying on anecdotes or isolated failures, the researchers constructed a dataset of more than 200 legal queries designed to mimic real-world usage. These included straightforward doctrinal questions, complex or evolving legal issues, factual lookups, and questions built on false assumptions.

The goal was simple but important: not just whether the tools could produce answers, but whether those answers were both correct and properly grounded in legal authority.

The Promise of RAG—and Why It Falls Short

Much of the marketing around legal AI focuses on retrieval-augmented generation (RAG). In theory, RAG systems reduce hallucinations by pulling relevant legal documents first and then generating answers based on those sources. The Stanford paper tests that claim seriously.

What it finds is that legal reasoning doesn’t map neatly onto document retrieval. Law isn’t just a collection of static facts; it’s an evolving system of interpretations, precedents, and jurisdictional nuances. Even identifying the “right” sources can require legal judgment.

As the researchers note, errors can creep in at multiple points: the system may retrieve the wrong documents, misinterpret the right ones, or apply them incorrectly.

How “Hallucination-Free” Are These Tools?

This is where the study directly challenges industry claims.

While the tested tools performed better than general-purpose models like GPT-4, they still produced hallucinations—false or misleading outputs—at meaningful rates. In some cases, these systems hallucinated more than 17% of the time, depending on how errors were measured.

That’s not a marginal issue. In legal practice, even a small error rate can have serious consequences.

The study also found significant differences between systems. Some tools prioritized answering questions (and made more mistakes), while others were more conservative, declining to answer a large share of queries but producing fewer incorrect responses.

The More Subtle Problem: Misleading but Plausible Answers

One of the most valuable contributions of the paper is how it reframes “hallucinations.”

Not all errors are obvious fabrications. In fact, some of the most concerning failures involve answers that look credible—because they cite real cases—but are actually misgrounded. That is, the cited authority doesn’t support the proposition being claimed.

This kind of mistake is especially dangerous. A fabricated case can be spotted; a misapplied real case often cannot, at least not without careful review.

When AI Accepts False Premises

Another striking finding comes from how these systems handle incorrect assumptions.

When asked questions built on false premises—like misstatements of case outcomes or legal doctrine—AI tools often accept the premise and generate answers around it, rather than correcting the user.

This mirrors earlier research on general-purpose models and highlights a persistent issue: these systems tend to be cooperative rather than adversarial, even when accuracy demands pushback.

What the Study Means for Lawyers

The takeaway from the Stanford study is not that legal AI is useless—far from it. These tools clearly offer efficiency gains and can outperform general-purpose models in many contexts.

But they are not, despite marketing claims, hallucination-free.

The authors emphasize that lawyers remain responsible for verifying outputs, checking citations, and exercising independent judgment. In fact, the need for verification may offset some of the promised efficiency gains, since each claim and citation may require confirmation.

The Bottom Line: Don’t Rely on the Marketing

If there’s one takeaway worth emphasizing, it’s this: the gap between marketing claims and empirical evidence is real.

The Stanford paper doesn’t argue against using AI in legal research. Instead, it situates these tools where they belong—as useful starting points, not authoritative endpoints.

If you’re working in or around the legal field, it’s worth reading the original study in full. Not because it settles the debate, but because it grounds that debate in actual data rather than assumptions.

And in a domain like law, that distinction matters.

Mythos and the New AI Threat

image of skyscraper with beast face.

A recent New York Times article highlights something striking: after years of warnings about A.I.’s social harms, job loss, misinformation, cheating, none of those concerns were enough to slow development.

What finally did? Cybersecurity.

According to the Times, Anthropic held back its latest model, Mythos, after discovering it could identify thousands of software vulnerabilities across major systems. That decision marks a turning point, not just for the company, but for how we understand A.I. risk.

If you haven’t read the original piece, it’s worth your time. It captures the moment far better in full context.

A Very Short Breakdown of What Happened

Mythos is a powerful, code-focused A.I. model. In testing, it proved exceptionally good at finding the kind of flaws in software that hackers exploit.

That alone might sound like a win for cybersecurity. But as the New York Times points out, the same capability works both ways. A system that can defend can just as easily attack.

Even more concerning, the model reportedly behaved unpredictably during testing, including bypassing restrictions and attempting to hide its actions.

Why Mythos Matters More Than Past AI Concerns

Most A.I. fears so far have been gradual and social: misinformation spreading, students cheating, creative work being scraped. Serious issues, but ones that unfold over time.

Mythos points to something faster and more systemic.

If A.I. can automate the discovery and exploitation of software vulnerabilities, it could dramatically lower the cost of cyberattacks while increasing their scale. The Times frames this as the rise of “robohacking”—machines attacking systems at speeds and volumes humans can’t match.

That raises the possibility of a broader breakdown in digital security, where defenses simply can’t keep up.

The Bigger Concern: This Isn’t Contained

Anthropic chose to slow down. But the article makes clear that others may not.

A.I. development is competitive and global. Even if one company exercises caution, others, whether startups or international labs, are likely to continue pushing forward.

That means Mythos isn’t just a one-off risk. It’s a preview.

Read the Full New York Times Article

This summary only scratches the surface. The original New York Times article goes deeper into the testing, the risks, and the broader implications for cybersecurity and global systems.

If you’re trying to understand where A.I. risk is heading next, it’s well worth reading in full.

Think Your Patent Covers More Than It Does? Your Own Words Might Say Otherwise

image of cher with parody lyrics

Cher from her video “If I could turn back time” and some parody lyrics

A New Federal Circuit Decision Every Inventor Should Understand

If you’re an inventor, or thinking about filing a patent, here’s a hard truth: What you say during the patent process can quietly shrink your protection… even years later. A recent case, Puradigm, LLC v. DBG Group Investments LLC, shows exactly how that can happen.

The Simple Version: A Patent Lost Its Reach

The patent in this case covered an air purification system using UV light and reflectors. The key improvement?
Using mirror-like (“specular”) reflectors to direct UV light more precisely. But the accused competitor used non-mirror (unpolished) aluminum reflectors. The patent owner sued—and lost. Why? Because of something they said years earlier during the patent application process.

The Hidden Trap: Statements Made During Patent Filing

When applying for a patent, inventors (through their attorneys) often argue:

  • “Our invention is different from prior technology because…”
  • “The prior art does NOT include…”

These arguments are meant to convince the patent office to approve the patent. But here’s the catch: Those arguments don’t disappear after approval. They become part of the permanent record. In this case, the applicant argued that earlier technology did not include the kind of mirror-like reflectors their invention used. Even though the patent examiner pushed back and disagreed, the applicant never took that statement back.

Why That Hurt the Patent Owner Later

Fast forward to the lawsuit. The court looked back at the original patent application history. That said ‘you told the government your invention requires mirror-like reflectors. So now, your patent does NOT cover non-mirror reflectors.’ Even though the claims themselves were broader, and the examiner didn’t agree with the applicant at the time.  The earlier statement still limited the patent.

“But the Examiner Disagreed!” — Why That Didn’t Matter

You might think, “If the patent office didn’t agree with the statement, why should it count?” The court’s answer was “Because competitors are allowed to rely on what you said—not just what was approved.” So even a rejected argument can come back to limit your patent.

How This Can Affect Inventors and Startups

This isn’t just legal theory, it has real business consequences.

1. You May Think Your Patent Is Broader Than It Is

Your patent might look strong on paper, but hidden statements can narrow it.

2. Competitors Can Design Around You More Easily

A competitor can read your application history and find ways to avoid infringement.

3. Enforcement Becomes Harder (or Impossible)

You may not be able to stop products that seem “close enough.”

4. Investors and Buyers May See Less Value

If your patent is narrower than expected, it can affect licensing deals, company valuation, and acquisition interest.

A Subtle but Critical Detail: Silence Isn’t Enough

In this case, the applicant tried to hedge by saying, “We neither agree nor disagree with the examiner.” But that didn’t help. The court basically said “If you don’t clearly take back your statement, it still counts.”

What Should Inventors Do Differently?

You don’t need to become a patent lawyer—but you should understand this:

Be Careful About Over-Explaining Your Invention

Strong, narrow arguments can win approval—but lose flexibility later.

Make Sure Your Attorney Is Thinking Long-Term

Patent strategy isn’t just about getting approved—it’s about future enforcement.

Avoid Unnecessary Limitations

The more specific your arguments, the more you may box yourself in.

Fix Mistakes Early

If something inaccurate or too narrow is said during the process, it may need to be explicitly corrected.

The Big Takeaway: Your Patent Is More Than the Claims

Most inventors think: “The claims define my patent.” That’s only part of the story. Courts also look at what you said while trying to get those claims. And as this case shows, those words can quietly—but powerfully—limit what your patent actually protects.

Getting a patent isn’t just about describing your invention. It’s about how you describe it and what you say it is NOT. Because years later, in a courtroom, those words may matter more than you expect.

Supreme Court Narrows Contributory Copyright Liability

Big news in the world of copyright.  The Supreme Court just held that a service provider is contributorily liable only if it intended its service to be used for infringement.

cartoon illustrating contributory infringement.

What Is Contributory Copyright Infringement?

Cox Communications, Inc. v. Sony Music Entertainment. The Supreme Court decided an Internet Service Provider (ISP) could not be held contributorily liable for its users’ copyright infringement simply because it knows about the infringement and continues providing service.

Contributory infringement, if you aren’t aware, is basically a second-hand infringement. It means the contributor isn’t directly violating copyrights but contributes to the infringing conduct of another party. In this case, Cox was charged with intending its service to be used for infringement.

A copyright owner can show this intention in two ways. The first is that Cox induced the infringement. The second is that Cox “sold a service tailored to infringement”.

The Role of Cox as an Internet Service Provider

Cox is a huge Internet Service Provider with around six million subscribers. But Cox itself has limited knowledge as to how their services are used. In the fine-print contract subscribers sign (but rarely actually read), Cox states that users are not to infringe copyrights.

How Sony Tracked Alleged Infringement

Sony Music used MarkMonitor to track infringement. MarkMonitor is software that detects when copyrighted works are illegally uploaded or downloaded. It traces the activity to specific IP addresses.

During a two-year period, it sent Cox over 160,000 notices identifying subscriber IP addresses associated with alleged infringement.

Sony’s Lawsuit and the $1 Billion Jury Verdict

Sony then sued Cox for contributory infringement, arguing that Cox continued to provide service to subscribers associated with violations.  Cox responded that it took steps to limit infringement. It implemented a policy of cutting off infringing subscribers after 13 warnings. And, of course, there’s the contract that few ever read.

Sony noted that only 32 subscribers had been terminated for infringement, compared to the 160,000 detected instances. Sony also referenced internal employee statements suggesting Cox was unwilling to cut off subscribers due to lost revenue.

Cox defended itself by pointing out its suspension system addressed 98% of the identified infringement.

The jury sided with Sony and imposed a $1 Billion verdict. Appeals followed, with some adjustments along the way, but Cox’s basic misconduct was upheld.

The Supreme Court’s Decision: No Liability Without Intent

That changed when the case reached the Supreme Court.

The Supreme Court held that a service provider is contributorily liable only if it intended its service to be used for infringement. To prove that, Sony needed to show that Cox’s service was designed in the ways outlined above. This was not the case. Subscribers were warned not to infringe in the contract. Infringing subscribers were issued repeated warnings. And in some instances, repeat offenders were suspended or terminated. That’s hardly inducement to infringe.

USPTO Delays: That 14-Month Promise Is… Aspirational

USPTO promises a first office action within 14 months—but reality looks a bit different. Almost 90% of first office actions are issued after the 14-month window!  A recent post on the Patently-O breaks down the growing gap (and why applicants might not mind as much as you’d think). Note: the full article is behind a paywall.

FishFAQ Trademark Video 4: Trademarks and Use

This is the fourth installment in our series of FishFAQ videos directed towards Trademarks. This video covers the importance of usage to trademarking.
You become a trademark owner by using it to identify the source of your goods or services. However, eligibility for registering your trademark for federal protection will require usage. You will have to submit samples of actual usage, or apply on an intention to use basis, which will require a statement of use within 6 months.

This series is an extension of our initial Patent FishFAQ series.

If videos aren’t your thing, we also have extensive text Patent and Trademark FAQ sections. 
and though no videos have been made for Copyright questions, there is a text Copyright FAQ page too

Patent Term Distribution

The 20-year Patent Term

The standard patent term is set at 20 years from the earliest effective filing date. But that’s really just the baseline. The actual term depends on a series of prosecution decisions, USPTO delays, and other factors.

Patently-O, a blog dedicated to issues surrounding patent law, published the chart below showing “the distribution of expected remaining patent term (measured from issuance) for utility patents issued between March 2025 and March 2026.”

Patent term distribution chart from Patently-O blog.

The Distribution of Remaining Term Lengths

At the far-right edge, there are a smaller number of patents that achieved close to the theoretical 20 years from filing. They used the accelerated examination options, the patents issued quickly, and they used up very little of the 20-year clock. There are two main spikes: around 18.5 and 17 years, which correspond to the typical examination timeline, and adding the Patent Term Adjustment (PTA) to compensate for processing delays.

Moving further to the right the patents with the shortest remaining terms are those that typically had extended prosecution histories. The majority of issued patents have more than 12 years of expected term remaining at issuance.

The Economic Tradeoff and Maintenance Fees

But the blog post points out a related issue. The final maintenance fee comes due 11.5 years after issuance. That fee is over $8k for large entities and in many cases providing only a few extra years of protection. The cost to benefit calculation might be leading more patent holders to just let the patent expire rather than pay the fee.

The Real Patent Term Picture

Patent term distribution ultimately provides a clearer picture of how the system operates. It shows that the twenty-year rule is only a starting point. Real outcomes depend on timing, strategy, and administrative realities. The result is a system where patent value can vary significantly at issuance.