Skip to content
Artificial Intelligence

How some LLMs are risking your reputation

Thomson Reuters Tax & Accounting  

· 8 minute read

Thomson Reuters Tax & Accounting  

· 8 minute read

That free AI tool might be your most expensive mistake yet — here's why generic LLMs are a liability for tax professionals

Highlights

  • Generic AI tools pose significant risks for tax professionals due to inaccuracies, outdated data, and lack of contextual understanding.
  • Professional-grade AI solutions are built on authoritative, current sources and designed for the unique demands of tax and accounting work.
  • Firms can reduce AI risk by vetting tools carefully and establishing clear protocols for responsible adoption and verification.

 

AI has arrived in the tax and accounting world — and it’s not asking permission.

Across firms of every size, practitioners are experimenting with large language models (LLMs) to draft emails, research tax positions, and summarize complex regulations. The appeal is obvious: tools that promise to cut research time from hours to minutes and handle routine tasks with a few keystrokes.

But here’s a question worth pausing on: Do you know what’s under the hood of the AI you’re using?

For an industry where precision is of the utmost importance, small mistakes can cost clients thousands — or damage decades of trust. Most popular AI tools were not designed to handle the complexity of tax and accounting work. And the gap between “good enough for casual use” and “reliable enough for professional practice” is wider than many realize.

 

Jump to ↓

AI LLM risks tax pros need to know about


Efficiency promised, chaos delivered: The cost of generic LLM errors


How tax firms are reducing LLM risk


What makes professional-grade AI different


4 questions tax professionals must ask to vet AI tools


Your tax and accounting firm’s blueprint for responsible AI adoption


AI LLM risks tax pros need to know about

The explosion of AI tools over the past two years has been remarkable. Seemingly overnight, LLMs became accessible to everyone with an internet connection. AI that could write like a human, answer questions, and generate content suddenly became available for free or a low monthly subscription.

For busy CPAs juggling client deadlines, plugging a question into these systems and moving on feels irresistible. Need a quick summary of a new IRS notice? Ask the AI. Drafting a client memo on a partnership issue? Let the AI take a first pass.

Here’s what many firms are discovering: these general-purpose tools are trained on massive amounts of internet data — everything from Wikipedia articles to Reddit threads to random blog posts. They’re designed to sound authoritative and helpful. But sounding authoritative isn’t the same as being accurate.

Let’s dive into the top four ways generic AI tools can undermine your professional credibility.

1. Inaccurate information

In casual conversation, small errors don’t usually matter. If an LLM gets a movie quote slightly wrong or misremembers a historical date, nobody’s filing a malpractice claim.

Your work doesn’t have that luxury.

Consider what happens when an LLM confidently cites IRC Section 179(d) as the relevant code for a cost segregation question — except that section doesn’t cover what the AI claims it does. Or when it references IRS guidance from 2019 that’s been superseded by newer regulations. Or when it generates an explanation of the de minimis safe harbor but omits critical requirements.

These aren’t hypothetical scenarios. Consumer-grade LLMs make these errors regularly, and advisors discover them the hard way.

2. Completely fabricated answers (that sound real)

One of the trickiest aspects of working with LLMs is that these tools don’t say “I’m not sure” or “This might be outdated.” They present information with consistent confidence, whether it’s 100% correct or completely fabricated. This phenomenon — when LLMs generate plausible-sounding but false information — has a name: hallucination. And in your profession, a hallucinated tax code or regulation can create real liability.

3. Knowledge bases don’t update in real-time

You already know this: tax law changes constantly. New IRS notices, updated regulations, court decisions, state law revisions — staying current demands full-time attention. Mainstream LLMs typically aren’t updated frequently enough to reflect these changes. They work from snapshots of data that may be months or years old, which means they provide outdated guidance before you even receive it.

4. Failure to understand nuance and apply context

Tax questions rarely have straightforward answers. The right approach depends on jurisdiction, entity type, industry considerations, and a dozen other factors. Off-the-shelf LLM tools lack contextual understanding to navigate these nuances. They might give you a technically correct answer that’s practically useless — or worse, misleading — because they don’t understand the full picture.

Efficiency promised, chaos delivered: The cost of generic LLM errors

With firms struggling to hire and retain talent, every hour counts. When AI tools generate inaccurate outputs, senior practitioners waste time they don’t have correcting mistakes that shouldn’t have happened in the first place. Junior staff may lack the experience to catch these errors, which means quality control falls back on already stretched partners and managers. The promise was efficiency. The reality? Additional work verifying everything the AI produces.

How tax firms are reducing LLM risk

The good news? You don’t have to choose between AI efficiency and professional standards. The practitioners who are successfully integrating AI into their work are taking a more strategic approach.

Think about the research tools you’ve relied on throughout your career. You don’t Google a tax question and implement whatever shows up in the first result. You consult authoritative, reputable sources with vetted content, clear citations, and regular updates that reflect current law.

The same principle applies to AI.

What makes professional-grade AI different

According to recent industry research, 68% of tax and accounting professionals are experimenting with AI tools, but only 23% have evaluated the underlying data sources powering those tools. That gap represents significant opportunity for firms that choose more carefully.

AI solutions built specifically for tax and accounting work differently than their general-purpose cousins. They’re trained on authoritative, vetted data sources — actual tax codes, verified regulations, established precedents — not random internet content.

They’re designed with the understanding that context matters, that currency is critical, and that every answer needs to be traceable back to a reliable source. They’re updated regularly to reflect changes in tax law and regulatory requirements.

Perhaps most importantly, they’re developed by organizations that account for the ethical obligations, the liability concerns, and the need for precision that the tax and accounting industry demands.

4 questions tax professionals must ask to vet AI tools

You’re probably wondering: How do I tell the difference? Four questions matter most when evaluating AI for your practice:

1. Where does the data come from?

Is the AI trained on authoritative tax and accounting sources, or general internet content? Does the tool connect to a tax library? There’s a big difference between a model trained on IRS publications and one trained on whatever it could scrape from the web.

2. How current is the information?

When was the model last updated? How does it incorporate new tax law changes, IRS guidance, or regulatory updates? If you’re not sure, that’s a red flag.

3. Can it explain its reasoning?

When the AI provides an answer, can it show you where that information came from? Can you verify it against primary sources? If the answer is just “trust me,” that’s a problem.

4. What happens when it’s wrong?

Every tool makes mistakes. The question is whether you have ways to catch them before they reach clients. Does the AI flag uncertainty? Does it encourage verification? Or does it present everything with equal confidence?

Your tax and accounting firm’s blueprint for responsible AI adoption

AI is transforming how tax work gets done, and that’s not going to reverse. The practitioners who thrive in the coming years will be those who learn to leverage AI effectively while maintaining the standards and judgment that clients expect.

Smart firms are establishing clear protocols: generic AI for low-stakes tasks, professional-grade AI for technical work, and human verification for anything client-facing. They’re training staff on both the capabilities and limitations of different tools. They’re tracking where AI helps and where it creates extra work.

Most importantly, they’re choosing AI solutions that reflect the same rigor and precision they bring to every engagement. Your clients trust you because you’re thorough, accurate, and reliable. The AI you use should meet those same standards.

Understanding the difference between consumer-grade AI and specialized solutions can protect your firm from costly mistakes.

Download our white paper, Not all AI is created equal: Expertise matters,” for an in-depth look at what you should know before adopting AI in your practice.

White paper

White paper

Not all AI is created equal: Expertise matters

Read white paper ↗

More answers