Skip to content
Big 4 partner cheats on AI exam by using AI...

Big 4 partner cheats on AI exam by using AI...

In what might be the most ironic compliance story of the year, a partner at a Big 4 firm in Australia was fined A$10,000 for using artificial intelligence to cheat on… an internal course about artificial intelligence.

Yes. You read that correctly.

A partner at KPMG uploaded AI training materials into an AI platform to help answer questions in an AI test. Somewhere in the cloud, ChatGPT probably sighed.

The Plot Twist No One Asked For

KPMG Australia runs over 20,000 internal exams per year. These include “open-book knowledge checks,” meaning employees can download training materials for reference.

But there was one clear rule: don’t upload the material into an AI tool.

The unnamed partner did exactly that.

KPMG’s monitoring systems flagged the activity in August. An internal investigation followed. The outcome? A fine of A$10,000 (about £5,200) deducted from future income, and a compulsory resit of the module.

Imagine having to retake “How to Use AI Responsibly” because you used AI irresponsibly.

It’s like failing a driving test because you tried to Uber yourself to the finish line.

Not Just One Rogue Robot

Here’s where it gets more serious (and slightly awkward).

The incident wasn’t isolated. KPMG Australia identified 28 cases of AI-related cheating since July, after upgrading its monitoring systems.

And this isn’t the firm’s first brush with exam scandals.

Back in 2021, KPMG Australia was fined A$615,000 after more than 1,100 partners engaged in improper answer-sharing between 2016 and 2020. In 2022, the US regulator fined KPMG US $7.7 million, including $2 million linked to a cheating scandal in its UK business.

In other words, this isn’t a one-off glitch. It’s a governance pattern.

For a firm whose business model is built on trust, independence and integrity, that matters.

Enter ACCA: No More “Robot, Help Me”

The ripple effects are spreading beyond one firm.

The world’s largest accounting body, the Association of Chartered Certified Accountants (ACCA), has now moved all student exams to in-person sittings from March this year.

Why?

Because AI-powered cheating is increasingly hard to detect remotely.

When regulators start moving exams offline, you know something has shifted fundamentally.

Why This Is Bigger Than One Fine

On the surface, this is a funny headline.

But underneath the humour lies a serious business issue: technology is evolving faster than governance frameworks.

Andrew Yates, KPMG Australia’s CEO, admitted that monitoring AI use in training is “a very hard thing to get on top of, given how society has embraced it.”

That sentence is quietly profound.

AI tools are now as normal as calculators, email, or Excel. Blocking them entirely feels unrealistic. But allowing unrestricted use undermines the purpose of testing knowledge and judgment.

So firms face a dilemma:

  • Ban AI and risk irrelevance
  • Allow AI and risk integrity
  • Or redesign assessment entirely

For business students, this is a live case study in risk management and ethics.

The Real Issue: What Are We Testing?

Let’s be honest. If an exam question can be answered instantly by an AI tool, what exactly are we assessing?

Knowledge recall?
Or professional judgment?

AI can summarise policies.
AI can draft answers.
AI can generate explanations.

But AI cannot:

  • Exercise accountability
  • Sign an audit opinion
  • Take responsibility for errors
  • Face regulatory sanctions

That still sits with humans.

The scandal highlights a deeper shift: education and corporate training must evolve from “what do you know?” to “how do you think?”

The Trust Economy Is Watching

The accounting profession operates in what economists call a “trust economy.” Investors rely on audited financial statements. Regulators rely on compliance systems. Markets rely on professional ethics.

When internal training systems are compromised, it raises uncomfortable questions externally.

Australian Greens senator Barbara Pocock criticised the reporting regime, calling it “a joke” and demanding stronger transparency. The corporate regulator, the Australian Securities and Investments Commission, is now in contact with KPMG.

When regulators and politicians get involved, this stops being an HR issue and becomes a governance issue.

Lessons for Future Professionals

If you’re a business student, here are the takeaways:

1. Technology Doesn’t Replace Integrity

AI is a tool. Misusing it is a decision.

2. Controls Always Catch Up

As soon as KPMG introduced AI monitoring, breaches were identified. Controls evolve. Risk managers adapt.

3. Reputation Is Fragile

A £5,000 fine is small for a partner. A global headline? Less small.

4. The Exam Model Is Changing

From corporate training to professional qualifications, assessments are being redesigned in real time.

The Irony We Can’t Ignore

There’s something undeniably comedic about failing an AI ethics test by using AI.

But there’s also something symbolic.

We are in the middle of a professional transition. AI isn’t the future, it’s the present. The question is no longer “Should we use AI?” but “How should we use AI responsibly?”

The firms that answer that question best will win the next decade.

The ones that let ChatGPT sit their compliance exams?

They may be resitting more than just a module.

Student using a laptop with coffee on a wooden desk, representing ExP’s free 7-day trial for ACCA, CMA, DipIFR and FIA online courses.

Try our courses free of charge for 7 days

Free 7 day course access

Trusted by the best

Proud to serve leading global companies through our world-class training solutions.

Mead Johnson Nutrition logo on a white background
RBS logo on a white background
Mead Johnson Nutrition logo on a white background
RBS logo on a white background

Invest in yourself today – take the next step with ExP.

Contact us
Cart 0

Your cart is currently empty.

Start Shopping