The risk nobody's talking about

Everyone's worried about AI hallucinating — making things up. That's a real risk, and it's getting smaller with every model update. The risk that isn't shrinking is the one nobody's watching: AI that gives you a perfectly accurate answer to 60% of the problem and presents it as the whole picture.

It doesn't simplify on purpose. It simplifies because that's what language models do — they pattern-match toward the most probable answer, not the most complete one. They drop requirements, ignore edge cases, and skip the parts that are hard to reason about. Then they hand you something polished and confident. Your team looks at it, thinks "this is good," and executes. Nobody checks what's missing because nothing looks missing.

That's the completeness gap. And it's the gap that hits your bottom line — not because the AI was wrong, but because it wasn't finished. The gap doesn't close with better models. It closes with better judgment — specifically, the judgment to pause between AI output and business action, and check what was left out.

That judgment is what I call the AI Instinct.

Your team already has it for people — you can tell when an employee is winging it versus when they've done the work. You don't have it for AI yet. Nobody does. That's what I build.

Where most businesses sit right now

AI CAPABILITY → HUMAN JUDGMENT CALIBRATION → Two years ago Most companies now THE TARGET Understands deeply. Hasn't deployed.

The journey from bottom-right to top-right isn't about deploying more AI. It's about organising the judgment your team already has.

What "correct but incomplete" looks like

This is a real AI output for a common business task. Tab one is what your team would see. Tab two is what they'd miss.

Task given to AI: "Draft a risk assessment for onboarding a major new subcontractor for a mid-sized construction company."
AI output:
Key Risks Identified:

1. Financial stability — Subcontractor may lack capital to sustain project through payment cycles. Recommend credit check and financial references before engagement.

2. Insurance coverage — Verify public liability, workers compensation, and professional indemnity coverage meets project minimums. Request certificates of currency.

3. Safety record — Review SWMS documentation and incident history. Require site-specific induction before commencement.

4. Capacity constraints — Assess whether subcontractor can resource the project without compromising existing commitments. Request current project list.

Recommendation: Proceed with onboarding subject to satisfactory completion of the above checks. Estimated due diligence timeline: 5–7 business days.
Task given to AI: "Draft a risk assessment for onboarding a major new subcontractor for a mid-sized construction company."
AI output — with completeness audit:
Key Risks Identified:

1. Financial stability — Subcontractor may lack capital to sustain project through payment cycles. Recommend credit check and financial references before engagement.

2. Insurance coverage — Verify public liability, workers compensation, and professional indemnity coverage meets project minimums. Request certificates of currency.

Completeness Gap — Contractual liability chain

No assessment of back-to-back liability provisions. If the subcontractor's insurance doesn't mirror head contract obligations, the principal contractor carries the residual risk. This is the clause that costs $200K+ when it goes wrong.

3. Safety record — Review SWMS documentation and incident history. Require site-specific induction before commencement.

Completeness Gap — Industrial relations exposure

No assessment of subcontractor's enterprise agreement status, union coverage, or workforce classification. On a union site, an unaligned subcontractor can trigger a site-wide stop. On a non-union site, misclassified workers create Fair Work liability.

4. Capacity constraints — Assess whether subcontractor can resource the project without compromising existing commitments. Request current project list.

Completeness Gap — Existing client conflict

No check for whether the subcontractor is currently engaged by a competitor or has contractual non-compete/exclusivity obligations that conflict with your project.

Recommendation: Proceed with onboarding subject to satisfactory completion of the above checks. Estimated due diligence timeline: 5–7 business days.
Completeness Gap — Termination and step-in rights

No assessment of what happens if the subcontractor fails mid-project. No step-in provisions, no performance bond requirement, no staged payment structure tied to milestones. The recommendation to "proceed subject to checks" assumes the checks will catch everything. They won't — because the checks themselves are incomplete.

100%
Accuracy — Nothing in the AI output is incorrect
45%
Completeness — Four critical risk categories missing

That's the completeness gap. That's what the AI Instinct catches.

Three layers of organised judgment

This isn't about making AI smarter. It's about making your team's oversight systematic — so the gap between AI output and business reality gets caught before anyone acts on it.

1

Plan-Evaluate-Patch

Every AI output gets audited before execution. The system forces the AI to plan without acting, then runs a second evaluator layer against your actual business logic — the rules, constraints, and exceptions your team knows but the AI doesn't. Gaps get flagged. The AI fills them before anyone sees the output. Your team gets the result after the completeness check, not before.

2

Intelligence Routing

Different tasks need different AI models with different levels of oversight. A routine email summary doesn't need the same governance as a risk assessment or a client proposal. I build the routing layer that matches model capability to task complexity and consequence — commodity models for narrow, defined tasks; frontier models with full audit trails for anything that touches your bottom line.

3

Single Source of Truth

The document your team approves is the same artifact that drives the AI. One file to update. When your team changes a process, the AI's behaviour updates with it. No version drift. No gap between what the documentation says and what the system does. This is what makes the architecture survive model changes — when the next AI ships, you update one file and the system adapts.

Before the system, I build the intent

Before building anything, I identify the trade-off hierarchy that governs every AI decision — when speed conflicts with thoroughness, which wins? Human operators absorb these rules through experience. I make them explicit on day one.

Three things that protect your investment

These aren't marketing lines. They're the rules I work by — and they exist because the alternative burns money.

I won't sell you technology you don't need

A failed AI project doesn't just waste money — it vaccinates your entire team against trying again. I start by asking what to stop doing altogether. If the honest answer is "you don't need AI for this," I'll say it.

I won't lock you into one AI vendor

Everything I build works across models. When the next model ships — and it will ship sooner than anyone expects — your systems keep running because the architecture is independent of any single AI. The judgment layer persists regardless of which model sits underneath it. You're not buying a dependency. You're buying a capability that survives the next model change.

I won't pretend AI output is trustworthy by default

AI produces output that is correct and incomplete — with total confidence. No hesitation, no "I'm not sure." Every output runs through a completeness audit before your team acts on it, with clear handoff points where a human makes the call.

Everything I build runs on frontier generative AI — Claude, ChatGPT — orchestrated into operational systems with judgment layers. No machine learning pipelines. No six-month model development. Working systems in weeks.

What this looks like in practice

100+
Specialised AI Workflows
BHP
20 Yrs Audit IT BD Ops
2–4
Weeks to First Result
0
Vendor Lock-In

3 hours → 15 minutes

Construction teams now generate complete briefing packs automatically — safety, compliance, site specs — from one source of truth.

Skipped regularly → Every job, automatically

Compliance checks that were being missed now run on every job, enforced consistently, zero additional workload.

Full day → Under an hour

Proposals assembled from five systems — with completeness checks that catch omissions people miss under deadline pressure.

2+ days → Same-day delivery

Monthly reports now consolidate automatically with a judgment layer that verifies narrative matches the numbers.

Background

Decades in industrial operations at BHP · Former Head IT Audit, IT Strategy, Corporate Planning, and VP International · Advised C-level executives on complex operational challenges. I know what operational pressure feels like from the inside.

Find out where the instinct gaps are

A structured diagnostic where we identify where AI is creating value in your operations and where it's creating invisible risk. You'll see the completeness gaps in your own workflows — the same kind of gaps the demo above revealed. Concrete findings, not a slide deck. If the honest answer is "not yet," I'll tell you.

No retainer required  ·  60 minutes  ·  If the answer is “not yet,” I’ll tell you.

Book the AI Instinct Diagnostic