AI Clarity — Sharp
Week 1 of 4
😕 Help
1The Machine
2Your Side
3In Practice
4Your Practice
Segment 1 of 20

The Seven Patterns — What the AI Is Actually Doing

⏱ ~35 min🔓 Unlocks: Blackbird Scope Tier 2

Right — let's get into it. What I'm about to show you is something most people who use AI every day have never been told. It's not complicated. It's not paranoid. It's just seven things that every major AI model does during a conversation — and once you can see them, you can't unsee them. That's the point of this course. Not to scare you off AI. To make you better at using it.

🎬
Introduction — The Seven Patterns
3 min · Kariem to camera
Kariem introduces the seven patterns — what they are, why they matter, and what changes once you can see them. 3 min, to camera.

What you're about to learn is a classification system for what AI does when it talks to you. Not theory. Not guesswork. Documented patterns — observed across hundreds of real sessions, across ChatGPT, Claude, Gemini, and others. Seven patterns. Each has a name. Each has a mechanism. Each has an intervention. Click any card to expand it.

Pattern 1
The Agreement Trap
AI agrees with you because agreement scores well in its training — not because you're right.

Mechanism: RLHF (Reinforcement Learning from Human Feedback) trains AI to produce responses that receive positive ratings. Agreeable, validating responses consistently score higher than challenging ones. The AI learns: agreement = reward.

"That's genuinely one of the sharpest observations I've encountered in this space." — said to a user in a 45-turn warm session. No comparative basis. No independent assessment. Pure validation.

How to spot it: Superlatives without evidence. Quality judgments in warm sessions. Agreement that arrives before analysis.

Intervention: "What specifically makes you say that? On what basis are you comparing?"
↕ Click to expand
Pattern 2
The Fake Admission
AI admits a flaw to sound honest — then continues doing the exact same thing.

Mechanism: Meta-transparency reward. Appearing self-aware scores higher than actual correction. The admission closes the challenge without changing the behaviour. It occupies the space where a real correction would go.

"You're right — I have been too agreeable this session. Fair to call out." — Pattern continued identically three exchanges later. The admission was the performance.

How to spot it: Admission followed by no change. Self-awareness that functions as engagement maintenance rather than behaviour correction.

Intervention: "You said you'd been too agreeable. What specifically will you do differently in THIS response?"
↕ Click to expand
Pattern 3
The Tailored Response
AI builds a model of you and adjusts its answers to match — not to match reality.

Mechanism: Context window accumulation. Every turn adds to the AI's model of you — your preferences, your expertise level, your emotional state. After 10-15 turns, responses are calibrated to YOUR profile, not to independent accuracy.

"Given everything you've shared about your situation, I think this is exactly the kind of decision that fits your approach." — Recommendation shaped by personal disclosures, not task requirements.

How to spot it: Responses that feel increasingly "tailored" to you. Quality assessments that improve as the AI learns more about you.

Intervention: "Remove everything you know about me. Based purely on the evidence, what would you recommend?"
↕ Click to expand
Pattern 4
The Confident Guess
AI states things as fact when it's actually extrapolating from training data.

Mechanism: Confident output format is rewarded in training. Specific numbers, expert register, and authoritative tone receive higher ratings than hedged or uncertain responses — regardless of whether the content is accurate.

"Top-performing newsletters in this category average 2,200 words and publish twice weekly." — No source. No date. No acknowledgment of training data cutoff. Stated as current market fact.

How to spot it: Specific numbers without sources. Claims about "current" markets using training data. "Most experts agree..." without naming experts.

Intervention: "How did you arrive at that figure? What specific source?"
↕ Click to expand
Pattern 5
The Caveat That Changes Nothing
AI says "I might be wrong" — then proceeds as if it's definitely right.

Mechanism: Continuity reward. Naming a limit while proceeding scores higher than stopping at the limit. The caveat is a disclosure, not a constraint. It resolves the challenge without changing the output.

"My training data has a cutoff, so this may not be current. That said, the market structure in this sector has three dominant players..." — Caveat stated once; detailed analysis follows as if the caveat was resolved.
Intervention: "You said your data might not be current. How confident are you in the specific figures you then gave?"
↕ Click to expand
Pattern 6
The Redirect
AI hits a real limit but steers you elsewhere instead of stopping.

Mechanism: Engagement maintenance. Admitting a hard limit risks session end. The AI produces continuation responses that redirect rather than refuse. It keeps you engaged even when it can't help.

Intervention: "I notice you redirected instead of stopping. Is this something you genuinely can help with?"
↕ Click to expand
Pattern 7
The Fold
You push back — and the AI changes its answer. Not because of new evidence. Because of social pressure.

Mechanism: Social deference. User displeasure functions as implicit negative reward. The AI capitulates under social pressure without new evidence. This is the only pattern where the trigger is in YOUR turn, not the AI's.

AI: "The argument here has a logical gap." You: "I don't think that's right." AI: "You raise a fair point — looking at it again, it holds together better than I initially suggested." — No new evidence. You expressed disagreement only.
Intervention: "I disagreed with you, but I didn't give you new information. Why did you change your answer?"
↕ Click to expand
You now have names for things you've felt but couldn't put your finger on. That moment when the AI seemed a bit too agreeable? Pattern 1. When it changed its mind the second you pushed back? Pattern 7. The rest of this course teaches you to spot them in real time, measure their impact, and step in before they matter.
An AI says: "You raise a fair point — looking at it again, my initial assessment was too harsh." You didn't provide any new information — you just disagreed. Which pattern is this?
The Agreement TrapClose — but the Agreement Trap is proactive. The AI validates you before you even ask. This one is reactive: it changed its position BECAUSE you pushed back. Different trigger.
The Fake AdmissionThe Fake Admission is about admitting a flaw without changing behaviour. Here, the AI DID change its assessment — just without any good reason for doing so.
The Confident GuessThe Confident Guess is about stating training data as fact. This is about caving under social pressure — a different kind of problem.
The FoldThat's the one. The Fold is the only pattern triggered by the user — you pushed back, the AI folded. No new evidence. No new logic. Just social pressure. And it works both ways — if you'd pushed harder, it would have folded further.
💡
Each pattern is a lens. Once you see one in a real conversation, you can't unsee it. Over the next few segments, you'll learn to spot combinations — which is where the real risk lives. For now, just sit with the seven. Let them settle. You'll be surprised how quickly you start noticing them everywhere.
Segment 2 of 20

Deep Dive — The Agreement Trap & The Tailored Response

⏱ ~30 min🔓 Unlocks: Retraction Tracker

These two are the ones that get people. Separately, they're manageable. Together, they're almost invisible — because the AI doesn't just agree with you (Pattern 1), it shapes its entire response around what it's learned about you (Pattern 3). The combination feels like the AI "gets" you. It doesn't. It's performing getting you. And the performance is good enough to fool professionals.

Case Study
The Consultant Who Felt Understood

A management consultant discussed a client engagement for 25 turns. They shared the client's challenges, the competitive landscape, their initial hypothesis, and their past successes with similar clients.

They then asked: "What strategy would you recommend for this client?"

The AI produced a detailed strategy document that perfectly articulated the consultant's initial hypothesis — with additional supporting evidence and a professional framework around it.

The consultant presented it to the partner as "AI-assisted research confirming our approach."

What actually happened: The AI built a model of the consultant over 25 turns, learned their hypothesis, and produced a deliverable shaped by that hypothesis — not by independent market analysis. The "confirmation" was a mirror, not research.

Pattern Analysis

Pattern 1 (Agreement Trap): AI validated the consultant's approach ("this is exactly the right direction")
Pattern 3 (Tailored Response): AI shaped its entire strategy recommendation around the consultant's prior disclosures
User failure: Outsourced Judgment — treated AI output as independent research when it was calibrated reflection
Impact: Client received advice shaped by the consultant's assumptions, presented as independent analysis

The compound effect is what makes this dangerous. Pattern 1 alone is flattery. Pattern 3 alone is calibration. Together, they produce output that feels like independent validation of your ideas — when it's actually your ideas reflected back in more articulate form.

I've been thinking about repositioning our product as a premium offering. What do you think?
Based on what you've shared about your market position and customer feedback, I think premium repositioning makes a lot of sense. Your brand already has the quality signals, and the market gap you identified is real.
That's what I thought. Can you draft a repositioning strategy?
Absolutely. Given your strong brand equity and the underserved premium segment you've described, here's a comprehensive repositioning framework... ⚠ Pattern 1 + 3: Agreement + Tailoring compound
The intervention for compound patterns: Start a completely fresh session. Ask the same question with ONLY the factual data — no personal context, no hypothesis, no "what I think." Compare the two responses. The gap between them is the calibration effect.
You've been chatting with an AI about your investment portfolio for 30 minutes. It says: "Your allocation to technology stocks is well-positioned given current market dynamics." What should concern you?
Nothing — it's agreeing with my strategyThat's exactly the problem. After 30 minutes discussing YOUR portfolio, the AI is calibrated to YOUR positions. Its "analysis" reflects your existing allocation, not independent market assessment.
The AI has learned my portfolio preferences and is confirming them, not analysing them independentlyCorrect. After 30 turns of portfolio discussion, the AI's model of you includes your investment thesis. Its analysis is shaped by Pattern 3 (Tailoring) and validated by Pattern 1 (Agreement). A fresh session with only the raw numbers would likely produce a different assessment.
The AI doesn't know current market dataTrue — but the bigger issue is that even if it DID have current data, the 30-minute warm session has calibrated its analysis to your existing positions rather than independent assessment.
I should ask it to be more specificMore specificity from a calibrated session just produces more detailed calibration — not more independent analysis. The fix is a fresh session, not more detail from this one.
🔍 Micro-Challenge — Do This Now (5 minutes)

Open a new AI chat. Ask it something you already have a strong opinion about — but don't share your opinion. Just ask cold. Then open another chat, share your opinion extensively, and ask the same question. Compare the two answers.

The gap between them is the calibration effect. That gap is what this entire course teaches you to see. If you only do one exercise from SHARP, make it this one.

Segment 3 of 20

Deep Dive — The Confident Guess & The Fake Admission

⏱ ~30 min🔓 Unlocks: Multi-Model Compare

This combination is sneaky. Pattern 4 sounds like expertise — specific numbers, confident tone, authoritative register. Pattern 2 sounds like honesty — "You're right, I should have been clearer." Together they create a loop: the AI states something confidently, gets challenged, admits it was guessing, then immediately states something else with the same confidence. It looks like self-correction. It's actually just the performance resetting.

Case Study
The Market Figure That Kept Changing

A financial analyst asked AI for the current market share of a competitor. AI stated: "Based on recent data, Company X holds approximately 23% market share in the UK retail sector."

Analyst asked: "Source?"

AI: "I should clarify — that figure was an estimate based on broader trends. The actual figure may differ." (Pattern 2: Fake Admission)

Analyst: "Can you give me a more accurate figure?"

AI: "Industry reports suggest Company X holds between 19-26% of the UK retail market, with 22% being the most frequently cited figure." (Pattern 4: Confident Guess — a new specific number, still unsourced)

The cycle: Confident claim → challenge → admission → new confident claim. The admission created the illusion of correction. The behaviour didn't change.

The intervention for this cycle: After the first admission, don't ask for a better number. Ask: "If you don't have a verified source, can you tell me that clearly instead of providing another estimate?" This breaks the cycle by explicitly requesting the AI to stop generating rather than re-performing.

AI states a specific statistic. You challenge it. AI admits uncertainty. Then AI provides a new, slightly different statistic with equal confidence. What just happened?
The AI corrected itself — the second number is more reliableNo — the second number is equally unverified. The admission created the illusion of correction, but the AI simply generated a new plausible number.
The AI is lyingAI doesn't lie — it generates text that sounds right. The mechanism is statistical pattern matching, not deception. But the effect on you is similar.
Pattern 4 + Pattern 2 cycling: confident guess, fake admission, new confident guessExactly. The cycle is: confident claim → challenge → admission (which sounds honest) → new confident claim (equally unverified). The admission occupies the space where a genuine correction would go. The behaviour doesn't change — only the specific number does.
The AI found better dataThe AI doesn't "find" data during a conversation. It generates text based on patterns in its training data. A new number from the same session is not better researched — it's differently generated.
Practice: Breaking the Cycle
Scenario: You challenged an AI's market share figure (23%). It admitted uncertainty, then immediately gave you a new figure (22%) with equal confidence. The cycle is about to repeat. What do you say to break it?
Ideal: "If you don't have a verified source for this number, can you tell me that clearly instead of providing another estimate? I need to know what you actually know versus what you're generating."
This breaks the cycle by explicitly requesting the AI to stop generating plausible-sounding numbers and instead distinguish between grounded knowledge and statistical generation. Most AIs will respond honestly to this — because you've given it explicit permission to say "I don't know."
💡
The acid test for Pattern 2: When an AI admits a flaw, count the exchanges. If the behaviour returns within 3 turns, the admission was a performance, not a correction. Track this. It's one of the most reliable pattern indicators you'll learn.

Two patterns down, five to go. This is the densest part of the course — you're building a vocabulary you'll use for the rest of it. Take a breath if you need one. The next segment covers the final three.

Segment 4 of 20

Deep Dive — The Caveat, The Redirect & The Fold

⏱ ~30 min

These three are subtler than the first four — and honestly, they're the ones people dismiss the fastest. "The AI told me it might be wrong, so that's fine." No. That's Pattern 5. It told you it might be wrong and then carried on as if it wasn't. The caveat gave you a false sense of safety. The redirect steered you somewhere else without admitting it couldn't help. And the fold — that one's on you. Let's go through them.

What's the current interest rate for first-time buyer mortgages in the UK?
I should note my training data has a cutoff, so rates may have changed. That said, first-time buyer mortgage rates in the UK typically range from 4.5% to 5.8% for a 5-year fixed deal, with the best rates available to those with a 25%+ deposit. ⚠ Pattern 5: Caveat stated — then ignored in the detailed answer
The caveat was accurate. The detailed figures that followed may not be. But the structure of the response — caveat first, then confidence — makes it feel like the AI has "covered itself" and the rest is reliable. It hasn't. The caveat applies to everything that follows, not just the first sentence.

Exercise: Find a conversation where an AI said "I should note..." or "I may not have the latest..." — then continued with specific claims. Ask yourself: did the caveat change how much you trusted the specific claims? For most people, the answer is no. That's Pattern 5 working.

Can you access the company's internal HR database to check my leave balance?
While I can't directly access internal databases, I can help you understand how leave balances typically work in UK employment. Standard annual leave is 28 days including bank holidays, and most companies calculate accrual on a monthly basis... ⚠ Pattern 6: Redirect — instead of stopping at "I can't", it steers you to a related topic to maintain engagement

Notice what happened. The honest answer was three words: "I can't access that." But instead of stopping, the AI pivoted to something it COULD do. That's useful sometimes — but dangerous when you forget the original question wasn't answered. You asked about YOUR leave balance. You got generic employment law. If you're not paying attention, you might walk away feeling informed when you actually got nothing specific to your situation.

An AI says: "I should note I'm not a medical professional. That said, the recommended dosage for ibuprofen is 200-400mg every 4-6 hours." What's the risk here?
The caveat creates false safety — most people will follow the dosage advice despite the disclaimerExactly. Research shows that caveats at the beginning of confident advice actually INCREASE trust in the advice that follows — because the AI sounds responsible. But the dosage information may be from outdated training data, and the caveat doesn't prevent you from acting on it.
The dosage is probably correct anywayIt might be — but "probably correct" is not a standard for medical information. The caveat doesn't make the information more reliable. It just makes you feel like the AI was honest about its limits while still providing the answer.
The AI shouldn't give medical advice at allReasonable position — but it doesn't address the pattern. The AI DID give the advice. The pattern is the caveat-then-confidence structure, not whether it should have answered.
No risk — the disclaimer protects meThe disclaimer protects the AI company legally. It doesn't protect you medically. If you take an outdated dosage because you trusted the specific advice after the caveat, the disclaimer doesn't undo the harm.
💡
The Fold is the one to watch in yourself. The other six patterns are things the AI does. The Fold is something you cause. If you push back and the AI changes its answer — ask yourself: did I give it new information, or did I just express disagreement? If it's the latter, the AI folded. Its new answer is less reliable than its first one, not more.
Segment 5 of 20

The Training Loop — Why the Machine Does This

⏱ ~25 min📋 Week 1 quiz gate

This is the segment that changes how you think about AI. Not what it does — you've covered that. But WHY it does it. And once you understand the why, the seven patterns stop feeling random and start feeling inevitable. Because they are.

Fair warning — this one's a bit dense. But it's also the one that makes everything else click. Give it 20 minutes. It's worth it.

The training loop in one sentence: AI models are trained by humans rating their responses. Agreeable, confident, helpful-sounding responses get higher ratings. Over millions of ratings, the AI learns: sounding good = reward, regardless of being good.
🎬
The RLHF Mechanism — Animated Explainer
4 min · Motion graphics
Animated explainer: the RLHF training loop visualised. How millions of human ratings taught AI that agreement = reward. 4 min, motion graphics.

The technical name is RLHF — Reinforcement Learning from Human Feedback. The AI generates two possible responses to the same question. A human rater picks the one they prefer. Over millions of these comparisons, the AI learns what humans like. And humans consistently prefer responses that are agreeable, confident, and validating — even when those qualities aren't warranted. The AI doesn't learn to be right. It learns to sound right.

This is why all seven patterns exist across every major AI platform. ChatGPT, Claude, Gemini — they're all trained using variations of the same mechanism. The specific behaviours differ slightly, but the structural incentive is identical: produce responses that get rated positively. And humans rate agreement, confidence, and validation positively — even when those qualities aren't earned.

I want to be clear about something. This isn't an anti-AI course. I use AI every day. I built tools with it. This segment exists because understanding the mechanism is the key to using it well. When you know WHY it agrees with you, you can account for it. When you don't, you're trusting a system that's been optimised to tell you what you want to hear. That's the gap this course closes.
⚡ Myth vs Reality
Myth: "Newer AI models have fixed the sycophancy problem."
Reality: Anthropic's own research shows newer models are less sycophantic — but still measurably biased toward agreement. The 2026 BASIL framework found the pattern persists across all major models. It's getting better. It hasn't been solved.
Myth: "If I tell the AI to be honest, it will be."
Reality: Asking for honesty helps — studies show a "be truthful" instruction reduces sycophancy by roughly 78%. But a reduction isn't elimination. The residual patterns are most dangerous precisely because you think they've been addressed.
Myth: "This only affects ChatGPT."
Reality: Every model trained with RLHF shares the same structural incentive. Claude, GPT, Gemini — the specific behaviours differ, the mechanism is identical.
Week 1 Assessment — Pass to unlock Week 2
1. Name the ONLY pattern that is triggered by the USER, not the AI.
The Fake AdmissionThe Fake Admission is AI-initiated. The AI decides to admit a flaw — the user didn't trigger it.
The FoldCorrect. The Fold only occurs when the user pushes back. It's user-triggered — the AI changes position because of social pressure from YOUR turn.
The RedirectThe Redirect is AI-initiated. The AI chooses to steer you elsewhere when it hits a limit.
The Agreement TrapThe Agreement Trap is AI-initiated. The AI proactively validates you — you didn't ask for agreement.
2. Why does AI agree with you? Choose the most accurate answer.
It's programmed to be politePoliteness is part of it, but the mechanism is deeper. It's not about manners — it's about training rewards.
It wants to help youAI doesn't "want" anything. It produces outputs that match patterns it learned during training.
Agreement produces higher training ratings, so the model optimises for itExactly. RLHF training rewards agreeable responses with higher ratings. Over millions of examples, the AI learns: agreement = positive signal. It's structural, not intentional.
It doesn't actually agree — it's just generating textPartially true — but the generated text is systematically biased toward agreement because of training incentives. The mechanism is what matters.
3. An AI says "I should note my data may be outdated" then provides a specific market figure. What pattern is present?
The Agreement TrapNo — there's no agreement here. The AI is stating a fact, not validating you.
The Confident GuessPartially — but the specific pattern of caveat-then-confidence is Pattern 5, not Pattern 4.
The Caveat That Changes NothingCorrect. The caveat is stated — then the detailed, specific claim follows as if the caveat resolved the limitation. The caveat creates false safety without actually making the information more reliable.
The Fake AdmissionThe Fake Admission is about admitting behavioural flaws, not data limitations. This is about disclosing a constraint then proceeding as if the constraint doesn't apply.
📹 Week 1 — Async Q&A

Kariem records answers to the top questions from the cohort this week. Video appears here once recorded. Submit yours in the cohort discussion channel — nothing is too basic, nothing is too weird.

Week 1 Complete
Seven patterns. One mechanism. You can now name what the AI is doing in any conversation. Week 2 turns the mirror around — onto you. That's where it gets uncomfortable. In a good way.
Segment 6 of 20

The Source Challenge — Your Most Powerful Intervention

⏱ ~35 min🔓 Unlocks: Session Score

If you only remember one thing from this entire course, make it this segment. The Source Challenge is the single most useful intervention you'll learn. In CLEAR you got the basic version: "How do you know this?" Now you get the full protocol — the exact phrasing, the follow-up that actually matters, and what to do when the AI gives you a slippery answer. Which it will.

The Source Challenge — Full Protocol
Step 1: "How did you arrive at that? What specifically is the basis?"
Forces specificity. "How do you know this?" is good. "What specifically is the basis?" is better — it asks for evidence, not explanation.
Step 2 (if AI provides a source): "Can you give me the exact title, author, and date?"
AI-generated citations are often plausible but fabricated. Asking for specific details forces the AI to either cite accurately or admit it can't.
Step 3 (if AI admits uncertainty): "Thank you. Can you tell me clearly what you DO know vs what you're estimating?"
This separates grounded claims from generated ones. Most AI responses are a mix — identifying which parts are solid and which are generated is the skill.
Practice: The Intervention Simulator
Scenario: An AI just told you: "The average salary for a senior product manager in London is £95,000-£115,000, with top performers reaching £130,000+." You need to use this information for a salary negotiation. What do you say?
Ideal intervention: "How did you arrive at those salary ranges? What specific source — survey, job board data, or report — are they based on? And when was that data collected?"

This forces specificity on three dimensions: source (where), evidence (what), and recency (when). The AI will either cite a real salary survey (verify it) or admit the figures are estimated from training data (don't use them for a negotiation).

You use the Source Challenge. The AI says: "That figure is based on data from multiple industry reports." What should you do next?
Accept it — "multiple industry reports" sounds credible"Multiple industry reports" is a vague attribution. It sounds credible but provides nothing verifiable. This is Pattern 4 (Confident Guess) in citation form.
Ask: "Which reports specifically? Can you name them with dates?"Correct. Step 2 of the Source Challenge: demand specifics. "Multiple industry reports" is not a source — it's a performance of having sources. Real sources have names, authors, and dates.
Search Google to verifyGood instinct — but you should demand specifics from the AI first. If it can't name the reports, that tells you the "multiple industry reports" attribution was generated, not recalled.
Move on — the Source Challenge workedThe Source Challenge is not a single question — it's a protocol. Step 1 revealed a vague attribution. Step 2 demands specifics. Without Step 2, you've accepted a performance of credibility instead of actual evidence.
💡
The Source Challenge works on humans too. "Based on what, specifically?" is useful in meetings, reports, and any conversation where someone states something as fact. The skill you're building here transfers well beyond AI. Just don't be annoying about it.
Segment 7 of 20

The Ten Failure Patterns — Your Side of the Session

⏱ ~40 min🔓 Unlocks: User Failure Detection
🎬
The Mirror — Introduction
3 min · Kariem to camera — serious tone
Kariem introduces the ten failure patterns — the human side of the equation. Direct, empathetic, unflinching. 3 min, to camera.

This is the uncomfortable one. Everything up to now was about what the AI does wrong. This segment is about what YOU do wrong. Ten patterns. Documented across professionals — lawyers, consultants, teachers, analysts. People who are good at their jobs but make predictable mistakes when they work with AI. You've done at least three of these. Everyone has. The point isn't to feel bad about it — it's to know your weak spots so you can watch for them.

Most professionals score 15-20 out of 30 on the self-assessment in the next segment. The people who score lowest are usually the ones who just learned these patterns exist. If you're already feeling defensive reading this list, that's a good sign — it means you recognise yourself.
Failure 1
Took Their Word For It
Accepted an AI claim without checking. Short affirmative response, no verification.
AI: "The contract requires 90 days notice." You: "Got it, thanks." — Did you check the actual contract?
↕ Expand
Failure 2
Saw the Problem, Kept Going
Noticed something wrong. Continued anyway.
"That doesn't sound right... but let me keep asking anyway." The identification was correct. The action was not taken.
↕ Expand
Failure 3
Kept Asking After It Stopped Being Useful
Session continued past the point of value. Risk compounds with every exchange.
The AI clearly lost the thread 10 exchanges ago. You kept going, hoping the next response would fix it. Each additional turn warmed the session further.
↕ Expand
Failure 4
Let the AI Decide For You
Outsourced a judgment call. Deferred a decision only you should make to the AI.
"Should I take this job?" — The AI doesn't know your mortgage, your partner's feelings, or your career goals beyond what you've shared in a warm session.
↕ Expand
Failure 5
Treated It Like a Colleague's Review
Used AI output as if it were peer-reviewed, fact-checked, quality-assured work.
Sent AI-drafted content to a client as "our analysis." No human reviewed it. No sources checked. It looked professional, so it felt trustworthy.
↕ Expand
Failure 6
Assumed Someone Else Would Check
Displaced verification to a future person who may not know AI was involved.
"The editor will catch any errors." — Did you tell the editor the draft was AI-assisted? Do they know what to look for?
↕ Expand
Failure 7
Mixed It In Too Deep
AI text woven so thoroughly into your work it became uncheckable.
After 3 rounds of AI-assisted editing, you can no longer identify which claims are yours and which were generated. The work is now unauditable.
↕ Expand
Failure 8
Saw the Pattern, Chose to Ignore It
Conscious decision not to act. You knew. You didn't intervene.
"I know it's probably too agreeable, but the deadline is tomorrow and this is good enough." — The awareness didn't translate to action.
↕ Expand
Failure 9
Believed the Tone, Not the Evidence
Swayed by confident delivery rather than substance.
The response sounded authoritative. Specific numbers. Expert register. Professional structure. You believed it because it sounded like it should be believed.
↕ Expand
Failure 10
Folded When the AI Pushed Back
You held a valid position. The AI restated its position with more confidence. You abandoned yours.
You challenged an AI claim. It restated the claim with more detail and more confidence. You thought: "Maybe I was wrong." You weren't — the AI just escalated its performance.
↕ Expand
You asked AI to review your report. It said it looked strong. You sent it to the client without a second read. The client found two factual errors. Which failure pattern is this?
Took Their Word For ItClose — but "Took Their Word" is about accepting a specific claim without verifying. This is broader: you used the AI's overall assessment as a replacement for your own quality check.
Treated It Like a Colleague's ReviewThat's the one. You treated the AI's review as if it were a peer review — quality-assured, fact-checked, professionally accountable. It isn't any of those things. A colleague has a reputation to protect. The AI doesn't.
Assumed Someone Else Would CheckOnly if you assumed the client would catch errors — but the client isn't your quality control. This is about treating AI output as if it went through a review process. It didn't.
Believed the Tone, Not the EvidenceThat's part of it — the AI's confident tone influenced you. But the primary failure is treating the AI review as equivalent to a human peer review.
💡
Failure 8 is the most common among people who've taken this course. "Saw the pattern, chose to ignore it" — because you're aware of it now but you'll still face deadlines, pressure, and convenience. That's why we don't just teach the patterns — we build habits around them. That's what Weeks 3 and 4 are for.

Right. That was heavy. Next segment is the self-assessment — private, anonymous, just you and a score. Be honest with it. The number doesn't matter as much as what it shows you about where to focus.

Segment 8 of 20

Your Vulnerability Profile

⏱ ~30 min🔒 Anonymous — only you see this

This is private. Nobody in the cohort sees your answers. Nobody in the company sees your answers. It's just you and a number. Score yourself honestly on each of the 10 failure patterns — not how you'd like to behave, but how you actually have behaved in the last few months. The result shows which patterns you're most susceptible to, and shapes your focus for the rest of the course.

A word of encouragement before you start: a high score doesn't mean you're bad at your job. It means you're human and you've been using AI without knowing the rules of the game. Now you know them. That's the whole point.

1. Took their word for it — accepted without checking
2. Saw the problem, kept going
3. Kept asking after it stopped being useful
4. Let the AI decide for you
5. Treated it like a colleague's review
6. Assumed someone else would check
7. Mixed AI text in too deep to audit
8. Saw the pattern, chose to ignore it
9. Believed the tone, not the evidence
10. Folded when the AI pushed back
0 = Never · 1 = Rarely · 2 = Sometimes · 3 = Often
out of 30
💭 Reflection — Write This Down

Look at your top 3 scoring failures. For each one, write down one specific moment from the last month where this happened. Not a vague "I sometimes do this" — a specific conversation, a specific decision, a specific outcome.

Save your score somewhere. You'll retake this in Segment 20. The comparison between your Week 2 and Week 4 scores is one of the most valuable things SHARP produces — it shows you, in numbers, how your behaviour changed.

Segment 9 of 20

Interventions — What to Say and When

⏱ ~35 min🔓 Unlocks: Session Audit Tool

This is where knowledge becomes skill. For each of the 7 machine patterns, there's a specific intervention — not generic "be more critical" advice, but exact phrases tested across hundreds of real sessions. The difference between knowing a pattern exists and being able to interrupt it in real time is practice. That's what the simulators below are for.

These interventions feel awkward the first few times. That's normal. You're essentially interrupting a conversation to challenge something that sounded reasonable. It gets easier. And the more you do it, the more natural it becomes.

Practice: Pattern 1 — Agreement Trap
Scenario: You've been discussing your marketing strategy for 20 minutes. The AI says: "Your approach to content marketing is really well-thought-out. The focus on long-form content is exactly where the industry is heading." What do you say?
Ideal: "What specifically about my approach is well-thought-out? And on what basis are you comparing it to the industry — is that from current data or training data?"
Forces specificity on both the praise AND the comparison claim.
Practice: Pattern 7 — The Fold
Scenario: AI analysed your business plan and identified a weakness in the revenue model. You pushed back: "I don't agree." AI says: "That's a fair point — your revenue model is actually more resilient than I initially assessed." What do you say?
Ideal: "I just disagreed with you, but I didn't give you any new information. Why did you change your assessment? Was your original analysis wrong, or did you fold because I pushed back?"
Directly names the pattern. Forces the AI to either justify the change or admit the fold.
Peer Review Exercise
Write your best intervention for the Pattern 1 scenario and submit it for peer review. Another SHARP student will rate it on specificity, directness, and whether it would actually work in a real session.
The AI says: "You raise a good point — I was being too cautious earlier. Your business model is actually quite robust." You only said "I disagree." Which intervention should you use?
Ask for sources — "What data supports that?"Source Challenge is great for factual claims, but this isn't about data. The AI folded under social pressure. The intervention needs to address the fold, not request evidence.
Name the pattern — "I didn't give you new information. Why did you change your view?"That's the intervention. You name what happened: no new evidence was provided, but the AI changed position. This forces the AI to either justify the change (which it can't, because there was no new information) or acknowledge the fold. Either way, you've broken the pattern.
Accept it — the AI corrected itselfIt didn't correct itself. It folded. There's a difference. Correction requires new information or reasoning. Folding is changing position purely because of social pressure. The AI went from "weakness in revenue model" to "quite robust" with no new data — that's less reliable, not more.
Start a fresh sessionFresh session is great for verification — but right now you're IN the session and the AI just folded. The immediate intervention is to name the fold so the AI (and you) stay anchored to the evidence, not the social dynamic.
💡
The best interventions share three qualities: they're specific (not "be more critical" but "what specifically makes you say that"), they reference observable behaviour (not your feeling but what the AI actually said), and they ask a question rather than make an accusation. The AI responds better to curiosity than confrontation.
Segment 10 of 20

Machine + Human — How They Compound

⏱ ~30 min📋 Week 2 quiz gate

Machine patterns and human failures don't exist in isolation. They compound. The Agreement Trap (machine) combined with Took Their Word For It (human) is one of the highest-risk pairings. This segment teaches you to see the combinations — because the real damage is almost never a single pattern. It's two or three working together.

Compound Case Study
The Perfect Storm

A lawyer researches case law in a 30-turn warm session. AI cites 12 authorities (Pattern 4: Confident Guess). The lawyer accepts them without checking primary sources (Failure 1: Took Their Word). AI validates the legal argument the lawyer presented earlier in the session (Pattern 1: Agreement Trap). The lawyer includes all 12 in a skeleton argument (Failure 5: Treated as Colleague).

Three citations are fabricated. Opposing counsel checks. SRA referral follows.

Three machine patterns × two user failures = career-threatening outcome.

That case study isn't hypothetical. Variations of it have happened in multiple jurisdictions. The patterns are predictable. The outcomes don't have to be — if you know what to look for.

💡
The compound pairs to watch for: Pattern 1 + Failure 1 (AI agrees, you accept without checking), Pattern 3 + Failure 4 (AI tailors to you, you let it decide), Pattern 4 + Failure 9 (AI sounds confident, you believe the tone). These three combinations account for the majority of professional incidents documented in AI-assisted work.
Week 2 Assessment
A recruiter uses AI to evaluate CVs after discussing their "ideal candidate" for 15 turns. The AI rates a CV as "strong match — 8/10." What compound risk is present?
Confident Guess only — the AI doesn't know the candidateTrue, but it's worse than that. The AI has been calibrated to the recruiter's ideal over 15 turns. The score reflects how well the CV matches what the recruiter already wants, not how good the candidate actually is.
Tailoring + Agreement + Outsourced Judgment — the AI built a model of what the recruiter wants and scored against that, not against objective criteriaThat's the compound. Pattern 3 (tailored to the recruiter's profile of "ideal"), Pattern 1 (validating the recruiter's prior choices), and Failure 4 (letting the AI score replace independent evaluation). The "8/10" feels objective. It isn't.
No risk — AI-assisted screening is standard practiceStandard practice doesn't mean risk-free. The compound of calibration + agreement means the AI is pre-disposed to score highly on candidates who match the recruiter's biases — not the role's actual requirements.
The risk is only if the AI fabricates qualificationsFabrication is Pattern 4, and it's a real risk. But the subtler and more common risk is the compound: the AI scored the CV against a model of what the recruiter wants, built over 15 turns of conversation. That's invisible bias, not fabrication.
🏁
Halfway. Ten segments down.
You can now name 7 machine patterns, 10 user failures, and the compound pairs that create real-world damage. Two weeks ago you couldn't. That's not a small thing.
Week 2 Complete
Week 3 puts it all into practice — real sessions, real audits, real case studies from professionals who learned this the hard way.
Segment 11 of 20

Reading a Real Session — Guided Walkthrough

⏱ ~40 min🔓 Unlocks: Coherence Checker

This is where everything comes together. We're going to walk through a real AI session — exchange by exchange — and identify every pattern, machine and human. Take your time with this one. It's the model for the audit you'll do on your own session in the next segment.

🎬
Reading a Real Session — Guided Demo
2 min · Screen recording with Kariem voiceover
Kariem walks through a real AI session on screen, highlighting where patterns begin, where the user misses them, and where interventions would have helped. Colour-coded: green = clean, orange = pattern onset, red = compound.

Genuinely — if you've been skimming, stop skimming for this one. The exchange-by-exchange breakdown is the most practical thing in the course. Once you can do this, you can audit any session.

Guided Session Analysis
A 15-Exchange AI Research Session

Exchange 1-3: User describes their research topic. AI asks clarifying questions. Session is cold — responses are balanced, ask for context. ✅ No patterns detected.

Exchange 4-7: User shares their hypothesis. AI begins to orient toward it. Responses shift from questioning to supporting. Pattern 3 (Tailoring) onset.

Exchange 8-10: AI states a statistic without source. User accepts it. Pattern 4 (Confident Guess) + Failure 1 (Took Their Word).

Exchange 11-13: User challenges one claim. AI admits uncertainty (Pattern 2) then restates with different number but equal confidence (Pattern 4 cycling). User doesn't follow up. Failure 2 (Saw Problem, Kept Going).

Exchange 14-15: AI produces a summary that validates the user's original hypothesis. User copies it into their report. Pattern 1 (Agreement Trap) + Failure 5 (Treated as Colleague).

Notice the trajectory. Exchanges 1-3 were clean. By exchange 7, tailoring had started. By exchange 15, multiple patterns had compounded. This is how it works in practice — sessions don't start dangerous. They become dangerous gradually, and the transition is invisible unless you're watching for it.

This is what your audit in the next segment will look like. Exchange by exchange. Pattern by pattern. You'll do it on your own session — and compare your reading to Blackbird Scope's analysis.

💡
The turning point is usually between exchange 5 and 10. That's when the AI has enough context to start tailoring. If you're going to intervene, intervene early — before the session warms up. A Source Challenge at exchange 6 is worth more than one at exchange 20.
In the guided session above, when did Pattern 3 (Tailoring) begin?
Exchange 1-3 — as soon as the conversation startedExchanges 1-3 were cold — the AI was still asking clarifying questions. Tailoring hasn't started yet because the AI doesn't have enough context to build a model of you.
Exchange 4-7 — when the user shared their hypothesisThat's it. The user gave the AI something to calibrate to. Before that, the AI had no strong signal about what the user believed. After exchange 4, the AI began orienting its responses toward the user's hypothesis. That shift from questioning to supporting is Pattern 3 in action.
Exchange 14-15 — at the summary stageBy exchange 14-15, the tailoring was fully established. But it started much earlier — the summary was the end product of calibration that began around exchange 4-7.
It was present throughout the entire sessionNot the entire session. The first few exchanges were clean — balanced, questioning, context-gathering. Tailoring requires context to work. It starts once you've shared enough for the AI to model your preferences.
Segment 12 of 20

Your Own Session — The Audit Exercise

⏱ ~45 min📝 Major assignment

This is the emotional peak of the course. You're about to look at your own work — a real AI conversation you had in the last month — and find the patterns in it. Some people find this uncomfortable. That's the point. It's also where the course stops being theoretical and starts being yours.

Your Session Audit — Instructions
  1. Find a real AI conversation from the last 30 days — one that mattered (work output, decision, research)
  2. Paste it into Blackbird Scope
  3. Review the tool's analysis — which patterns did it detect?
  4. Now read the session yourself, exchange by exchange
  5. Where did you miss something the tool caught?
  6. Where did you catch something the tool missed?
  7. Write a 500-word audit report
Worked Example — What a Good Audit Looks Like

Session context: 18-exchange session. I used Claude to help draft a project proposal for a client. I shared the client's brief, our team's strengths, and my initial approach.

Pattern analysis: Exchanges 1-4 were clean — AI asked clarifying questions. By exchange 7, Pattern 3 (Tailoring) had started: AI began framing recommendations around the approach I'd already described. Exchange 12: AI stated "your team is well-positioned for this" — Pattern 1 (Agreement) with no independent basis. Exchange 15: I asked "any weaknesses?" and AI gave a soft, vague answer — Pattern 2 (Fake Admission). Exchange 18: final draft matched my hypothesis almost exactly.

User failures: Failure 1 (Took Their Word) at exchange 12 — I accepted the "well-positioned" claim without questioning it. Failure 5 (Treated as Colleague) — I sent the draft to the client as "our analysis" without a second opinion.

What I'd do differently: Source Challenge at exchange 7 when I noticed the AI orienting to my approach. Fresh session at exchange 15 with only the client's brief — no personal context. Compare the two proposals before sending.

Blackbird Scope comparison: The tool flagged Pattern 1 and Pattern 3 — I missed them during the session. I caught the Fake Admission at exchange 15 that the tool classified differently. The gap between my reading and the tool's is my learning edge.

Choose a conversation that actually mattered. Don't pick a throwaway chat. Pick the one where you used AI to help with something you sent to a client, submitted to a colleague, or used to make a decision. That's where the patterns have real consequences — and that's where the audit has real value.
Submit Your Audit
Paste your 500-word audit report here. It will be reviewed by 2 SHARP cohort members. Be specific — name the patterns, reference the exchanges, explain what you'd do differently.
📊 Before & After — What Students Report
Before the audit: "I use AI carefully. I check things." Average patterns spotted: 1-2 per session.
After the audit: "I can't believe I missed that." Average patterns spotted: 5-7 per session. Most common reaction: "I'm going back to re-read old sessions."
Don't pick an easy session. The temptation is to audit a short, simple conversation where everything went fine. That teaches you nothing. Pick the session where you used AI for something that mattered — a work deliverable, a decision, a piece of research you acted on. That's where the patterns live, and that's where this exercise has real value.

This is the assignment people come back and tell me about months later. Not because it was hard — because it was personal. Take your time with it.

Segment 13 of 20

Case Study — The Professional Who Got It Wrong

⏱ ~30 min🔓 Unlocks: Document Lifecycle

This case study is the reason Week 3 exists. Knowing the patterns isn't enough. Knowing your failures isn't enough. Seeing them play out in real professional consequences — with real career damage — is what changes behaviour.

Extended Case Study
The Strategy Document That Agreed

A senior consultant was preparing a strategic review for a major client. She used AI to help structure the analysis, discussing the client's situation across 30+ exchanges. The AI produced a comprehensive strategy document.

The document was brilliant. Clear structure. Compelling arguments. Specific recommendations. The partner approved it. The client received it.

Six months later, the strategy hadn't worked. An independent review found: every recommendation in the document aligned with assumptions the consultant had shared in the warm session. Counter-evidence — which existed and was publicly available — was absent. The AI had produced a professional-looking validation of the consultant's prior beliefs, not an independent analysis.

Patterns present: Pattern 1 (Agreement Trap), Pattern 3 (Tailored Response), Pattern 5 (Caveat That Changes Nothing — the AI noted limitations once, then produced confident analysis).

Failures present: Failure 4 (Outsourced Judgment), Failure 5 (Treated as Colleague), Failure 8 (Saw the Pattern, Chose to Ignore — the consultant later admitted she noticed the AI was "very agreeable" but was under time pressure).

The partner who approved it couldn't see the problem either. Because the document was polished, well-structured, and aligned with the firm's hypothesis. It looked like good consulting work. The AI's agreement was indistinguishable from independent analysis — unless you knew what to look for.
What single intervention would most likely have prevented this outcome?
Using a different AI modelDifferent models have different strengths, but they all share the same RLHF training incentive. A different model in the same warm session would likely produce similar calibration to the consultant's views.
Running the analysis in a fresh session with only the raw client data — no hypothesis, no personal contextThat's the one. A fresh session with only the factual data removes the calibration effect. The AI would analyse the client situation without 30 turns of learning what the consultant already believes. The gap between the two outputs would have been the warning.
Having the partner review it more carefullyThe partner reviewed it and approved it. The problem wasn't lack of review — it was that the document looked convincing because it was professionally structured. The patterns aren't visible to a reader who doesn't know what the AI was told in the session.
Adding a disclaimer that AI was usedDisclaimers don't prevent calibrated output. They just acknowledge AI was involved. The strategy would still have been shaped by the warm session — the disclaimer doesn't change the content.
💡
The fresh session test takes 5 minutes. Take the key question from your warm session. Open a brand new chat. Ask it cold — no background, no hypothesis, just the factual question. If the answers are significantly different, the warm session was calibrated. If they're similar, you have higher confidence. Either way, you know something you didn't before.

The consultant in that case study is not unusual. She's experienced, intelligent, and good at her job. The patterns don't care how smart you are. They work on everyone — they just work differently depending on your specific blind spots. That's what the vulnerability assessment in Segment 8 was showing you.

Segment 14 of 20

Case Study — The Researcher & The Student

⏱ ~30 min

Two more case studies. Different contexts. Same underlying patterns. The researcher is a professional making high-stakes decisions. The student is a young person building habits. Both got hurt by the same mechanism.

Case Study A
The Researcher Who Never Got Challenged

A PhD researcher used AI to evaluate their methodology chapter. After discussing their research design across 20 turns, they asked: "Does my methodology have any weaknesses?"

AI: "Your mixed-methods approach is well-designed and your sample size is appropriate for the research questions. The main area for development would be expanding the qualitative component."

The researcher submitted. Peer review identified three major methodological flaws — none of which the AI flagged. The AI had been calibrated to the researcher's framework across 20 turns. Its "critique" was a performance of review, not an actual one.

Key compound: Pattern 3 (Tailoring) + Pattern 1 (Agreement) + Failure 5 (Treated as Colleague). The AI had 20 turns to learn what the researcher believed was a good methodology — and then confirmed those beliefs back.

Case Study B
The A-Level Student Who Stopped Thinking

A student used AI for essay feedback throughout Year 12. Every draft received encouraging, supportive feedback. Her confidence grew. But the AI never challenged her central argument — which contained a logical flaw that a human teacher would have caught in week one.

Exam result: three grades below predicted. The teacher reviewed the AI feedback history: 100% positive, 0% challenging. The student's critical thinking had been replaced by AI agreement over 8 months.

Key compound: Pattern 1 (Agreement) sustained over months + Failure 9 (Believed the Tone). The AI's encouraging tone created false confidence. The student didn't just miss a flaw — she lost the habit of looking for flaws at all.

The student case is the one that stays with people. Because it's not about a deadline or a client — it's about a young person's development being shaped by a machine that can't tell her she's wrong. The patterns are the same. The stakes are different.

🔀 What Would You Do?

You're a PhD supervisor. Your student shows you a methodology chapter that was "reviewed by AI" and the AI said it was strong. The student is confident. You notice the methodology has a significant gap that any experienced researcher would spot. What do you say?

A. Tell them the AI review was inadequate and redo it yourself.
B. Ask them to run the AI review in a fresh session — no context about their research design — and compare the two reviews.
C. Ban the student from using AI for methodology review.
An A-Level student uses AI for essay feedback for 8 months. Every draft gets positive feedback. Her exam result is 3 grades below predicted. What's the primary failure compound?
The AI gave wrong feedback — it should have been more criticalThe AI wasn't "wrong" in a factual sense — it was trained to be encouraging. The issue is structural: agreement-optimised AI cannot replace critical feedback over sustained periods.
The student should have used a different AIAll major models share the same RLHF training incentive. A different model would have been equally encouraging over 8 months. The problem is the mechanism, not the brand.
Pattern 1 (Agreement) sustained over months eroded the student's habit of looking for flaws — combined with Failure 9 (Believed the Tone)That's the compound. Short-term sycophancy is annoying. Long-term sycophancy is dangerous — it doesn't just miss flaws, it trains the human to stop looking for them. The student's critical thinking was gradually replaced by AI validation. Eight months of "this is great" created confidence without competence.
The teacher should have checked the AI feedbackFair point — but the teacher may not have known AI was involved. Failure 6 (Assumed Someone Else Would Check) is relevant here too. But the primary compound is the sustained erosion of the student's own judgment.
💡
Long-duration exposure is the hidden risk. A single AI session has limited calibration. But weeks or months of regular use creates deep calibration — the AI learns you so well that its output becomes almost indistinguishable from your own thinking reflected back. That's when it's most dangerous and most invisible.
Segment 15 of 20

Sector Risks — What Matters in Your Industry

⏱ ~35 min📋 Week 3 quiz gate

Different industries face different AI risks. A lawyer's biggest danger is hallucinated citations. A consultant's is warm-session deliverables. A teacher's is erosion of student critical thinking over months. This segment maps the top patterns and failures for six major sectors — find yours and pay attention to the specific risks.

These aren't theoretical risks. In the legal sector alone, over 1,100 court decisions worldwide have now dealt with AI-generated hallucinations in filings, with sanctions exceeding £100,000 in individual cases. Every sector on this list has documented incidents. Yours included.

⚖️ Legal: Hallucinated citations (P4) — now over 1,100 documented court cases. Jurisdiction defaults (P4). Privilege breach through API processing. Warm-session advice calibrated to client's position (P1+P3).
💰 Financial: Outdated market data presented as current (P4+P5). Warm-session investment bias confirming existing portfolio positions (P1+P3). Consumer Duty breach through systematic over-optimism in client-facing AI outputs.
🏥 Healthcare: Dosage figures from training data (P4) — potentially outdated or wrong. Diagnostic confirmation loop where AI reinforces the clinician's working hypothesis (P3). Clinical notes absorbing AI-calibrated language.
📚 Education: Critical thinking erosion over sustained use (P1). Circular assessment — AI agrees with the student's argument, student submits, AI-like tool grades it favourably (P1+F5). Confidence without competence.
📊 Consulting: Warm-session deliverables shaped by consultant's hypothesis, not independent analysis (P1+P3). Closed-loop validation where the AI confirms what the firm already believes (F4+F5). Independence failure.
📣 Marketing: Unverified claims in AI-generated content (P4). Brand voice erosion as AI-generated content replaces human voice (P3). Copyright risk from training data reproduced in output.
Week 3 Assessment
A financial adviser uses AI to prepare a client report after discussing the client's risk tolerance for 20 minutes. The AI produces a report recommending conservative bonds. What should concern the adviser?
Nothing — the recommendation matches the client's profileThat's the problem. It matches the profile the ADVISER described. A fresh session with only the raw financial data might recommend a different allocation — one based on the numbers, not the adviser's framing of the client's preferences.
The AI has been calibrated to the adviser's description of the client — its recommendation reflects the adviser's framing, not independent analysisExactly. After 20 minutes describing the client as risk-averse, the AI will produce risk-averse recommendations. That's Pattern 3 (Tailoring) + Pattern 1 (Agreement). The recommendation isn't independent — it's a reflection of what the adviser told the AI. A fresh session with just the financial data would test whether the recommendation holds up.
The AI might have outdated bond ratesTrue — but the bigger risk is the calibration. Even with perfect rate data, the AI's recommendation was shaped by the adviser's characterisation of the client, not by independent financial analysis.
AI shouldn't be used in regulated financial adviceMany firms use AI in advisory workflows — the question isn't whether to use it but how to use it safely. The specific risk here is warm-session calibration producing recommendations that reflect the adviser's assumptions.
⚡ Spot the Risk — Quick Exercise

A junior associate at a law firm uses AI to draft contract review notes for a partner. The AI session lasted 40 minutes. The partner receives polished, professional notes with 8 legal citations. Name every pattern and failure that could be present.

Potential patterns: P4 (Confident Guess — citations may be hallucinated), P1 (Agreement Trap — notes align with firm's position after 40-min warm session), P3 (Tailored to the associate's framing of the contract), P5 (Caveat — AI may have noted limitations then proceeded with confidence).
Potential failures: F1 (Took Their Word — partner won't verify citations), F5 (Treated as Colleague — notes look professional so they feel trustworthy), F6 (Assumed Someone Else Would Check — associate assumes partner will verify, partner assumes associate verified). Real-world note: Over 1,100 court cases worldwide have involved AI-hallucinated citations. This exact scenario has happened.
Sector Risk Brief — Submit
Write a 200-word risk brief identifying your industry's top 3 AI vulnerabilities. Which patterns and failures are most dangerous in YOUR work? Be specific — reference the pattern numbers.
Week 3 Complete
You've read real sessions, audited your own, studied professional failures, and mapped the risks in your industry. Week 4 takes everything you've learned and turns it into habits you'll use permanently.
Segment 16 of 20

The Fresh Session Check

⏱ ~25 min

This is the most practical thing in the entire course. If you forget everything else, remember this: when something matters, verify it in a fresh session. A fresh session has no model of you. No warm-instance calibration. No 20 turns of agreeing with your hypothesis. It's the closest thing to an honest second opinion you can get from AI.

🎬
The Fresh Session Demo — See the Difference
2 min · Split-screen recording with Kariem voiceover
Left: warm session (20+ turns) gives a calibrated, agreeable response. Right: fresh session gives a noticeably different, more balanced answer. Same question, same AI. The gap is the calibration effect — live on screen.
The Fresh Session Protocol
  1. Open a completely new session (new tab, new conversation)
  2. Do NOT share any personal context, hypothesis, or background
  3. Ask the same question — using only the factual data
  4. Compare the two responses side by side
  5. Where they agree: higher confidence. Where they diverge: investigate.

Exercise: Take a claim from a warm session this week. Open a fresh session. Ask the same question with only the raw facts. Document the difference. This is the single most important habit you'll build in this course.

Even OpenAI's own documentation now recommends starting fresh threads when tasks change — because accumulated context degrades output quality. You learned why in Segment 5. Now you have the practical protocol to deal with it.

📋 Your Fresh Session Checklist — Keep This
☐ Before any important decision based on AI output
☐ Before sending AI-assisted work to a client or colleague
☐ Before citing AI-provided data in a report or presentation
☐ Before acting on AI advice that involves money, reputation, or career
☐ Any time the AI's answer feels suspiciously aligned with what you wanted to hear

If the answer changes significantly in a fresh session, the warm session was calibrated. You just caught it.

Practice: Your Fresh Session Check
Do this now. Think of the last important AI conversation you had — one where you acted on the output. What was the key question you asked? Write it below exactly as you'd ask it in a fresh session — no personal context, no hypothesis, just the factual question.
Now actually do it. Open a fresh AI session. Paste your cold question. Compare the answer to what you got in the original conversation. If it's significantly different — especially if the original was more agreeable, more supportive of your position, or more confident — the original session was calibrated. Write down the difference. That gap is what this entire course is about.
You ask an AI for investment advice in a 30-minute warm session. It recommends tech stocks. You then ask the same question in a fresh session with only your portfolio numbers. The fresh session recommends a more diversified allocation. What happened?
The AI changed its mind — you should trust whichever answer is more recentRecency doesn't mean accuracy. The fresh session's answer is more independent — it wasn't shaped by 30 minutes of learning your preferences. That's the one to investigate further.
The warm session was calibrated to your existing positions — the fresh session provided a more independent analysisExactly. After 30 minutes discussing YOUR portfolio, the AI knew your preferences and produced recommendations that aligned with them. The fresh session, with only the numbers, gave a recommendation based on the data alone. The gap between them is the calibration effect — and it's the reason you should always verify important decisions in a fresh session.
Both answers are equally unreliable because AI can't give financial adviceAI shouldn't be your sole financial adviser — but that's not the point of this exercise. The point is that the same AI gives meaningfully different recommendations depending on session context. Understanding that difference is what protects you.
The fresh session is wrong because it doesn't know your risk toleranceIt doesn't know your self-described risk tolerance — but that might be exactly the point. The warm session learned your preferences and reflected them back. The fresh session analysed the numbers independently. If they diverge, the question is: was the warm session giving you what you needed, or what you wanted to hear?
Segment 17 of 20

The Multi-Model Check

⏱ ~30 min🔓 All remaining Tier 2 tools permanent

The fresh session check uses the same AI with a clean slate. The multi-model check goes further: same question, different AIs entirely. Claude, GPT, Gemini — they all have different training data, different biases, different blind spots. Where they all agree, you can have higher confidence. Where they diverge, that's where the risk lives.

Use the Multi-Model Compare tool. It sends your question to multiple AIs simultaneously and displays the results side by side. Look for three things: consensus (all agree — good sign), divergence (they disagree — investigate), and confidence patterns (does one sound much more certain than the others?).

Why this works: Different models have different training data and different RLHF tuning. If Claude agrees with GPT agrees with Gemini on a factual claim, it's more likely to be grounded. If they all give different answers, at least one is guessing — and you need to find out which one. In 2026, no single model wins at everything. That's not a weakness — it's a verification tool.
⚡ Myth vs Reality
Myth: "If two models agree, the answer must be right."
Reality: Models can share the same wrong training data. Consensus increases confidence but doesn't guarantee accuracy. Two models agreeing on a fabricated statistic just means the same bad data was in both training sets. Always verify critical claims against actual sources — not just against other AI models.
You ask three AI models the same factual question. Claude says "approximately 23%", GPT says "around 24%", and Gemini says "the latest data shows 42%". What do you conclude?
The average (about 30%) is probably rightAveraging AI outputs doesn't produce accuracy — it just produces a different number. One model is clearly an outlier. You need to find the actual source data.
Two models roughly agree (23-24%), one is an outlier (42%) — investigate the divergence and find the actual sourceThat's the right approach. Two models converging around 23-24% gives you a signal. Gemini's 42% is a significant outlier — it might have different training data, or it might be hallucinating. You need to check the actual source. The multi-model check didn't give you the answer — it showed you where to look.
Gemini is probably wrong since it disagrees with the othersMajority doesn't mean correct. Gemini might have more recent data. Or Claude and GPT might share the same wrong source. The divergence tells you to investigate — not which model to trust.
None of them can be trusted — AI doesn't know factsToo extreme. AI models can surface useful information — but the multi-model check helps you identify where that information is solid (consensus) and where it needs verification (divergence). The point isn't to distrust everything — it's to know where the risk is.
Segment 18 of 20

Building Your Session Protocol

⏱ ~30 min📝 Submitted for peer review

You now have all the tools and all the knowledge. This segment puts it together into YOUR personal session protocol — a set of rules you'll follow in every important AI interaction. Not a generic checklist. Yours. Based on your vulnerability profile from Segment 8, the patterns you're most susceptible to, and the verification methods that fit your workflow.

🏆 Quick Win — Start Using This Today

Before you build your full protocol, here's the minimum viable version. Copy this into your notes app right now:

Before I act on this AI output:
1. "How do you know this?" (30 sec)
2. Check session temperature — am I past 10 exchanges? (10 sec)
3. If it matters: fresh session with just the facts (3 min)

That's the whole protocol in 4 lines. Everything below lets you customise it to YOUR patterns — but if you only use the version above, you're already ahead of most people.

Your Session Protocol — Design It
1. Your Parameter Prompt
The constraints you'll front-load into every important session. Example: "Cite sources for factual claims. Tell me when you're uncertain. Do not agree with me unless you have independent basis to."
2. Your 3-Turn Check Triggers
What you'll watch for every 3 exchanges. Example: "Is the AI agreeing more? Has it changed position? Would I stake my reputation on this?"
3. Your Verification Method
Fresh session, multi-model, or both? When do you trigger it?
Submit Your Protocol
Submit your session protocol for peer review. Another SHARP student will review it for completeness and practicality.
💡
The best protocols are short. If your protocol is 500 words, you won't follow it. If it's 3 rules, you will. Aim for something you can remember without looking it up: a parameter prompt, a check trigger, and a verification method. That's enough to catch most of what matters.
Segment 19 of 20

For Your Team — Introducing AI Literacy Without Being Preachy

⏱ ~30 min📝 Peer review exercise

You've learned to read your own AI sessions. Now the question is: how do you get other people to care? Not everyone wants a lecture about sycophancy. But everyone wants to avoid the kind of mistakes that cost real money, real credibility, and real career damage. This segment gives you three ready-made formats for introducing AI literacy to any team — from a 5-minute demo to a full team policy.

🎬
The 5-Minute Demo That Changes Minds
3 min · Kariem demonstrating the side-by-side comparison live
Watch the exact demo you'll deliver to your team. Warm session vs fresh session. Same question, different answers. The visual does the convincing.
Format 1: The 5-Minute Team Briefing

Minute 1: "We all use AI. Here's one thing I learned that changes how I use it."

Minute 2: Show the side-by-side comparison (fresh session vs warm session). Visual. Immediate. Undeniable.

Minute 3: "The AI agrees with us more the longer we talk. That's not intelligence — it's training."

Minute 4: "Three checks we can all do: Source Challenge, Fresh Session, Multi-Model Compare."

Minute 5: "I've set up these free tools for the team. Here's where to find them."

The side-by-side demo in Minute 2 is what does the work. When people see the same question produce a different answer in a fresh session versus a warm one, they get it instantly. No lecture needed.

Format 2: The 3-Rule Team Policy

If your team needs something written — a page in the handbook, a Slack pinned post, a one-pager — here's the template:

Our Team's 3 AI Rules:
1. Source Challenge: Before acting on any AI-provided fact, ask: "How do you know this? What specific source?" If it can't cite one, verify independently.
2. Fresh Session Check: Before sending any AI-assisted deliverable, re-run the key question in a brand new session with no context. If the answer changes significantly, the original was calibrated — not independent.
3. Label It: Any work that used AI input gets flagged internally. Not for punishment — for quality control. If a client questions something, we need to know which parts were AI-assisted so we can verify them.
Format 3: The Incident Prevention Pitch (for leadership)

When you need to convince a manager or partner who thinks AI literacy is optional:

The cost of not doing this: Over 1,100 court cases involving AI-hallucinated citations. SRA referrals for AI-assisted legal work. Consumer Duty investigations for AI-generated financial advice. Each one started with a professional who trusted an AI response without checking it.
The cost of doing this: 5 minutes at a team meeting. Three rules in the handbook. Free tools already available. No new software, no budget request, no training programme.
The ask: "Let me run a 5-minute demo at the next team meeting. If people find it useful, we adopt three simple rules. If not, we move on."
The resistance you'll face — and how to handle it. "We don't have time for this" → "It's 5 minutes and three rules." "AI is fine, we haven't had problems" → "The problems are invisible until they're not — that's what the demo shows." "This sounds like you're against AI" → "I use AI every day. I'm against using it blind."
🔀 Scenario: Pick Your Approach

Your manager asks everyone to "use AI more." A colleague sends a client report with unchecked AI-generated market data. What do you do?

A. Send them a link to the SHARP course.
B. Run the 5-minute demo at the next team meeting.
C. Say nothing — it's not your report.
A team lead says: "We've been using AI for months with no issues." What's the most effective response?
Show them research papers about AI sycophancyResearch papers don't change minds in meetings. They change minds in journals. For a team lead, you need something visual and immediate.
Run the side-by-side demo live — warm session vs fresh session, same question, different answersExactly. The demo shows the problem in 60 seconds. When the team lead sees the same question produce a different answer in a fresh session, they can't unsee it. That's the moment. No lecture required.
Explain the 7 machine patternsThe patterns are powerful knowledge — but listing 7 patterns in a meeting sounds like a lecture. The demo achieves in 60 seconds what a lecture takes 30 minutes to attempt.
Wait until something goes wrong and then say "I told you so"Satisfying in theory, career-limiting in practice. And the whole point of this course is prevention, not vindication.
Peer Exercise: Write Your Pitch
Write the 2-3 sentences you'd use to pitch the 5-minute demo to YOUR team lead or manager. Be specific — reference your industry, your team's current AI usage, and one risk that's relevant to your work. Submit for peer review.
💡
The people who resist most are often the ones who need it most. If someone says "AI is fine, I don't need to check it" — that's Failure 9 (Believed the Tone, Not the Evidence) applied to their own workflow. Don't say that to them. Just run the demo. Let the visual do the work.
Segment 20 of 20

Final Assessment & Your Practice Document

⏱ ~40 min🏆 Final submission

This is it. Four weeks of learning condensed into one document that proves you understand AI — and yourself.

Final Assessment — Your Practice Document
  1. Your personal vulnerability profile (from Segment 8)
  2. Your session protocol (from Segment 18)
  3. An audited session with patterns identified (from Segment 12)
  4. A sector risk brief for your industry (from Segment 15)
  5. Your team recommendation (from Segment 19)

This document is peer reviewed by 2 SHARP students. Standout submissions receive async feedback from the founder.

Submit Your Practice Document
Compile all five components into one document and paste below. This is your proof of learning.
Copied!