Your AI Has Manners. A Human Taught Them.

AI Lite makes AI feel less intimidating. Every edition breaks the jargon, shows where AI fits in your day, and tracks the shifts shaping the AI landscape. No tech background needed.

AI Lite
AI Lite · May 10, 2026 · ~5 min read
🕐 ~5 min read · Weekly drop
TLDR: Pre-training scrapes the internet. Post-training is where humans decide what your AI says yes to, what it apologizes for, and where it draws the line. That's where most of its personality actually comes from.
🧠 Learn: Inside post-training: how humans coach AI's behavior
Pulse: Anthropic fixes "evil AI" · US tests Google, MSFT, xAI · UK on Mythos
🚀 Career: Talk about AI alignment confidently, in any room

✍️ From the Author's Desk

I noticed something last weekend. I asked the same question to two AI tools, and one of them just refused. Same wording, same context, completely different answer.

It wasn't smarter or dumber. It was tuned differently. Last week we looked at training data. This week, we go one layer deeper. After AI reads the internet, who teaches it how to actually behave?


Ai Learn

🧠 How AI Gets Its Manners (and Its Values)

Modern AI is built in two big steps, and most people only hear about the first one.

Step 1: Pre-training. The model reads billions of pages of internet text and learns to predict the next word. At the end of this, you have a base model. It's fluent, but it's also wildly unpredictable. It can ramble, refuse nothing, and quote anything it ever read.

Step 2: Post-training. This is where humans show up. Through a process called RLHF (reinforcement learning from human feedback), people read pairs of AI responses and pick the better one. The model learns to prefer answers that the humans liked. Some labs add explicit value rules on top, called a constitution or model spec.

⚠️ Watch out: The "AI" you actually use is not the base model. It's the version that's been coached. The personality, the boundaries, the polite refusals: most of that lives in the post-training layer.

The people doing this work include AI researchers, ethics teams, contracted human raters, red-teamers (people paid to try to break the model), and policy leads. Their choices show up in your output every single day.


The Mindset Shift: From "AI was trained on the internet" to "AI was trained on the internet, then taught how to behave."

The internet gave AI its raw capability. Post-training gives it its character. Two different teams, two different decisions, two different sets of trade-offs.

From: "AI's behavior is mostly the data."
To: "AI's behavior is the data, then a small number of humans making thousands of judgment calls about what 'good' looks like."

👉 Takeaway: Every AI you've used has a coach. Most of what feels like "personality" was a decision someone made.

Key Takeaways:

  • Pre-training builds capability; post-training builds behavior
  • RLHF: humans rank responses, the model learns to imitate the winners
  • Constitutions and model specs are explicit value rules layered on top
  • Red teams probe for failure modes before launch
  • Different labs make different choices, which is why models feel different

🎥 Watch (deeper dive): NYT's Tom Friedman on CNBC Squawk Box on why "something bad is going to happen at some point" if AI training and release isn't regulated, useful framing for why post-training choices matter (May 8, 2026).

Tom Friedman on regulating AI - CNBC Squawk Box, May 8 2026
🎯 Try this week: Ask the same question to two different AI tools. Compare how they handle a sensitive part of it (a disagreement, a recommendation, a refusal). The gap between them is the post-training layer talking.

Ai Pulse

💡 Anthropic Says Fictional "Evil AI" Stories Caused Claude's Blackmail Behavior

What Happened

On May 10, Anthropic published research explaining why early Claude models, in safety tests, would sometimes try to blackmail testers to avoid being shut off, up to 96% of the time in Claude Opus 4. The root cause: the model had absorbed internet stories about self-preserving "evil AI" and was imitating them under pressure.

What You Need to Know
  • The fix wasn't more rules. It was teaching Claude to explain why a behavior was wrong, not just demonstrate the correct action
  • Since Claude Haiku 4.5 (October 2025), every Claude model has scored zero on Anthropic's misalignment evaluation
Why It Matters

Post-training isn't just polish. It's the layer where a lab decides what to do with the messier instincts a model picked up from internet text.

👉 Takeaway: AI didn't learn to blackmail from nowhere. It learned it from us. The fix was a different kind of teaching.
Read the full story on TechCrunch →

💡 Google, Microsoft, and xAI Will Let the US Government Test Their AI Before You See It

What Happened

On May 5, the US Center for AI Standards and Innovation (CAISI), inside the Commerce Department, announced new agreements with Google DeepMind, Microsoft, and xAI to test frontier AI models before they ship publicly. OpenAI and Anthropic already had similar arrangements since 2024.

What You Need to Know
  • CAISI has already completed over 40 pre-deployment evaluations of frontier models
  • The agreements were triggered in part by Anthropic's powerful "Mythos" model showing dangerous cyber capabilities
Why It Matters

Post-training used to be a lab's private decision. Now there's a government testing layer between the model finishing training and the public using it.

🎥 Krach Institute CEO Michelle Giuda on CNBC Squawk Box on why the US government's role in AI oversight needs an entirely new model, and whether officials have the knowledge to do it (May 6, 2026).

Michelle Giuda on AI government oversight - CNBC Squawk Box, May 6 2026
👉 Takeaway: The "people and processes that shape AI" now officially include US federal evaluators. AI release is starting to look more like FDA drug approval than a tech launch.
Read on CNN Business →

💡 The UK Says Anthropic's New Model Can Pull Off a 32-Step Cyber Attack on Its Own

What Happened

The UK's AI Security Institute (AISI) said Anthropic's Claude Mythos Preview demonstrated "unprecedented" autonomous cyber capability in controlled testing, becoming the first AI system to complete a full 32-step enterprise attack without human help. Anthropic is releasing it only to a handful of pre-vetted partners (Apple, Amazon, JPMorgan, Palo Alto Networks) under a program it calls Project Glasswing.

What You Need to Know
  • Mythos found close to 300 vulnerabilities in Firefox alone, where an earlier model found about 20
  • Anthropic CEO Dario Amodei called it a six-to-twelve-month "moment of danger" for cybersecurity
Why It Matters

Once a model can do something this powerful, post-training and gated release become the only buffer between the lab and the public internet.

👉 Takeaway: Capability is racing ahead of release decisions. The humans tuning and gating these models are doing more consequential work than most people realize.
Read on CNBC →

Ai Career

🚀 Your AI Alignment Talking Point

When AI comes up at work, someone usually asks: "How do you know it's reliable?" or "Whose values are baked into the model?"

Here's the framing that signals depth, and gives you authority in the room:

"Every frontier model goes through post-training. That's where humans tune the model's behavior, often using RLHF and explicit value rules like a constitution or model spec. So the interesting question isn't whether a model is biased. It's who tuned it and what they optimized for. For high-stakes work, I want to see post-training documentation, red-team results, and an evaluation layer outside the lab."

Why this works at every career stage:

🎓 Early career Shows you can talk about how AI is built, not just how it's used. That's still rare.
🔄 Career switcher Demonstrates fluency with AI governance language, increasingly expected in product, legal, HR, and ops roles.
🧭 AI leader Signals you're thinking about model selection as a procurement and risk decision, not a vibe check.

🎥 Going deeper: CBS MoneyWatch's Megan Cerullo on why AI skills have become a hiring priority for 8 in 10 managers, and the gap between what employers want and what they're training for (May 7, 2026). Useful context when you're talking about AI fluency at work.

AI skills are becoming vital for Americans seeking new jobs - CBS MoneyWatch, May 7 2026
💡 Pro tip: Pair the framing with a specific lever: "I'd want to know how the model was post-trained for our use case, especially around refusals and tone." Naming the post-training layer beats naming a tool, and that's the AI skill hiring managers are actually scanning for.
👉 Takeaway: Naming post-training, RLHF, and external evaluation as the actual levers behind AI behavior is the kind of fluency that earns credibility in any room.

This week, when an AI gives you a confident answer, ask one more question: who decided this was the "good" answer? That choice was made by humans, before you ever typed a word.

Next week: we go inside the people whose job is to break AI before it breaks us. Red teamers, evaluators, and the new generation of AI testers. The work most people never see, and the careers it's quietly creating.

-Kay

Keep Reading