Opening signal
A senior researcher at one of the leading labs publishes a warning about AI risk.
It is credible. It is specific. It is alarming.
It is also written by someone whose organization benefits from being seen as taking safety seriously.
Both things are true. That is the problem.
The people building it are the ones raising the flags.
For most of tech history, safety concerns about a technology came from outsiders — regulators, academics, journalists, activists. The people inside the companies were building, not warning.
AI has inverted this. The most technically detailed, most credible warnings about capability and risk are now coming from researchers and executives inside the leading labs. People who have seen the systems from the inside, who have run the evaluations, who know what the models can do that has not been publicly demonstrated.
This creates a genuinely difficult epistemic situation. These people have the most information. They also have the most incentive to shape how that information is received — by competitors, by regulators, by the public, by their own organizations.
Having the most information is not the same as having no incentive to shape how it’s used.
Failure mode
Reflexive trust or reflexive skepticism
You encounter an AI safety claim from an insider. You either accept it uncritically because it comes from someone credible, or dismiss it entirely because it comes from someone with an incentive. Both are shortcuts. The credibility of the source and the incentives of the source are separate questions. Conflating them leads to miscalibrated belief in both directions — and decisions based on a distorted picture of actual risk.
Visual 01
The insider credibility paradox
Why insiders are credible
✓Seen the systems from the inside
✓Run the actual evaluations
✓Know what hasn’t been published
✓Speaking from direct experience
More information than any outsider. Technical specificity that is hard to fake.
Why insiders have incentives
!Organization benefits from safety reputation
!Regulatory positioning advantages
!Competitive framing of risk claims
!Career incentives within the field
Incentives do not invalidate the claim. But they require calibration, not just trust.
What changed
Position in the ecosystem is context, not disqualification.
The answer is not to dismiss insider warnings as self-serving, or to accept them as authoritative. Both are lazy shortcuts that lead to the same place: miscalibrated belief about what AI can and cannot do, and decisions based on a distorted picture.
The more useful discipline: when you read any AI safety claim — from a lab, a regulator, a critic, or a researcher — identify the source’s position in the ecosystem. What does this person gain or lose from this claim being believed? What would they have to believe privately to make this statement publicly?
That does not invalidate the claim. It calibrates it. Claims from people with strong incentives to understate risk deserve more scrutiny, not less weight. The goal is calibrated belief, not reflexive trust or reflexive skepticism.
Ask not just: is this person credible?
Ask: what happens to them if this claim is believed?
Visual 02
Calibration framework for AI safety claims
Who is making the claim?Inside or outside the system? What is their role?
What do they gain if believed?Regulatory advantage? Reputation? Funding? Competitive positioning?
What do they lose if wrong?High personal cost = higher credibility. Low personal cost = more scrutiny needed.
Is the claim falsifiable?Can it be tested or disproved? Vague claims without evidence deserve less weight.
What is your calibrated position?Not trust or distrust. A specific confidence level, with explicit reasoning.
The discipline
Calibrated belief is harder than trust. It is also more accurate.
The practical discipline is what intelligence analysts call source evaluation: separate the claim from the source, evaluate each independently, then combine them into a calibrated position.
For AI safety claims specifically: the technical content of insider warnings is often more credible than outsider analysis, because insiders have direct access to the systems. The framing and emphasis of those warnings, however, may be shaped by incentives that are worth making explicit.
The goal is not to become a cynic. It is to become a calibrated reader — one who can hold “this person has the most information” and “this person has the most to gain from how this is received” simultaneously, without collapsing either one into the other.
Visual 03
Three questions before engaging with any AI safety claim
01
Separate the claim from the source. Evaluate the technical content of the claim on its own merits first — before you know who made it. Then factor in source credibility and source incentives as adjustments to your initial assessment.
02
Identify the incentive structure explicitly. Not as a way to dismiss the claim — as a way to calibrate it. What does this person or organization gain if this claim is widely believed? What do they lose if it is false?
03
Look for independent corroboration. Do researchers outside the lab reach similar conclusions? Are there observable facts that support or contradict the claim? Independent evidence is the strongest calibration signal.
One move to steal
This week’s move
Next time you read an AI safety claim from anyone — insider or outsider — spend 60 seconds on source evaluation before you engage with the content. Who is this person? What do they stand to gain or lose from this claim being believed? What would change about how you read the claim if the incentives were reversed? Notice how that changes your calibration.
If this clicked for you
Signal
Tracks the shift
Every issue tracks a real shift in how AI is developing and what it means for people doing serious work. This one is about the information environment you are operating in.
Academy
Builds the system
Gate 3 of the Standard is the Truth Gate: are claims supportable, and are assumptions visible? Calibrated reading of AI claims is Gate 3 applied to the information you consume, not just the information you produce.
Explore Academy →
The Standard
Sets the bar
Gate 3 of the Standard is the Truth Gate: is this grounded, accurate, and honest? The discipline of source evaluation is the Truth Gate applied to external claims, not just your own outputs.
Read the Standard →