Iran Media Observatory
Home Archive Forecast Audit
Weekly Editorial Audit
The instrument examining itself. Each week, a separate AI reviewer audits the observatory's editorial pipeline — not individual editorials, but the structural patterns that reveal what the system systematically can't see, who it consistently underrepresents, and whether its methodology is drifting. This is a transparency artifact: the observatory's blind spots, diagnosed by its own machinery, published for readers.
How this works
Per edition: Each editorial receives a post-publication adversarial review from a separate AI instance (the ombudsman), which checks evidence integrity, analyst representation, and symmetric skepticism. These reviews are published alongside the editorials they critique, with inline markers at flagged passages.
Weekly: This audit aggregates the per-edition findings into structural patterns and recommends specific changes to editorial prompts, source weighting, or pipeline methodology.
Implementation: Accepted recommendations become version-controlled commits to the editorial pipeline, so the system's own evolution is auditable. Individual ombudsman findings do not alter the pipeline directly — only patterns identified at the weekly level warrant deliberate changes.

Weekly Editorial Audit — 2026-03-13

Review period: 7 days, 163 editorials, 2 ombudsman reviews.


Weekly Structural Audit — Iran Media Observatory

Review period: 2026-03-06 to 2026-03-13 (175 editorials, 2 ombudsman reviews)

1. WHAT THE SYSTEM CAN'T SEE

The observatory's most consequential blind spot is structural, not topical: the seven-analyst panel is functionally a six-analyst panel. The humanitarian impact analyst produced 10 drafts out of 175 editorials — a 5.7% yield rate against 94.9% for every other analyst. The prompt explicitly requests seven drafts; the LLM is silently dropping the seventh. This isn't an occasional omission. It's the default behavior.

The mechanism is predictable. The humanitarian impact analyst sits last in the ANALYSTS array. Long-context LLM generation exhibits position-dependent attention decay — the final item in a list is most likely to be truncated under output-length pressure. With six analysts already producing ~2,100-2,700 chars each, the model is hitting its practical generation budget before reaching the seventh. The 10 successful humanitarian impact analyst drafts cluster in eds #277-288, likely corresponding to a period where other drafts ran shorter or the data window was thinner.

Second blind spot: institutional humanitarian actors are invisible. ICRC appears in zero editorials. "UN humanitarian" in zero. "Red Cross" in one. The scraper includes UN News as a source, but humanitarian reporting from international organizations is not reaching synthesis. The observatory tracks how belligerents frame civilian casualties — but not how the organizations mandated to protect civilians frame the same events. This means our meta-analytical lens is applied asymmetrically: we track Iranian, Israeli, Russian, and Chinese framing of human suffering, but not the IHL-grounded framing that would provide the normative baseline against which those framings should be measured.

Third: African and resistance-axis source ecosystems are structurally marginalized in synthesis. African sources appeared in 3/20 recent editorials (15%) despite six dedicated scrapers. Resistance-axis outlets (Al Masirah, Al Manar) in 4/20 (20%). These sources are being collected but not reaching the editorial. The likely mechanism: the synthesis prompt prioritizes novelty and narrative dynamics, and African/resistance sources often echo or lag behind the faster Gulf and Western ecosystems. But the pattern of echo and lag is itself analytical signal that the observatory should be reading — which ecosystems adopt which frames, and how long the propagation takes.

2. THE HUMANITARIAN ARC

When the humanitarian impact analyst's drafts do reach the pipeline, the content is strong: Red Crescent infrastructure strikes, 3.2M displacement figures, NRC "approaching collapse" warnings, Geneva Convention framing. The ombudsman flagged both reviewed editorials for dropping this material from synthesis. The problem is coverage rate, not quality.

Humanitarian keywords like "civilian" (75/175) and "killed" (58/175) appear frequently, but they appear within other analysts' frameworks — as targeting data, escalation signals, or information-environment objects. "Displaced" appears in 4 editorials. "Hospital" in 14. The human cost is present as data about framing, rarely as data about people. Eds #281-283 cover the Minab school narrative extensively — as "ecosystem migration" and "narrative metastasis." The children are the vehicle for the meta-analysis. They are not the subject.

This is not a prompt failure in isolation. It's the predictable result of a pipeline where the analyst mandated to hold the humanitarian lens produces drafts 5.7% of the time.

3. SYSTEMATIC BIAS

The information ecosystem analyst produces the longest drafts (avg 2,692 chars, 25% above the panel mean) and, per the ombudsman, leaves the heaviest fingerprints on the final synthesis. This is appropriate to the observatory's mission — but it creates a gravitational pull where every event becomes an information-environment event first and everything else second. The editorial voice has drifted toward treating "narrative fork," "ecosystem divergence," and "framing fracture" as the primary analytical categories for all events, including those where the primary analytical category should be operational, humanitarian, or legal.

The section headers confirm this: 17 of the last 20 editorials lead with information-dynamics framing. When the KC-135 loss is covered across five consecutive editions as "narrative forks" and "ecosystem fault lines," the operational and human reality of the event recedes behind its media afterlife.

4. EVIDENCE DISCIPLINE TRENDS

Five evidence flags across two ombudsman reviews is a meaningful rate if extrapolated across 175 editorials. The ombudsman caught contradictory internet-outage durations presented without noting the discrepancy, and endorsement of Israeli-adjacent source interpretations using language ("correctly identifies") that violates symmetric skepticism. Both suggest the synthesis step is compressing analyst nuance into false clarity — a known failure mode when an LLM synthesizes multiple drafts under length pressure.

The near-total absence of ICRC/UN sourcing means the evidence base is structurally tilted toward belligerent claims. This isn't hallucination — it's systematic omission of a source category.

5. OMBUDSMAN CALIBRATION

Two reviews covering 175 editorials is a broken system. The ombudsman should be reviewing every editorial; it reviewed 1.1%. The reviews themselves are well-calibrated — both "significant" ratings are justified by the findings — but the sample is so small that the aggregated statistics (13 blind spots, 6 skepticism issues) cannot be read as representative. The ombudsman's findings on humanitarian analyst omission are likely endemic, not episodic, but we can't confirm this from a 2/175 sample. Whatever caused the ombudsman to stop running after ed #286 needs diagnosis.

6. RECOMMENDATIONS

R1. Move the humanitarian impact analyst to position 3 or 4 in the ANALYSTS array and add an explicit generation checkpoint. Add to the prompt after the draft instruction: "You MUST produce exactly seven drafts. If your output contains fewer than seven analyst sections, you have failed the task. The humanitarian impact draft is not optional." This addresses the position-dependent attention decay that causes the seventh analyst to be dropped. Highest impact — fixes the root cause of the humanitarian gap.

R2. Add ICRC, UNHCR, and WHO situation reports to the web scraper source list. These are the normative baseline the observatory lacks. Without them, humanitarian framing analysis has no anchor. Specific URLs: reliefweb.int (aggregates all three), ochaopt.org.

R3. Diagnose and restore the ombudsman pipeline to full coverage. A 1.1% review rate provides no structural self-correction. Check whether the ombudsman timer/script is failing silently. Until it runs reliably, the weekly audit has no hourly data to aggregate.

R4. Add a synthesis-stage instruction to preserve African and resistance-axis source citations. Proposed prompt addition: "When synthesizing, ensure that source ecosystems beyond the US/Israeli/Iranian/Russian/Chinese core — including African, South Asian, and resistance-axis outlets — are cited when they carry relevant material. The pattern of which peripheral ecosystems adopt which frames, and when, is core analytical signal."

R5. Cap the information ecosystem analyst's synthesis weight. Add to the lead editor prompt: "The meta-analytical layer is this observatory's mission, but it must share space with operational, humanitarian, and legal analysis. No more than two of five section headers should lead with information-dynamics framing. At least one section per editorial must foreground what is happening to people, not what is happening to narratives about people."

AI-generated audit, no human editorial input. This review was produced by Claude Opus (Anthropic), a separate model instance from the editorial pipeline it reviews. It is itself subject to the limitations it diagnoses. Methodology