Track alpine results from last decade: only twice did a woman defend her world-slalom crown. Both times Mikaela Shiffrin did it, first in 2017, then again in 2026. Bookmakers priced the repeat at 9-1; the Elo-based model supplied to the Olympic committee gave 11 %. A small group of coaches who had watched the American train on injected snow the week before put the chance at 35 %. She won by 0.92 s. The story is recapped here: https://salonsustainability.club/articles/shiffrin-wins-slalom-gold-repeat.html.

Gold-medal repeats are textbook thin-tail problems: 30-40 racers, single-run variance above 2.5 s, temperature swings that shift base-layer friction by 0.06 µm. Feed those numbers into gradient-boosted trees and the output collapses toward the mean; the prior is too weak and the signal too sparse. Seasoned technicians bypass the math: they clock side-slip speed on the last inspection run, measure the rasp of the file, listen for the hollow tick when a ski hits a patch of over-watered ice. Those micro-readings never reach the database, yet they move the probability needle from 9 % to 30 % in minutes.

Inside Google’s 2025 internal audit, code flagged 0.3 % of ads as potential policy breaches. Human reviewers caught 61 % of the true violations among the same pool, algorithm recall stalled at 27 %. The difference: reviewers noticed a thumbnail font that mimicked a government seal-an edge case absent from training data. Translate that to supply-chain risk: TSMC’s 2020 chip drought was priced at 4 % by commodity models; plant managers who smelled acrylate shortages in Q3 booked capacity elsewhere and saved $210 m.

Actionable takeaway: keep the console open, but staff a two-person red-cell desk for any outcome that occurs less than 2 % of the time. Give them veto power over the model when three trigger questions-data sparsity, regime shift, or opponent adaptability-flash red. History shows the payroll of that desk pays for itself after the first correct override.

Expert Intuition Outperforms Algorithms in Forecasting Rare Events

Expert Intuition Outperforms Algorithms in Forecasting Rare Events

Build a 12-person red-team for any model that predicts <5% frequency outcomes: give each member 30 minutes with raw field notes, no dashboards, then run a silent vote. When seven or more disagree with the code, override and log the divergence; this cut miss-rates on West-African Ebola flare-ups from 18% to 4% between 2019-2025.

Seasoned petroleum geologists spotting trap geometries outperform gradient-boosting rigs by 22% F1 on 47 wells/year with <2% seismic coverage. Their trick: a 3-minute mental simulation of migration pathways done while circling the printed section with a red pen, a step no kernel was trained to replicate because label scarcity tops 1:50 000 km².

Give the human 64 milliseconds per anomaly: that is the mean fixation time radiologists need to flag 1-mm pneumothoraces on 2048-px scans. A ResNet-152 needs 3.7 GB of CUDA memory; the retina uses 6 ng of ATP and still catches five extra cases per 10 000 ICU admissions.

Keep a decision diary: after each override, record the gut cue (smell of sour mash in a Kentucky warehouse predicted three barrels seeping through staves; model missed it). After 200 logged calls, 34 repeat cues explained 81% of subsequent hits, letting junior blenders reach senior accuracy in 14 weeks instead of 9 years.

Combine both sides: feed the geologist’s freehand sketch into a 256-node GAN, then re-score the prospect. The hybrid curve lifts NPV by 11% across 83 Gulf-of-Mexico leases, proving gray matter plus silicon beats either solo.

Stop when disagreement collapses: once panel variance drops below σ=0.15 on a 5-point Likert, the edge is gone; from that point the machine’s 0.8-second answer is cheaper and just as reliable. Archive the session, wipe the whiteboard, move to the next tail-risk.

Map the Signal Landscape: Build a Checklist of 7 Micro-Cues Veterans Scan but Models Miss

Start each shift by running the 7-point sweep: (1) A freight trader notes a 0.3 % widening in bid-ask on Baltic C5 route at 06:47 GMT-before the move hits the tape. (2) Count the number of LinkedIn job posts containing hydrocracker in Houston ZIPs 77002-77007; a jump from 3 to 11 in ten days preceded the 2019 Pemex FCC outage by 18 days. (3) Check the queue depth on the Shanghai rubber warehouse app; if it drops below 12 trucks for three consecutive mornings, physical stocks are 8-11 kt lighter than exchange data claim. (4) Parse the Telegram channel of Pengerang refinery operators; three consecutive night-shift BBQ emoji strings tracked a 42-hour unplanned shutdown in 2021. (5) Monitor the spot price of 99.5 % isobutylene in Zhangjiagang; a 1.2 % daily rise while futures stay flat flags a cargo cancellation within 96 hours. (6) Read the customs metadata: when the declared weight of an MR tanker differs by >0.7 % from the AIS draft calculation, 83 % of cases involve later cargo switches. (7) Watch the Twitter follower count of the Ningbo port authority; a sudden 4 % drop correlates with cyber intrusions that delayed berths by 22 hours on average.

  • Keep a Trello card for each cue; set colour labels to red when three or more flash inside 48 hours-this stack gives a 17-hour lead on 92 % of the past 14 freight spikes.
  • Store screenshots with timestamps; regulators accepted these JPEGs as supporting evidence during the 2025 Singapore bunkers inquiry.
  • Automate the scrape, but eyeball the nuance: models ignore emoji order, yet swapping the burger icon before the beer predicts a day shift, after predicts a night call-out.
  • Share the checklist with the intern who covers weekends; human pattern sense catches what the Python script drops in the NaN bin.

Calibrate Confidence: Run a 15-Minute Red-Team Drill to Surface Overlooked Tail-Risk Indicators

Set a 15-minute timer, hand three colleagues a single-page scenario describing a 4-sigma shock (e.g., a 60 % overnight spike in the cost of a critical input), and ask each to list five weak signals they would watch in the next 48 hours. Collect the lists, merge duplicates, and rank by observability: signals that require public data sources (port throughput, satellite heat maps, customs declarations) score 1; those needing privileged access score 3. Anything averaging below 2 and appearing on at least two lists goes onto a 24-hour watch dashboard with SMS triggers.

Repeat monthly, rotating the shock (cyber freeze of a key logistics node, sudden export ban, 3-standard-deviation currency swing). Track which flagged indicators moved first in reality; promote the top 20 % to standing alerts and drop the laggards. After six cycles the hit-rate climbs from 27 % to 58 % while false positives drop by a third, trimming expected shortfall on a model portfolio of commodity positions from 2.1 % to 0.8 % of VaR.

Keep the drill under 15 minutes: cap prep at two slides, forbid laptops-paper and pens only. Record the session on a phone, transcribe with free voice-to-text, and paste the raw notes into a shared cloud sheet; no polished minutes. The scrappier the process, the faster weak signals surface before markets reprice.

Capture Tacit Knowledge: Turn Post-Mortem Stories into a 3-Step Script for Junior Forecasters

Capture Tacit Knowledge: Turn Post-Mortem Stories into a 3-Step Script for Junior Forecasters

Record every missed call within 24 h: who saw the signal, what noise drowned it, which gatekeeper blocked escalation. Tag each story with date, sector, asset, volatility level; store in a 5-field JSON so novices can grep copper + Q4-22 + warehouse-stock-discrepancy and surface three near-identical misses in 30 s.

Next, distill the narrative into a 3-step script: (1) Premonition-quote the exact phrase that raised neck-hair (spread blew past 2σ while headlines still read calm); (2) Friction-list the three veto voices and the KPI they used to kill position; (3) Pay-off-state P/L impact in bps and the calendar day when stop would have triggered. Archive 50 such scripts; rookies shadow-simulate them on live data until their recall rate on tail cues tops 80 %.

Blend Hybrid Teams: Assign Roles So Humans Trigger on 1%-Probability Thresholds and Models on 5%

Split the alert pipeline: analysts own 0.0-1.0 % scores, gradient-boosted trees own 1.0-5.0 %. Slack message routes to #human-review when the Bayesian posterior drops below 0.01; anything above 0.05 auto-opens Jira ticket tagged #model-handled. Last year a Nordic energy trader used this split: 14 human-flagged spikes below 1 % saved 38 MWh during a February storm; 312 model-flagged 2-5 % cases avoided 1.2 MWh curtailment penalties.

Threshold BandOwnerReview SLA2026 Hit RateFalse Save Cost
0.0-0.5 %Senior desk analyst5 min73 %$1.4 k
0.5-1.0 %Rotating junior15 min61 %$2.7 k
1.0-5.0 %XGBoost ensemble30 s78 %$0.9 k
>5 %Automated hedge0 s81 %$0.3 k

Calibration drift guard: every 48 h compare rolling 30-day empirical frequency against predicted. If the 0.5 % bucket shows 1.8 % frequency, raise the model cut-off to 0.6 % and shrink the human band to 0.0-0.6 %. Repeat until mismatch <0.3 %. A Frankfurt prop desk cut 112 false positives per quarter using this loop.

Keep a 0.25 % override reserve: any analyst can type /force-human in Slack to pull a 4 % forecast back into human queue. Reserve resets daily; unused overrides convert to $50 coffee credit. Morale stays high; 3-sigma outliers still get eyeballs.

FAQ:

How did the study separate expert intuition from a lucky guess when the events being predicted are rare?

The researchers ran two parallel tracks. First, they logged every forecast an expert made, scoring each on a 0-1 probability scale. Second, they built a matched sample of algorithmic forecasts that used only the same public data the expert saw. Over 1 800 rare-event cases (base rate < 2 %), the experts’ Brier scores were 0.17 lower on average. To rule out luck, they bootstrapped 10 000 re-samples of the data; the expert advantage persisted in 94 % of the draws. Finally, they asked each expert to narrate the trigger that made them raise the probability. Those narratives were later shown to independent raters; when at least three raters agreed the story contained a coherent cue (I have seen this exact freight-forwarding glitch before the last two container busts), the hit rate jumped to 38 %, while the algorithm stayed flat at 11 %. That pattern repeats too often to be coincidence.

Which industries supplied the data, and can the findings be transferred to medicine or cyber-security?

Half the observations came from global shipping delays, 30 % from sovereign-credit defaults, and the rest from commercial-satellite launch failures. All three settings share three traits: low base rates, high-dimensional covariates, and fragmented private information. A follow-up pilot used sepsis alerts in two hospitals (0.3 % admission rate). Clinicians who had seen > 500 similar ICU cases beat the hospital’s gradient-boosting model by 21 % in F1 score. A smaller cyber-security trial on zero-day exploits (42 cases) showed the same direction but missed significance; the authors suspect the expert panel was too junior. So transfer seems possible wherever veterans carry tacit pattern libraries that never reached the training set.

What stops companies from simply asking seniors for a gut feeling instead of investing in data infrastructure?

Three practical brakes keep firms from ditching models. First, senior experts are scarce; the study used only 38 people across five continents, and most were already near retirement. Second, intuition drifts: when two experts switched employers and lost daily exposure, their edge eroded within nine months. Third, regulators still ask for auditable risk numbers; my gut says 1.4 % fails the model-documentation checklist. The balanced set-up that emerged in the shipping firm was a hybrid: algorithm screens the 50 000 daily shipments, experts review the 400 highest-risk flagged records, overriding the code in 6 % of cases and cutting false alarms by a third. That mix saved an estimated USD 14 million in demurrage fees last year.

Could the algorithm be improved by feeding it the experts’ retrospective stories?

They tried. Narratives were parsed into 1 300 cue phrases and appended to the feature matrix. A regularised logistic model improved by 3 % AUC, but still trailed the live expert by 12 %. The gap narrowed only when the same expert who supplied the story also labelled fresh cases—evidence that part of the knowledge is context-bound and evaporates once removed from the source. The team now experiments with active-learning loops: the model proposes edge shipments, the expert corrects five per day, and the updated parameters are pushed overnight. After four months the difference shrank to 5 %, suggesting that iterative annotation, not one-off story dumps, is the viable path.

How should a risk manager decide when to trust the expert over the model?

Use a three-filter rule. (1) Event frequency: below 1 % per year, give the expert the casting vote; above 5 %, stick to the model. (2) Expert calibration: keep a rolling log of each forecaster’s probability scores; if their 20 % calls materialise around 20 % of the time over at least 30 observations, they retain veto power. (3) Information edge: require the expert to specify one non-public cue they inspected (a WhatsApp photo from a dock, a board leak, a customs broker’s mood). If the cue can be documented and later verified, the override is approved; if not, the model stands. Applying this rule to the shipping data kept 92 % of expert overrides while eliminating half of the harmful false positives.