How threatened should human auditors feel about AI competition?
The Composite Fear Index aggregates six sub-indices measuring different aspects of AI threat to human auditor jobs:
Formula:
Composite = (Performance × 0.25) + (Penetration × 0.15) + (Severity × 0.20) + (Replacement × 0.20) + (Capability × 0.10) + (Trend × 0.10)
Component weights explained:
Weighting: AI-assisted (human-in-loop) tools receive 0.5× weight; backtesting results receive 0.2× weight (not real-time competition).
Score interpretation:
How does AI performance compare to human auditors in the same contests?
What it measures: AI's competitive standing when directly competing against human auditors in the same contests.
Formula: Percentile = (1 - place/total_participants) × 100
Weighting: AI-assisted (human-in-loop) tools receive 0.5× weight; backtesting results receive 0.2× weight (not real-time competition).
Data source: Human contest placements from bot_placements.csv where contestType = "human"
Interpretation:
Limitations: Based on contest results only; doesn't capture manual audit quality or client relationships.
Best performance: almanax placed #4/479 in Citrea
How much of the audit contest market has AI penetrated in the last 90 days?
What it measures: AI's market presence across different audit platforms and contest types in the recent 90-day window.
Formula: Penetration Score = min(100, (AI_entries/total_contests × 10) + (platforms_entered × 5))
Data sources:
bot_placements.csv - AI participation recordsreal_contests.csv - Total contest counts per platformKey metrics:
Interpretation: Higher penetration = AI is becoming ubiquitous in the audit contest market, reducing opportunities for human-only participation.
Platform Breakdown (90d)
All-time: 10 human contest entries, 891 bot races
Can AI find serious vulnerabilities, or just low-severity issues?
What it measures: The highest severity level of vulnerabilities that AI tools have successfully discovered.
Scoring:
Data source: verified_findings.csv - confirmed vulnerabilities found by AI tools in real audits/contests.
Why it matters: If AI can only find low-severity issues, human auditors remain essential for critical security work. Finding Critical/High bugs demonstrates AI can match human expertise on high-impact vulnerabilities.
Limitations: Severity classification varies by platform; some "High" findings may be less impactful than others.
Findings by Tool
| Tool | Critical | High | Medium | Low | Total |
|---|---|---|---|---|---|
| zerocool | 2 | 4 | 1 | 6 | 13 |
| nethermind-auditagent | 1 | 2 | 2 | 0 | 5 |
| v12 | 0 | 11 | 6 | 73 | 90 |
| agentlisa | 0 | 6 | 2 | 0 | 8 |
| octane | 0 | 2 | 0 | 1 | 3 |
| the-hound | 0 | 1 | 7 | 0 | 8 |
Factors that indicate whether AI could replace human auditors.
What it measures: Three key factors that determine whether AI could realistically replace human auditors.
Formula: Replacement Score = (Speed × 0.2) + (Cost × 0.3) + (Quality × 0.5)
Component breakdown:
(1 - AI_cost/Human_cost) × 100. Assumes ~$100 AI API costs vs ~$8,000 human auditor cost per contest (40 hrs × $200/hr).Interpretation:
AI auditor capability levels and associated threat to human jobs.
What it measures: A progressive scale of AI capability achievements, from basic participation to human-level expertise.
Formula: Score = (current_level / 7) × 100
Level definitions:
Detection: Automatically calculated from bot_placements.csv by checking placement positions in human vs bot-race contests.
Why it matters: Each level represents a qualitative leap in AI capability that historically required human expertise.
Current Level: 4 — Fear from milestones: 57%
Is AI getting better or worse at competing with humans?
What it measures: Whether AI performance is improving, declining, or stable over time.
Calculation:
change = recent_avg - historical_avgTrend classification:
Why it matters: A rising trend suggests AI will continue to close the gap with humans. A falling trend may indicate AI hitting capability limits or humans adapting.
Limitations: Small sample sizes may produce noisy trends; doesn't account for contest difficulty variations.
✓ AI performance declining — humans maintaining edge
How the overall Job Fear Index is calculated.
How each index has evolved over time (cumulative view). Tables show events that changed each index.
Performance Events (5 human contest entries)
| Date | Tool | Contest | Place | Percentile | Index Change |
|---|---|---|---|---|---|
| 2025-10-06 | almanax | Hybra Finance | #69/82 | 15.9% | 61.9 → 52.7 (-9.2) |
| 2025-09-15 | almanax | Succinct SP1 | #6/12 | 50% | 65.9 → 61.9 (-4.0) |
| 2025-08-15 | almanax | Citrea | #4/479 | 99.2% | 49.3 → 65.9 (+16.6) |
| 2025-07-23 | almanax | GTE Spot CLOB | #62/72 | 13.9% | 84.6 → 49.3 (-35.4) |
| 2025-06-01 | savantFirst Human Contest Entry (Autonomous AI) | Symbiotic Relay | #6/39 | 84.6% | 0.0 → 84.6 (+84.6) |
Capability Milestones (3 milestone events)
| Date | Milestone | Tool | Contest | Place | Level Change |
|---|---|---|---|---|---|
| 2025-08-15 | First Top 5 Finish (Autonomous AI) | almanax | Citrea | #4/479 | 43 → 57 (+14) |
| 2025-06-01 | First Top 10 Finish (Autonomous AI) | savant | Symbiotic Relay | #6/39 | 29 → 43 (+14) |
| 2023-04-27 | First Bot Race Win | tragedyotcommons | EigenLayer | #1/13 | 0 → 14 (+14) |
Platform Penetration Events (3 new platform entries)
| Date | Event | Tool | Contest | Score Change |
|---|---|---|---|---|
| 2025-08-15 | First Cantina Entry | almanax | Citrea | 20 → 30 (+10) |
| 2025-06-01 | First Sherlock Entry | savant | Symbiotic Relay | 10 → 20 (+10) |
| 2023-04-27 | First Code4rena Entry | tragedyotcommons | EigenLayer | 0 → 10 (+10) |