Methodology
How we analyze and measure AI's impact on occupations. Our methodology is grounded in peer-reviewed research, transparent data processing, and clearly defined metrics.
Data Sources
Our analysis integrates multiple authoritative data sources to provide a comprehensive view of AI's impact on occupations. We continuously monitor new research to update our assessments.
- Massenkoff & McCrory (2026) - Labor market impacts of AI: observed exposure metric from real Claude usage dataPrimary
- Eloundou et al. (2023) - GPTs are GPTs: theoretical task exposure framework (beta scores 0, 0.5, 1)
- Brynjolfsson et al. (2025) - Canaries in the Coal Mine: observed employment effects using ADP payroll microdata
- U.S. Bureau of Labor Statistics (2024) - Employment Projections 2024-2034 with occupation-level growth rates
- O*NET SOC Classification System - standardized occupation and task taxonomy used across all assessments
Metrics Explained
We use four primary metrics to quantify AI's impact on each occupation. Each metric captures a different dimension of how AI interacts with work tasks.
- Overall Exposure
- A combined metric that synthesizes theoretical and observed exposure data to provide a single summary score (0-100) of how much an occupation's tasks overlap with current AI capabilities.
- Theoretical Exposure
- Measures what AI could potentially automate based on academic research and capability assessments. Derived primarily from Eloundou et al. (2023) beta task exposure scores, which rate each task's susceptibility to LLM automation.
- Observed Exposure
- Measures what AI actually does in practice based on real-world usage data. Derived from Anthropic's analysis of millions of Claude conversations mapped to O*NET occupational tasks, reflecting actual adoption patterns.
- Automation Risk
- The probability of significant job displacement within the assessed time frame. Combines exposure metrics with employment trend data, wage levels, and task substitutability to estimate the likelihood of workforce reduction.
Exposure Level Classification
| Level | Score Range |
|---|---|
| Very High | > 70 |
| High | 50 - 70 |
| Medium | 30 - 50 |
| Low | 15 - 30 |
| Very Low | < 15 |
Analysis Framework
We decompose each occupation into individual tasks and assess the degree to which current and near-future AI systems can automate each task. This task-level approach provides more nuanced insights than whole-occupation estimates.
- Task-Level Decomposition
- Each occupation is broken down into its constituent tasks using O*NET's Detailed Work Activities (DWAs). We assess each task independently rather than making blanket judgments about entire occupations.
- Beta Score Methodology
- Following Eloundou et al. (2023), each task receives a beta score: 0 (no exposure), 0.5 (partial exposure with human oversight), or 1 (full exposure to AI automation). These scores are aggregated to produce occupation-level metrics.
- Time Series Construction (2023-2028)
- We construct time series data using actual measurements for 2023-2025 and estimated projections for 2026-2028. Actual data is clearly distinguished from estimates in all visualizations using solid vs. dashed lines.
- Projection Methodology
- Forward-looking estimates (2026-2028) are based on observed trend rates, announced AI capability improvements, and BLS employment projections. All projected values are clearly marked as estimates.
Data Quality & Limitations
Transparency about our limitations is essential to responsible analysis. Users should consider these factors when interpreting our data.
- Sample Sizes
- Our observed exposure data is based on millions of Claude conversations, providing robust statistical coverage for high-usage occupations. However, less common occupations may have smaller sample sizes and wider confidence intervals.
- Geographic & Occupational Coverage
- Currently covering 55 occupations with plans to expand to 200+. Data primarily reflects the U.S. labor market and English-language AI interactions, which may not fully represent global patterns.
- Update Frequency
- Core metrics are updated when new research publications or data releases become available. BLS projections are updated annually. Observed exposure data is updated as new analyses are published by Anthropic.
- Theoretical vs. Observed Gap
- There is often a significant gap between theoretical exposure (what AI could do) and observed exposure (what AI actually does). Adoption barriers, regulatory constraints, and organizational inertia mean actual AI impact typically lags behind technical capabilities.
Update History
We maintain a transparent record of major data updates and methodology changes.
Initial Launch
Launched with 55 occupations across 14 categories. Integrated Anthropic labor market report data, Eloundou theoretical exposure framework, and BLS 2024-2034 employment projections.
Expansion to 200+ Occupations
Gradual expansion of occupation coverage using AI-assisted analysis combined with manual expert review. Additional data sources and regional labor market data will be incorporated.
Key References
All data points are linked to their original sources. We provide full citation information for transparency and independent verification. See the complete reference list below.
References
All data sources and research papers cited in our analysis.
12 references
- [1]Report
Anthropic Research Team
“Labor market impacts of AI: A new measure and early evidence”
Anthropic, 2026.
Introduces 'observed exposure' metric combining theoretical LLM capabilities with real-world Claude usage data. Finds Computer Programmers at 75% coverage, while actual adoption remains far below theoretical capacity.
- [2]Report
Ruth Appel, Maxim Massenkoff, Peter McCrory, Miles McCain, Ryan Heller, Tyler Neylon, Alex Tamkin
“Anthropic Economic Index report: economic primitives”
Anthropic, 2026.
Defines five economic primitives for AI task classification: complexity, skills, use case, autonomy, and success rate. 34% of Claude.ai usage in Computer & Math occupations.
- [3]Working Paper
Andrew Johnston, Christos Makridis
“The Labor Market Effects of Generative AI: A Difference-in-Differences Analysis”
SSRN, 2025.
Applies difference-in-differences methodology to measure generative AI's labor market effects across occupations.
- [4]Paper
Erik Brynjolfsson, Bharat Chandar, Ruyu Chen
Stanford Digital Economy Lab, 2025.
Young software developers (22-25) see ~20% employment decline from 2022 peak. 13% decline for early-career workers in AI-exposed occupations. Uses ADP payroll microdata.
- [5]Report
Kunal Handa, Alex Tamkin, Miles McCain, Saffron Huang, Esin Durmus, Sarah Heck, Jared Mueller, Jerry Hong, Stuart Ritchie, Tim Belonax, Kevin K. Troy, Dario Amodei, Jared Kaplan, Jack Clark, Deep Ganguli
“Which Economic Tasks are Performed with AI? Evidence from Millions of Claude Conversations”
Anthropic, 2025.
Analyzes millions of Claude conversations to map AI usage to O*NET occupational tasks. 36% of occupations have significant AI task coverage.
- [6]Working Paper
Menaka Hampole, Dimitris Papanikolaou, Lawrence DW Schmidt, Bryan Seegmiller
“Artificial Intelligence and the Labor Market”
National Bureau of Economic Research, 2025.
Instruments for firm-level AI adoption using historical university hiring networks. Firms with AI-related hiring history face lower adoption costs.
- [7]Article
Sarah Eckhardt, Nathan Goldschlag
“AI and Jobs: The Final Word (Until the Next One)”
Economic Innovation Group (EIG), 2025.
Finds AI effect on jobs 'invisible' by conventional metrics. Highly exposed workers show 0.30pp unemployment increase vs. 0.94pp for less exposed. Only ~9% of businesses report using AI.
- [8]Dataset
U.S. Bureau of Labor Statistics
“Employment Projections: 2024-2034”
U.S. Bureau of Labor Statistics, 2024.
Projects 5.2M new jobs 2024-2034 (+3.1% total). Computer & Math +10.1%. Retail declining. AI impacts incorporated for first time in BLS projections.
- [9]Paper
Xiang Hui, Oren Reshef, Luofeng Zhou
“The Short-Term Effects of Generative Artificial Intelligence on Employment”
Organization Science, 2024.
Studies effects of generative AI on freelance platforms. Finds immediate negative impact on earnings and employment for workers in AI-exposed tasks.
- [10]Paper
Tyna Eloundou, Sam Manning, Pamela Mishkin, Daniel Rock
“GPTs are GPTs: An early look at the labor market impact potential of large language models”
arXiv, 2023.
80% of U.S. workforce could have 10%+ tasks affected by LLMs. 19% may see 50%+ tasks impacted. Introduces beta task exposure metric (0, 0.5, 1).
- [11]Paper
Daron Acemoglu, David Autor, Jonathon Hazell, Pasciano Restrepo
“Artificial Intelligence and Jobs: Evidence from Online Vacancies”
Journal of Labor Economics, 2022. DOI: 10.1086/718327
Analyzes AI's impact on job postings using vacancy data. Finds AI adoption displaces some tasks while creating demand for new AI-complementary skills.
- [12]Paper
Daron Acemoglu, Pasciano Restrepo
“Robots and Jobs: Evidence from US Labor Markets”
Journal of Political Economy, 2020. DOI: 10.1086/705716
Estimates that one additional robot per thousand workers reduces employment-to-population ratio by 0.2pp and wages by 0.42%.