The Commonplace
Home Dashboard Papers Evidence Digests 🎲
← Papers

A skills-focused simulation suggests AI’s technical reach in the U.S. labor market is far larger and more geographically widespread than visible adoption indicates: current AI capabilities overlap with about 11.7% of U.S. wages (~$1.2tn), compared with roughly 2.2% (~$211bn) reflected in present adoption concentrated in tech hubs.

The Iceberg Index: Measuring Workforce Exposure in the AI Economy
Ayush Chopra, Santanu Bhattacharya, DeAndrea Salvador, Ayan Paul, Teddy Wright, Aditi Garg, Feroz Ahmad, Alice C. Schwarze, Ramesh Raskar, Prasanna Balaprakash · Fetched March 10, 2026 · Robotics
semantic_scholar descriptive low evidence 8/10 relevance DOI Source
Using a skills-centered Iceberg Index and a 151M-agent simulation, the paper finds current AI capabilities technically overlap with roughly 11.7% (~$1.2 trillion) of U.S. wage value—about five times the wage share visible in current AI adoption (~2.2%, ~$211 billion)—and that this exposure is geographically diffuse beyond tech hubs.

Artificial Intelligence is reshaping America’s over $9.4 trillion labor market, with cascading effects that extend far beyond visible technology sectors. When AI automates quality control in automotive plants, consequences spread through logistics networks, supply chains, and local service economies. Yet traditional workforce metrics cannot capture these ripple effects: they measure employment outcomes after disruption occurs, not where AI capabilities overlap with human skills before adoption crystallizes. Project Iceberg addresses this gap using Large Population Models to simulate the human–AI labor market, representing 151 million workers as autonomous agents executing over 32,000 skills across 3,000 counties and interacting with thousands of AI tools. It introduces the Iceberg Index, a skills-centered metric that measures the wage value of skills AI systems can perform within each occupation. The Index captures technical exposure, where AI can perform occupational tasks, not displacement outcomes or adoption timelines. Analysis shows that visible AI adoption concentrated in computing and technology (2.2% of wage value, approximately $211 billion) represents only the tip of the iceberg. Technical capability extends far below the surface through cognitive automation spanning administrative, financial, and professional services (11.7%, approximately $1.2 trillion). This exposure is fivefold larger and geographically distributed across all states rather than confined to coastal hubs. Traditional indicators such as GDP, income, and unemployment explain less than 5% of this skills-based variation, underscoring why new indices are needed to capture exposure in the AI economy. By simulating how capabilities may spread under alternative scenarios, Project Iceberg enables policymakers and business leaders to identify exposure hotspots, prioritize training and infrastructure investments, and test interventions before committing billions to implementation. Iceberg is built with the AgentTorch framework.  

Summary

Main Finding

Project Iceberg introduces a skills-centered measure (the Iceberg Index) and a large-scale agent-based simulation showing that AI technical capability in the U.S. labor market is substantially larger and far more geographically diffuse than visible adoption suggests. Visible AI adoption concentrated in computing/technology represents about 2.2% of U.S. wage value (~$211 billion), while cognitive automation potential across administrative, financial, and professional services amounts to 11.7% (~$1.2 trillion). Traditional macro indicators explain <5% of the resulting geographic variation, so skill-level exposure metrics and simulations are needed to anticipate and plan for AI’s broader labor-market effects.

Key Points

  • Problem: Conventional workforce metrics capture outcomes after disruption (employment, wages) but not where AI capabilities overlap with human skills before adoption.
  • Iceberg Index: A skills-centered metric measuring the wage value of the specific skills AI systems can perform within each occupation. It quantifies technical exposure (capability overlap), not displacement, adoption timelines, or realized outcomes.
  • Scale & resolution: Model represents 151 million U.S. workers as autonomous agents, covers 32,000+ skills across ~3,000 counties, and links agents to thousands of AI tools.
  • Key quantitative results:
    • Visible AI adoption in tech/computing: ~2.2% of wage value (~$211B).
    • Broader cognitive automation potential in admin, financial, professional services: ~11.7% (~$1.2T) — ~5× larger and spread across all states, not just coastal hubs.
    • Traditional indicators (GDP, income, unemployment) explain <5% of state- and county-level variation in skills-based exposure.
  • Purpose: The framework lets policymakers and firms simulate alternative scenarios to locate exposure hotspots, prioritize investments (training, infrastructure), and test interventions prior to large deployments.
  • Implementation: Model and simulations built with the AgentTorch framework.

Data & Methods

  • Modeling approach: Large Population Models / agent-based simulation (agents = 151M workers) representing individual workers’ occupations, skills portfolios, wages, and county locations. Agents interact with a catalog of thousands of AI tools mapped to task/skill capabilities.
  • Skill taxonomy: ~32,000 distinct skills mapped to occupations and tasks; each skill assigned wage value contribution within occupations to compute the Iceberg Index.
  • Iceberg Index computation: For each occupation/geography, sum the wage value associated with skills that current AI capabilities can perform. The index is a measure of technical exposure (potential substitutability/complementarity), not an adoption or displacement probability.
  • Geographic granularity: Analysis across ~3,000 U.S. counties and all states, enabling spatial dispersion analysis.
  • Scenarios & simulations: Counterfactual scenarios that vary capability diffusion, adoption rates, policy interventions, and firm behavior to explore how exposures might translate into outcomes.
  • Validation & limitations noted: The Index captures capability overlap but not firm adoption choices, regulatory constraints, social acceptance, complementarity effects, or reallocation dynamics. Simulations depend on mapping of AI tools to skills and assumptions embedded in AgentTorch agent behaviors.

Implications for AI Economics

  • Measurement: Skills-centered exposure metrics (like the Iceberg Index) are essential complements to traditional employment and macro indicators for anticipating AI impacts before outcomes materialize.
  • Policy targeting: Because exposure is geographically widespread and concentrated in service and administrative work as well as tech, policies should be spatially and sectorally granular (county- or state-level interventions, not only coastal/hub strategies).
  • Labor-market interventions: Use the model to prioritize retraining, upskilling, and credentialing where wage-value exposure is high; align training investments to skills where AI is unlikely to fully substitute and where complementarities can raise productivity and wages.
  • Infrastructure & economic development: Identify local infrastructure needs (broadband, compute, AI adoption support for SMEs) in exposure hotspots to avoid uneven benefits and displacement.
  • Scenario testing for costly decisions: Firms and governments can simulate policy levers (subsidies, tax incentives, public-training programs, adoption incentives) to estimate likely labor-market trajectories before committing large budgets.
  • Research agenda: Encourage more granular skill-to-AI-capability mappings, longitudinal tracking of adoption vs. exposure, and integration of firm behavior and regulatory dynamics into agent-based models to move from exposure assessment toward outcome prediction.
  • Caution: The Iceberg Index signals where capability exists, not where job losses will occur or when. Policy responses should combine exposure analysis with empirical monitoring of adoption, complementarities, and local labor reallocation.

Assessment

Paper Typedescriptive Evidence Strengthlow — The study provides a detailed, model-based measure of technical exposure (capability overlap) but does not observe or identify causal effects on employment, wages, or firm outcomes; results depend on mapping assumptions, simulated agent behavior, and hypothetical adoption scenarios rather than empirical variation in realized outcomes. Methods Rigormedium — The work uses large-scale, high-resolution inputs (151M simulated workers, ~32k skills, county-level geography, and thousands of AI tools) and a transparent agent-based framework, which support rich scenario analysis; however, key mappings (skills-to-capabilities, wage-value attribution) and behavioral rules are inherently subjective, sensitivity to assumptions is likely, and firm/adoption dynamics and institutional constraints are not fully modeled or validated against longitudinal empirical outcomes. SampleAn agent-based simulation representing ~151 million U.S. workers linked to ~3,000 counties, ~3,000 occupations, a taxonomy of ~32,000 skills, and a catalog of thousands of AI tools; skill-to-occupation wage-value contributions are used to compute the Iceberg Index (wage exposure measures) and generate counterfactual diffusion/adoption scenarios using the AgentTorch framework. Themesproductivity labor_markets GeneralizabilityU.S.-only; results may not hold in other countries with different occupational structures or institutions, Time-specific to the set of AI capabilities and tools mapped at the time of the study; capability changes will alter exposure estimates, Relies on skill-to-AI and skill-to-wage mappings that may have measurement error or subjective coding choices, Measures technical exposure, not realized adoption or labor-market outcomes—doesn't account for firm behavior, regulation, or social acceptance, Agent behavior and macro reallocation dynamics are simulated with simplifying assumptions that may not reflect real-world frictions, May underrepresent informal, gig, or undocumented work and firm heterogeneity across regions

Claims (14)

ClaimDirectionConfidenceOutcomeDetails
AI technical capability in the U.S. labor market is substantially larger and far more geographically diffuse than visible adoption suggests. Automation Exposure mixed medium difference between skills-based exposure (Iceberg Index) and visible AI-adoption wage-share, and geographic dispersion of exposure across counties/states
n=151000000
0.05
Visible AI adoption concentrated in computing/technology represents about 2.2% of U.S. wage value (~$211 billion). Labor Share negative medium percent of U.S. wage value attributed to visible AI adoption (2.2%) and corresponding dollar estimate (~$211B)
n=151000000
2.2% (~$211 billion)
0.05
Broader cognitive automation potential across administrative, financial, and professional services amounts to 11.7% (~$1.2 trillion). Labor Share negative medium percent of U.S. wage value exposed to current AI capabilities (Iceberg Index = 11.7%) and dollar estimate (~$1.2T)
n=151000000
11.7% (~$1.2 trillion)
0.05
The broader cognitive automation potential is roughly five times larger than visible adoption and is geographically widespread (present across all states, not only coastal hubs). Automation Exposure negative medium ratio of Iceberg Index wage-share to visible-adoption wage-share (~5×) and geographic distribution of Iceberg Index values across counties/states
n=151000000
≈5× (Iceberg Index vs visible adoption); geographically widespread
0.05
Traditional macro indicators (GDP, income, unemployment) explain less than 5% of the state- and county-level variation in skills-based exposure. Automation Exposure negative medium percent variance explained (R^2) in the Iceberg Index by traditional macro indicators (<5%)
n=3000
<5% explained variance (R²) by macro indicators
0.05
The Iceberg Index is a skills-centered metric that measures the wage value of specific skills AI systems can perform within each occupation; it quantifies technical exposure (capability overlap), not displacement, adoption timelines, or realized outcomes. Automation Exposure null_result high Iceberg Index value (wage-value of automatable skills per occupation/geography)
0.09
The simulation model represents 151 million U.S. workers as autonomous agents, covers 32,000+ distinct skills, links agents to thousands of AI tools, and provides county-level resolution (~3,000 U.S. counties). Other null_result high model scope metrics: number of agents (151M), skills (~32k), counties (~3k), and linked AI tools (thousands)
n=151000000
151,000,000 agents; ~32,000 skills; ~3,000 counties; thousands of AI tools
0.09
The framework supports counterfactual scenario simulations that vary capability diffusion, adoption rates, policy interventions, and firm behavior to explore how exposures might translate into outcomes. Other positive high simulated labor-market trajectories under alternative counterfactual parameterizations (no single numeric outcome)
0.09
Model and simulations are implemented with the AgentTorch framework. Other null_result high implementation platform (AgentTorch)
Implemented in AgentTorch
0.09
The Iceberg Index captures capability overlap but does not capture firm adoption choices, regulatory constraints, social acceptance, complementarity effects, or worker reallocation dynamics. Other null_result high scope/limitations of the Iceberg Index (what it does not measure)
0.09
The framework can help policymakers and firms locate exposure hotspots, prioritize investments in training and infrastructure, and test interventions prior to large deployments. Governance And Regulation positive medium decision-support capabilities: identification of exposure hotspots and evaluation of intervention scenarios (qualitative outcome)
0.05
Because exposure is geographically widespread and concentrated in service and administrative work as well as tech, policy responses should be spatially and sectorally granular (county- or state-level interventions rather than only coastal/hub strategies). Governance And Regulation positive medium recommended policy targeting granularity based on spatial and sectoral distribution of Iceberg Index values
n=3000
0.05
Research should prioritize more granular skill-to-AI-capability mappings, longitudinal tracking of adoption vs. exposure, and integration of firm behavior and regulatory dynamics into agent-based models to move from exposure assessment toward outcome prediction. Research Productivity positive low proposed research directions (not an empirical measurement)
0.03
The Iceberg Index indicates where capability exists but does not indicate whether or when job losses will occur. Automation Exposure null_result high distinction between capability exposure (Iceberg Index) and realized job loss/adoption timing (not measured)
0.09

Notes