The Commonplace
Home Dashboard Papers Evidence Digests 🎲
← Papers

Task-level approaches reveal that low-skilled tasks face the largest displacement risk while high-skilled work is more likely to be augmented, but much of the literature rests on static, pre-2020 mappings and limited causal evidence that rapid LLM advances have already outpaced.

Recent Methodologies on AI and Labour - a Desk Review
Elena Papagiannaki, Joana Geisler · March 06, 2026 · Edinburgh Napier Research Repository (Edinburgh Napier University)
openalex review_meta n/a evidence 7/10 relevance DOI Source PDF
Recent AI–labor economics has moved to task-level, NLP/LLM-based measurement showing heterogeneous effects—higher displacement risk for lower-skilled tasks and augmentation for higher-skilled tasks—but many studies rely on static assumptions and weak causal designs and some pre-2020 conclusions are already outdated by rapid AI progress.

This desk review toolkit brings together recent methodologies used to analyse AI’s effects on employment, wages, and productivity. Shift from Occupation-Level to Task-Level Analysis: Recent methodologies increasingly focus on tasks rather than entire occupations, recognising the heterogeneity within jobs. For example, Felten et al. (2023) and Eloundou et al. (2023) use AI exposure indices to measure task-specific impacts.Integration of Advanced Technologies: Methods now incorporate Natural Language Processing (NLP) (e.g., BERT, LSTM) and LLMs (e.g., GPT-4) to analyse job descriptions and predict automation risks (Xu et al., 2025; Hampole et al., 2025).Scenario Planning and Policy Focus: Think tanks like TBI and IPPR, and organisations like IMF emphasise scenario-based modelling to inform policy, highlighting the need for reskilling and labour market reforms (TBI, 2024; IPPR, 2024, Korinek, 2023). They assume different ‘initial conditions’ in adoption to estimate different conclusions in employment, wages, productivity ect. Studies consistently find that AI’s impact varies by skill level, with low-skilled workers facing higher displacement risks and high-skilled workers benefiting from augmentation (Brynjolfsson et al., 2023; Chen et al., 2024). Many methodologies struggle with static assumptions, lack of causal evidence, and overreliance on theoretical models. For instance, Acemoglu & Restrepo (2022) assume fixed comparative advantage, while Webb (2020) ignores adaptation. As of year 2025, conclusions of some of the pre-2020 research is already redundant in terms of what professions will remain or not. The pre-2020 conclusion that creative and intellectual occupations will remain in high demand, has already been disproven.

Summary

Main Finding

Recent methodologies in AI–labor economics have shifted from occupation-level to task-level analysis, increasingly using NLP and large language models to measure AI exposure and predict effects on employment, wages, and productivity. These methods reveal heterogeneous impacts across tasks and skill levels: lower-skilled tasks face higher displacement risk while higher-skilled workers are more likely to be augmented. However, many studies rely on static assumptions and theoretical models, limiting causal inference and making some pre-2020 conclusions about which professions will endure already outdated by 2025.

Key Points

  • Task-level focus
    • Research is moving from treating occupations as homogeneous to mapping task-by-task exposure to AI (e.g., Felten et al., 2023; Eloundou et al., 2023).
    • Task approaches capture within-occupation heterogeneity in automation/augmentation risk.
  • Use of advanced ML and LLMs
    • Methods increasingly apply NLP (BERT, LSTM) and LLMs (GPT-4) to parse job descriptions, map skills/tasks, and predict automation risk (Xu et al., 2025; Hampole et al., 2025).
  • Scenario and policy-oriented modelling
    • Think tanks and international organisations (TBI, IPPR, IMF) emphasise scenario planning with differing adoption initial conditions to inform reskilling and labour-market policy (TBI, 2024; IPPR, 2024; Korinek, 2023).
  • Distributional findings
    • Robust pattern: AI’s effects vary by skill level — displacement risk concentrated among lower-skilled tasks, augmentation and wage gains more likely for high-skilled tasks (Brynjolfsson et al., 2023; Chen et al., 2024).
  • Methodological limitations
    • Many studies use static assumptions (fixed comparative advantage, no adaptation), limited causal identification, and model-dependent projections (critique of Acemoglu & Restrepo, 2022; Webb, 2020).
    • Rapid technological change (post-2020 advances) has already made some earlier profession-level conclusions obsolete by 2025.

Data & Methods

  • Data sources (commonly used / implied)
    • Job postings and descriptions, occupational task databases (e.g., O*NET style), employer/household surveys, administrative payroll data, firm-level productivity measures.
  • Measurement approaches
    • AI exposure indices: task-level scores indicating susceptibility to AI automation or augmentation (Felten et al., 2023; Eloundou et al., 2023).
    • NLP/LLM pipelines: use BERT/LSTM/Transformer models to extract tasks/skills from free-text job ads and to map tasks to AI capabilities (Xu et al., 2025; Hampole et al., 2025).
    • Scenario modelling: counterfactual simulations with different adoption speeds, policy responses, and initial conditions to bound possible employment/wage/productivity trajectories (TBI, IPPR, IMF).
  • Analytical strategies and shortcomings
    • Cross-sectional exposure correlations and panel-difference analyses are common but often lack strong causal identification (endogeneity of adoption, unobserved confounders).
    • Static equilibrium/representative-agent models neglect dynamic reallocation, task re-bundling, and firm-level heterogeneity.
    • Rapid model and capability change (LLMs and multimodal AI) challenge out-of-sample validity of indices developed pre-2020.

Implications for AI Economics

  • Research priorities
    • Prioritise causal, dynamic analyses: exploit natural experiments, firm-level adoption variation, and longitudinal microdata to identify causal effects on employment, wages, and productivity.
    • Update and validate task–AI mapping continuously using state-of-the-art LLMs and human-in-the-loop validation to keep exposure indices current.
    • Incorporate task reallocation and worker adaptation into models (reskilling, on-the-job learning, task switching).
  • Policy implications
    • Design flexible reskilling and retraining programs targeted at high-risk tasks and low-skilled workers, informed by task-level exposure maps.
    • Use scenario planning to prepare labour-market institutions for a range of adoption speeds; policy should focus on smoothing transitions (unemployment insurance, portable benefits, active labour-market policies).
    • Consider incentives for firms to invest in augmentation (human+AI) rather than pure replacement where socially beneficial.
  • Measurement and forecasting
    • Move away from occupation-level forecasts toward task-level, continuously updated indicators linked to real-world adoption measures (firm purchases, API usage, procurement).
    • Develop standards for documenting AI-capability changes and their mapping to tasks so conclusions remain transparent and updatable.
  • Cautions for interpretation
    • Be wary of projections based on pre-2020 assumptions; rapid AI capability improvements (LLMs, multimodal models) alter which tasks are automatable.
    • Emphasise uncertainty and distributional outcomes — aggregate productivity gains can coincide with concentrated displacement without complementary policy.

Assessment

Paper Typereview_meta Evidence Strengthn/a — This is a synthesis of recent literature rather than original causal empirical analysis; the underlying studies vary in quality and causal identification, so a single evidence-strength rating is not applicable. Methods Rigormedium — The review systematically summarizes newer task-level and NLP/LLM-based measurement approaches and highlights methodological shortcomings (static assumptions, weak causal identification), but it does not appear to follow a formal systematic-review or meta-analytic protocol with pre-specified inclusion criteria and quantitative synthesis. SampleAggregates evidence from multiple data sources used across studies: online job postings and free-text job descriptions, occupational task databases (O*NET-style), employer and household surveys, administrative payroll and firm-level productivity data, and scenario-model outputs from think tanks and international organisations. Themeslabor_markets productivity skills_training adoption GeneralizabilityMany indices and mappings are US/English-centric (O*NET, job-posting datasets) limiting cross-country applicability, Pre-2020 task–AI mappings may be invalid post-LLM and multimodal advances, reducing temporal generalizability, Firm-level heterogeneity and selection into adoption mean exposure scores may not translate to realized impact across all industries, Job-posting data omit informal work and incumbents' on-the-job task changes, biasing representativeness, Scenario and model-based projections depend on assumed adoption speeds and policy regimes, limiting external validity

Claims (15)

ClaimDirectionConfidenceOutcomeDetails
Recent research in AI–labor economics has shifted from occupation-level analysis to task-level analysis, mapping task-by-task exposure to AI. Research Productivity null_result high granularity of exposure measurement (occupation-level vs. task-level AI exposure)
0.04
Task-level approaches capture within-occupation heterogeneity in automation and augmentation risk that occupation-level analyses miss. Automation Exposure null_result high heterogeneity in automation/augmentation risk across tasks within occupations
0.04
Methods increasingly apply advanced NLP and large language models (BERT, LSTM, GPT-4) to parse job descriptions, map skills/tasks, and predict automation risk. Research Productivity null_result medium task/skill extraction and AI-exposure prediction accuracy from free-text job descriptions
0.02
Think tanks and international organisations are emphasising scenario planning with differing adoption initial conditions to inform reskilling and labour-market policy. Governance And Regulation positive medium policy scenario outputs (projected employment/wage/productivity under alternative adoption assumptions)
0.02
A robust empirical pattern in the literature is that AI’s effects vary by skill level: displacement risk is concentrated among lower-skilled tasks while augmentation and wage gains are more likely for higher-skilled tasks. Skill Obsolescence mixed medium displacement risk, augmentation incidence, and wage changes disaggregated by skill level
0.02
Many studies rely on static assumptions (fixed comparative advantage, no adaptation) and theoretical models, which limits causal inference and makes projections model-dependent. Research Productivity null_result high strength of causal identification and robustness of projected employment/wage outcomes
0.04
Rapid post-2020 advances in AI (LLMs and multimodal models) have already rendered some pre-2020 profession-level conclusions obsolete by 2025. Research Productivity negative medium validity/applicability of pre-2020 profession-level forecasts in 2025
0.02
Commonly used data sources for measuring AI exposure include job postings and descriptions, occupational task databases (O*NET-style), employer/household surveys, administrative payroll data, and firm-level productivity measures. Research Productivity null_result high coverage and types of data used for AI exposure and labour-outcome measurement
0.04
Researchers construct AI exposure indices at the task level to indicate susceptibility to AI automation or augmentation. Automation Exposure null_result high task-level AI exposure scores
0.04
NLP/LLM pipelines are used to extract tasks and skills from free-text job ads and to map those tasks to AI capabilities. Research Productivity null_result medium task/skill extraction performance and task-to-capability mapping
0.02
Scenario modelling in the reviewed literature typically uses counterfactual simulations with different adoption speeds, policy responses, and initial conditions to bound possible employment, wage, and productivity trajectories. Research Productivity null_result medium range of projected employment/wage/productivity trajectories across scenarios
0.02
Common empirical strategies (cross-sectional exposure correlations and panel-difference analyses) often lack strong causal identification due to endogeneity of adoption and unobserved confounders. Research Productivity null_result high validity of causal estimates of AI adoption effects on labour outcomes
0.04
Static equilibrium and representative-agent models neglect dynamic reallocation, task re-bundling, and firm-level heterogeneity, limiting their realism for forecasting labour outcomes under AI adoption. Research Productivity null_result high completeness/realism of economic models used to forecast labour-market effects
0.04
Policy should prioritise flexible reskilling and retraining programs targeted at high-risk tasks and low-skilled workers, informed by task-level exposure maps. Training Effectiveness positive medium effectiveness of reskilling/training programs in mitigating displacement and improving employment/wages for targeted workers
0.02
Measurement and forecasting should move away from occupation-level forecasts toward task-level, continuously updated indicators linked to real-world adoption measures (firm purchases, API usage, procurement). Research Productivity positive medium forecast accuracy and timeliness of AI exposure indicators
0.02

Notes