Task-level approaches reveal that low-skilled tasks face the largest displacement risk while high-skilled work is more likely to be augmented, but much of the literature rests on static, pre-2020 mappings and limited causal evidence that rapid LLM advances have already outpaced.
This desk review toolkit brings together recent methodologies used to analyse AI’s effects on employment, wages, and productivity. Shift from Occupation-Level to Task-Level Analysis: Recent methodologies increasingly focus on tasks rather than entire occupations, recognising the heterogeneity within jobs. For example, Felten et al. (2023) and Eloundou et al. (2023) use AI exposure indices to measure task-specific impacts.Integration of Advanced Technologies: Methods now incorporate Natural Language Processing (NLP) (e.g., BERT, LSTM) and LLMs (e.g., GPT-4) to analyse job descriptions and predict automation risks (Xu et al., 2025; Hampole et al., 2025).Scenario Planning and Policy Focus: Think tanks like TBI and IPPR, and organisations like IMF emphasise scenario-based modelling to inform policy, highlighting the need for reskilling and labour market reforms (TBI, 2024; IPPR, 2024, Korinek, 2023). They assume different ‘initial conditions’ in adoption to estimate different conclusions in employment, wages, productivity ect. Studies consistently find that AI’s impact varies by skill level, with low-skilled workers facing higher displacement risks and high-skilled workers benefiting from augmentation (Brynjolfsson et al., 2023; Chen et al., 2024). Many methodologies struggle with static assumptions, lack of causal evidence, and overreliance on theoretical models. For instance, Acemoglu & Restrepo (2022) assume fixed comparative advantage, while Webb (2020) ignores adaptation. As of year 2025, conclusions of some of the pre-2020 research is already redundant in terms of what professions will remain or not. The pre-2020 conclusion that creative and intellectual occupations will remain in high demand, has already been disproven.
Summary
Main Finding
Recent methodologies in AI–labor economics have shifted from occupation-level to task-level analysis, increasingly using NLP and large language models to measure AI exposure and predict effects on employment, wages, and productivity. These methods reveal heterogeneous impacts across tasks and skill levels: lower-skilled tasks face higher displacement risk while higher-skilled workers are more likely to be augmented. However, many studies rely on static assumptions and theoretical models, limiting causal inference and making some pre-2020 conclusions about which professions will endure already outdated by 2025.
Key Points
- Task-level focus
- Research is moving from treating occupations as homogeneous to mapping task-by-task exposure to AI (e.g., Felten et al., 2023; Eloundou et al., 2023).
- Task approaches capture within-occupation heterogeneity in automation/augmentation risk.
- Use of advanced ML and LLMs
- Methods increasingly apply NLP (BERT, LSTM) and LLMs (GPT-4) to parse job descriptions, map skills/tasks, and predict automation risk (Xu et al., 2025; Hampole et al., 2025).
- Scenario and policy-oriented modelling
- Think tanks and international organisations (TBI, IPPR, IMF) emphasise scenario planning with differing adoption initial conditions to inform reskilling and labour-market policy (TBI, 2024; IPPR, 2024; Korinek, 2023).
- Distributional findings
- Robust pattern: AI’s effects vary by skill level — displacement risk concentrated among lower-skilled tasks, augmentation and wage gains more likely for high-skilled tasks (Brynjolfsson et al., 2023; Chen et al., 2024).
- Methodological limitations
- Many studies use static assumptions (fixed comparative advantage, no adaptation), limited causal identification, and model-dependent projections (critique of Acemoglu & Restrepo, 2022; Webb, 2020).
- Rapid technological change (post-2020 advances) has already made some earlier profession-level conclusions obsolete by 2025.
Data & Methods
- Data sources (commonly used / implied)
- Job postings and descriptions, occupational task databases (e.g., O*NET style), employer/household surveys, administrative payroll data, firm-level productivity measures.
- Measurement approaches
- AI exposure indices: task-level scores indicating susceptibility to AI automation or augmentation (Felten et al., 2023; Eloundou et al., 2023).
- NLP/LLM pipelines: use BERT/LSTM/Transformer models to extract tasks/skills from free-text job ads and to map tasks to AI capabilities (Xu et al., 2025; Hampole et al., 2025).
- Scenario modelling: counterfactual simulations with different adoption speeds, policy responses, and initial conditions to bound possible employment/wage/productivity trajectories (TBI, IPPR, IMF).
- Analytical strategies and shortcomings
- Cross-sectional exposure correlations and panel-difference analyses are common but often lack strong causal identification (endogeneity of adoption, unobserved confounders).
- Static equilibrium/representative-agent models neglect dynamic reallocation, task re-bundling, and firm-level heterogeneity.
- Rapid model and capability change (LLMs and multimodal AI) challenge out-of-sample validity of indices developed pre-2020.
Implications for AI Economics
- Research priorities
- Prioritise causal, dynamic analyses: exploit natural experiments, firm-level adoption variation, and longitudinal microdata to identify causal effects on employment, wages, and productivity.
- Update and validate task–AI mapping continuously using state-of-the-art LLMs and human-in-the-loop validation to keep exposure indices current.
- Incorporate task reallocation and worker adaptation into models (reskilling, on-the-job learning, task switching).
- Policy implications
- Design flexible reskilling and retraining programs targeted at high-risk tasks and low-skilled workers, informed by task-level exposure maps.
- Use scenario planning to prepare labour-market institutions for a range of adoption speeds; policy should focus on smoothing transitions (unemployment insurance, portable benefits, active labour-market policies).
- Consider incentives for firms to invest in augmentation (human+AI) rather than pure replacement where socially beneficial.
- Measurement and forecasting
- Move away from occupation-level forecasts toward task-level, continuously updated indicators linked to real-world adoption measures (firm purchases, API usage, procurement).
- Develop standards for documenting AI-capability changes and their mapping to tasks so conclusions remain transparent and updatable.
- Cautions for interpretation
- Be wary of projections based on pre-2020 assumptions; rapid AI capability improvements (LLMs, multimodal models) alter which tasks are automatable.
- Emphasise uncertainty and distributional outcomes — aggregate productivity gains can coincide with concentrated displacement without complementary policy.
Assessment
Claims (15)
| Claim | Direction | Confidence | Outcome | Details |
|---|---|---|---|---|
| Recent research in AI–labor economics has shifted from occupation-level analysis to task-level analysis, mapping task-by-task exposure to AI. Research Productivity | null_result | high | granularity of exposure measurement (occupation-level vs. task-level AI exposure) |
0.04
|
| Task-level approaches capture within-occupation heterogeneity in automation and augmentation risk that occupation-level analyses miss. Automation Exposure | null_result | high | heterogeneity in automation/augmentation risk across tasks within occupations |
0.04
|
| Methods increasingly apply advanced NLP and large language models (BERT, LSTM, GPT-4) to parse job descriptions, map skills/tasks, and predict automation risk. Research Productivity | null_result | medium | task/skill extraction and AI-exposure prediction accuracy from free-text job descriptions |
0.02
|
| Think tanks and international organisations are emphasising scenario planning with differing adoption initial conditions to inform reskilling and labour-market policy. Governance And Regulation | positive | medium | policy scenario outputs (projected employment/wage/productivity under alternative adoption assumptions) |
0.02
|
| A robust empirical pattern in the literature is that AI’s effects vary by skill level: displacement risk is concentrated among lower-skilled tasks while augmentation and wage gains are more likely for higher-skilled tasks. Skill Obsolescence | mixed | medium | displacement risk, augmentation incidence, and wage changes disaggregated by skill level |
0.02
|
| Many studies rely on static assumptions (fixed comparative advantage, no adaptation) and theoretical models, which limits causal inference and makes projections model-dependent. Research Productivity | null_result | high | strength of causal identification and robustness of projected employment/wage outcomes |
0.04
|
| Rapid post-2020 advances in AI (LLMs and multimodal models) have already rendered some pre-2020 profession-level conclusions obsolete by 2025. Research Productivity | negative | medium | validity/applicability of pre-2020 profession-level forecasts in 2025 |
0.02
|
| Commonly used data sources for measuring AI exposure include job postings and descriptions, occupational task databases (O*NET-style), employer/household surveys, administrative payroll data, and firm-level productivity measures. Research Productivity | null_result | high | coverage and types of data used for AI exposure and labour-outcome measurement |
0.04
|
| Researchers construct AI exposure indices at the task level to indicate susceptibility to AI automation or augmentation. Automation Exposure | null_result | high | task-level AI exposure scores |
0.04
|
| NLP/LLM pipelines are used to extract tasks and skills from free-text job ads and to map those tasks to AI capabilities. Research Productivity | null_result | medium | task/skill extraction performance and task-to-capability mapping |
0.02
|
| Scenario modelling in the reviewed literature typically uses counterfactual simulations with different adoption speeds, policy responses, and initial conditions to bound possible employment, wage, and productivity trajectories. Research Productivity | null_result | medium | range of projected employment/wage/productivity trajectories across scenarios |
0.02
|
| Common empirical strategies (cross-sectional exposure correlations and panel-difference analyses) often lack strong causal identification due to endogeneity of adoption and unobserved confounders. Research Productivity | null_result | high | validity of causal estimates of AI adoption effects on labour outcomes |
0.04
|
| Static equilibrium and representative-agent models neglect dynamic reallocation, task re-bundling, and firm-level heterogeneity, limiting their realism for forecasting labour outcomes under AI adoption. Research Productivity | null_result | high | completeness/realism of economic models used to forecast labour-market effects |
0.04
|
| Policy should prioritise flexible reskilling and retraining programs targeted at high-risk tasks and low-skilled workers, informed by task-level exposure maps. Training Effectiveness | positive | medium | effectiveness of reskilling/training programs in mitigating displacement and improving employment/wages for targeted workers |
0.02
|
| Measurement and forecasting should move away from occupation-level forecasts toward task-level, continuously updated indicators linked to real-world adoption measures (firm purchases, API usage, procurement). Research Productivity | positive | medium | forecast accuracy and timeliness of AI exposure indicators |
0.02
|