Generative AI lifts workplace productivity through automation and decision support, but gains are uneven and accompanied by worker concerns about skill obsolescence and job security; evidence is concentrated in developed economies, leaving the effects in developing countries poorly understood.
The growing adoption of generative artificial intelligence (AI) in workplace settings has generated significant interest in its implications for productivity, employee perceptions, and job security. This systematic literature review synthesises findings from 40 empirical and conceptual studies published between 2020 and 2025 across organisational and professional contexts to evaluate the multifaceted impact of generative AI on organisational and workforce outcomes. Following the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) framework, a structured search was conducted across Google Scholar and Dimensions.ai, yielding 3,252 database records, with 8 additional hand-searched studies, of which 40 met the inclusion criteria. The review identifies consistent evidence of productivity improvements driven by task automation, decision support, and knowledge augmentation. However, these gains are accompanied by mixed employee perceptions, with increased efficiency and job satisfaction coexisting alongside concerns about skill obsolescence and role displacement. Job insecurity emerges as a critical mediating factor influencing employee attitudes and behavioral responses, including upskilling intentions and resistance to technological change. Importantly, the review reveals a significant research gap in the comparative understanding of generative AI's impact across developed and developing economies, where differences in technological infrastructure, labor market dynamics, and skill distributions may lead to uneven outcomes. The findings highlight that the effects of generative AI are heterogeneous and context-dependent, shaped by job roles, skill levels, and institutional environments. By integrating fragmented literature into a cohesive framework, this study contributes to the emerging discourse on AI-driven workplacetransformation and offers implications for managers and policymakers to ensure more balanced, inclusive, and context-sensitive AI adoption strategies.
Summary
Main Finding
Generative AI (LLMs and tools like ChatGPT and Copilot) produces substantial and robust productivity gains in knowledge-intensive tasks, but these gains are highly heterogeneous and accompanied by mixed employee perceptions and rising job insecurity. The net effects depend on task-technology fit, organizational deployment logic (automation vs. augmentation), worker skill, and institutional context—with a pronounced informational and capability divide between developed and developing economies.
Key Points
-
Productivity
- Consistent empirical evidence of large gains in knowledge work: examples reported include ~40% faster task completion (Noy & Zhang, 2023), ~15% higher customer-service output (Brynjolfsson et al., 2025), and ~55.8% faster coding with Copilot (Peng et al., 2023).
- Gains are conditional: LLMs improve performance within their capability frontier but can degrade outcomes on tasks beyond that frontier (Dell'Acqua et al., 2023).
- Mechanisms: task automation, decision support, and knowledge augmentation; strongest effects in language-, synthesis-, and information-heavy tasks.
- Heterogeneity: effects vary by occupation, individual openness/skill, and implementation design; evidence concentrated in developed, knowledge-intensive sectors.
-
Employee perceptions
- Ambivalent attitudes: many workers report increased efficiency, confidence, and satisfaction when AI is framed/implemented as a complementary resource.
- Negative reactions cluster around identity threat, reliability/bias concerns, surveillance, and poor change management, producing resistance even when productivity benefits exist.
- Implementation factors (training, psychological safety, participation) materially affect perceptions.
-
Job insecurity and labor effects
- Job insecurity is a key mediator linking AI adoption to employee wellbeing, commitment, and behavior (e.g., upskilling, turnover intention).
- LLM exposure is greatest among high-wage, high-education occupations—reversing earlier automation patterns and producing potential middle-skill displacement and labor-market polarization.
- Reskilling programs exist but are uneven; anticipatory insecurity causes organizational harms even before displacement materializes.
-
Developing vs developed economies
- Literature is geographically skewed toward developed countries; evidence from South Asia, Sub-Saharan Africa, and Latin America is sparse.
- Developing-economy barriers: weak digital infrastructure, talent shortages, regulatory gaps, limited change-management capacity.
- AI benefits in developing contexts appear concentrated in large firms and high-skill workers, risking widened inequality (the "Generative AI Divide").
-
Theoretical framing
- Review draws on Task-Technology Fit (task-specific gains), Job Demands-Resources (AI as both resource and demand), and the Automation–Augmentation paradox (trade-offs between short-term substitution gains and long-term engagement/adaptability).
Data & Methods
- Review type: Systematic literature review following PRISMA; narrative synthesis (no meta-analysis due to heterogeneous designs/measures).
- Timeframe and scope: Studies published January 2020–December 2025; focus on generative AI, LLMs, ChatGPT, Copilot in workplace contexts.
- Search and selection:
- Databases searched: Google Scholar and Dimensions.ai; plus hand-searching of references (8 additional records).
- Initial hits: 3,252 records (1,940 Google Scholar; 1,312 Dimensions.ai); after deduplication 2,892 unique records.
- Screening: title and abstract screening reduced pool to full-text review; final included studies = 40.
- Inclusion criteria: studies addressing generative AI in workplace settings reporting on productivity, perceptions, job insecurity, displacement or skill change; empirical, review, or conceptual studies; English; full text accessible.
- Evidence types: randomized controlled trials, field experiments, controlled experiments, surveys, qualitative studies, systematic reviews, conceptual papers, country analyses.
- Quality assessment: three-tier framework applied by study design (internal validity for experiments, representativeness and measurement validity for surveys, analytical rigor for qualitative/conceptual). Greater weight given to peer-reviewed experiments and quasi-experiments when resolving conflicts.
- Limitations of methods noted: reliance on Google Scholar/Dimensions (Scopus/Web of Science not accessed), geographic bias in primary studies, and heterogeneity precluding pooled effect estimates.
Implications for AI Economics
-
For empirical modeling and measurement
- Treat productivity gains as conditional and task-specific rather than uniform: incorporate task-technology fit measures and heterogeneity across worker skill distributions.
- Include job insecurity and psychological responses as mediating variables when estimating labor supply, effort, and turnover costs.
- Expand outcome sets beyond output/time (quality, autonomy, innovation, wellbeing, reskilling behavior).
- Prioritize causal designs (field RCTs, quasi-experiments) and representative longitudinal surveys to capture short- and medium-run dynamics.
-
For labor-market and distributional analysis
- Expect potential polarization: rising demand at top-skilled and low-skilled tasks with displacement risk for middle-skilled administrative roles; model heterogeneous wage and employment effects accordingly.
- Incorporate firm-size and sectoral heterogeneity—large firms and frontier firms may capture disproportionate gains, amplifying within-country inequality.
- Evaluate redistribution and social-insurance policies (training subsidies, wage insurance, active labor-market policies) in welfare and policy-counterfactual models.
-
For cross-country and development economics
- The "Generative AI Divide" implies different aggregate productivity and distributional outcomes across countries; models should account for infrastructure, human capital, regulatory capacity, and adoption asymmetries.
- Research and policy should prioritize capacity-building: digital infrastructure, training ecosystems, regulation that addresses bias and surveillance, and support for SMEs to adopt augmentation-focused AI.
-
For firm strategy and policy prescriptions
- Augmentation-first deployment (design AI to complement workers) is likelier to yield sustainable productivity and preserve human capital—models of firm behavior should include managerial choice over automation vs. augmentation.
- Policies encouraging transparent change management, worker participation, and subsidized upskilling can reduce insecurity-driven frictions and improve realized productivity.
- Regulatory and governance frameworks need to be context-sensitive: universal rules may misfire where institutional capacity and labor-market structures differ.
-
Research gaps / priorities
- Comparative empirical work across developing economies and sectors beyond knowledge-intensive settings.
- Longer-term longitudinal studies on career trajectories, wages, and firm-level investment dynamics.
- Meta-analytic synthesis once standardized effect measures accumulate.
- Studies measuring how specific implementation choices (incentives, monitoring, governance) mediate productivity vs. wellbeing trade-offs.
Summary conclusion: Generative AI can raise aggregate and individual productivity substantially, but economic impacts depend critically on tasks, institutions, and deployment choices. AI economics research and policy should shift from asking whether generative AI matters to asking how, for whom, and under what governance regimes the gains are realized and equitably distributed.
Assessment
Claims (9)
| Claim | Direction | Confidence | Outcome | Details |
|---|---|---|---|---|
| This systematic literature review synthesised findings from 40 empirical and conceptual studies published between 2020 and 2025 using the PRISMA framework (search across Google Scholar and Dimensions.ai), yielding 3,252 database records plus 8 hand-searched studies, of which 40 met the inclusion criteria. Other | null_result | high | systematic review sample and search yield (records screened/included) |
n=40
0.4
|
| There is consistent evidence of productivity improvements from generative AI in workplace settings, driven by task automation, decision support, and knowledge augmentation. Organizational Efficiency | positive | high | productivity improvements (via task automation, decision support, knowledge augmentation) |
n=40
0.24
|
| Generative AI adoption is associated with mixed employee perceptions: some studies report increased efficiency and higher job satisfaction. Worker Satisfaction | positive | high | reported efficiency gains and job satisfaction |
n=40
0.24
|
| These positive perceptions coexist with employee concerns about skill obsolescence related to generative AI. Skill Obsolescence | negative | high | concerns about skill obsolescence |
n=40
0.24
|
| Employees express concerns about role displacement (job loss or role changes) associated with generative AI adoption. Job Displacement | negative | high | perceived risk of role displacement / job loss |
n=40
0.24
|
| Job insecurity emerges as a critical mediating factor influencing employee attitudes and behavioural responses to generative AI, including upskilling intentions and resistance to technological change. Skill Acquisition | negative | high | upskilling intentions and resistance to technological change (mediated by job insecurity) |
n=40
0.24
|
| There is a significant research gap in comparative understanding of generative AI's impact across developed and developing economies; differences in infrastructure, labour markets, and skill distributions may lead to uneven outcomes. Inequality | null_result | high | comparative evidence on generative AI impacts across developed vs. developing economies (absence/lack) |
n=40
0.4
|
| The effects of generative AI on work and organisations are heterogeneous and context-dependent, shaped by job roles, skill levels, and institutional environments. Task Allocation | mixed | high | heterogeneity of AI effects across roles/skills/institutions |
n=40
0.24
|
| The review integrates fragmented literature into a cohesive framework and offers implications for managers and policymakers to pursue more balanced, inclusive, and context-sensitive AI adoption strategies. Governance And Regulation | positive | high | guidance for managerial and policy decision-making regarding AI adoption |
n=40
0.04
|