The Commonplace
Home Dashboard Papers Evidence Syntheses Digests 🎲
← Papers

Generative AI lifts workplace productivity through automation and decision support, but gains are uneven and accompanied by worker concerns about skill obsolescence and job security; evidence is concentrated in developed economies, leaving the effects in developing countries poorly understood.

Generative AI in the Workplace: A Systematic Review of Productivity Effects, Employment Perceptions, and Job Insecurity
Varnita Dubey · Fetched April 11, 2026 · International journal of research and innovation in social science
semantic_scholar review_meta medium evidence 7/10 relevance DOI Source PDF
Generative AI tends to improve workplace productivity through task automation and decision support, but effects are heterogeneous and accompanied by mixed employee perceptions and job insecurity that shape upskilling and resistance to adoption.

The growing adoption of generative artificial intelligence (AI) in workplace settings has generated significant interest in its implications for productivity, employee perceptions, and job security. This systematic literature review synthesises findings from 40 empirical and conceptual studies published between 2020 and 2025 across organisational and professional contexts to evaluate the multifaceted impact of generative AI on organisational and workforce outcomes. Following the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) framework, a structured search was conducted across Google Scholar and Dimensions.ai, yielding 3,252 database records, with 8 additional hand-searched studies, of which 40 met the inclusion criteria. The review identifies consistent evidence of productivity improvements driven by task automation, decision support, and knowledge augmentation. However, these gains are accompanied by mixed employee perceptions, with increased efficiency and job satisfaction coexisting alongside concerns about skill obsolescence and role displacement. Job insecurity emerges as a critical mediating factor influencing employee attitudes and behavioral responses, including upskilling intentions and resistance to technological change. Importantly, the review reveals a significant research gap in the comparative understanding of generative AI's impact across developed and developing economies, where differences in technological infrastructure, labor market dynamics, and skill distributions may lead to uneven outcomes. The findings highlight that the effects of generative AI are heterogeneous and context-dependent, shaped by job roles, skill levels, and institutional environments. By integrating fragmented literature into a cohesive framework, this study contributes to the emerging discourse on AI-driven workplacetransformation and offers implications for managers and policymakers to ensure more balanced, inclusive, and context-sensitive AI adoption strategies.

Summary

Main Finding

Generative AI (LLMs and tools like ChatGPT and Copilot) produces substantial and robust productivity gains in knowledge-intensive tasks, but these gains are highly heterogeneous and accompanied by mixed employee perceptions and rising job insecurity. The net effects depend on task-technology fit, organizational deployment logic (automation vs. augmentation), worker skill, and institutional context—with a pronounced informational and capability divide between developed and developing economies.

Key Points

  • Productivity

    • Consistent empirical evidence of large gains in knowledge work: examples reported include ~40% faster task completion (Noy & Zhang, 2023), ~15% higher customer-service output (Brynjolfsson et al., 2025), and ~55.8% faster coding with Copilot (Peng et al., 2023).
    • Gains are conditional: LLMs improve performance within their capability frontier but can degrade outcomes on tasks beyond that frontier (Dell'Acqua et al., 2023).
    • Mechanisms: task automation, decision support, and knowledge augmentation; strongest effects in language-, synthesis-, and information-heavy tasks.
    • Heterogeneity: effects vary by occupation, individual openness/skill, and implementation design; evidence concentrated in developed, knowledge-intensive sectors.
  • Employee perceptions

    • Ambivalent attitudes: many workers report increased efficiency, confidence, and satisfaction when AI is framed/implemented as a complementary resource.
    • Negative reactions cluster around identity threat, reliability/bias concerns, surveillance, and poor change management, producing resistance even when productivity benefits exist.
    • Implementation factors (training, psychological safety, participation) materially affect perceptions.
  • Job insecurity and labor effects

    • Job insecurity is a key mediator linking AI adoption to employee wellbeing, commitment, and behavior (e.g., upskilling, turnover intention).
    • LLM exposure is greatest among high-wage, high-education occupations—reversing earlier automation patterns and producing potential middle-skill displacement and labor-market polarization.
    • Reskilling programs exist but are uneven; anticipatory insecurity causes organizational harms even before displacement materializes.
  • Developing vs developed economies

    • Literature is geographically skewed toward developed countries; evidence from South Asia, Sub-Saharan Africa, and Latin America is sparse.
    • Developing-economy barriers: weak digital infrastructure, talent shortages, regulatory gaps, limited change-management capacity.
    • AI benefits in developing contexts appear concentrated in large firms and high-skill workers, risking widened inequality (the "Generative AI Divide").
  • Theoretical framing

    • Review draws on Task-Technology Fit (task-specific gains), Job Demands-Resources (AI as both resource and demand), and the Automation–Augmentation paradox (trade-offs between short-term substitution gains and long-term engagement/adaptability).

Data & Methods

  • Review type: Systematic literature review following PRISMA; narrative synthesis (no meta-analysis due to heterogeneous designs/measures).
  • Timeframe and scope: Studies published January 2020–December 2025; focus on generative AI, LLMs, ChatGPT, Copilot in workplace contexts.
  • Search and selection:
    • Databases searched: Google Scholar and Dimensions.ai; plus hand-searching of references (8 additional records).
    • Initial hits: 3,252 records (1,940 Google Scholar; 1,312 Dimensions.ai); after deduplication 2,892 unique records.
    • Screening: title and abstract screening reduced pool to full-text review; final included studies = 40.
  • Inclusion criteria: studies addressing generative AI in workplace settings reporting on productivity, perceptions, job insecurity, displacement or skill change; empirical, review, or conceptual studies; English; full text accessible.
  • Evidence types: randomized controlled trials, field experiments, controlled experiments, surveys, qualitative studies, systematic reviews, conceptual papers, country analyses.
  • Quality assessment: three-tier framework applied by study design (internal validity for experiments, representativeness and measurement validity for surveys, analytical rigor for qualitative/conceptual). Greater weight given to peer-reviewed experiments and quasi-experiments when resolving conflicts.
  • Limitations of methods noted: reliance on Google Scholar/Dimensions (Scopus/Web of Science not accessed), geographic bias in primary studies, and heterogeneity precluding pooled effect estimates.

Implications for AI Economics

  • For empirical modeling and measurement

    • Treat productivity gains as conditional and task-specific rather than uniform: incorporate task-technology fit measures and heterogeneity across worker skill distributions.
    • Include job insecurity and psychological responses as mediating variables when estimating labor supply, effort, and turnover costs.
    • Expand outcome sets beyond output/time (quality, autonomy, innovation, wellbeing, reskilling behavior).
    • Prioritize causal designs (field RCTs, quasi-experiments) and representative longitudinal surveys to capture short- and medium-run dynamics.
  • For labor-market and distributional analysis

    • Expect potential polarization: rising demand at top-skilled and low-skilled tasks with displacement risk for middle-skilled administrative roles; model heterogeneous wage and employment effects accordingly.
    • Incorporate firm-size and sectoral heterogeneity—large firms and frontier firms may capture disproportionate gains, amplifying within-country inequality.
    • Evaluate redistribution and social-insurance policies (training subsidies, wage insurance, active labor-market policies) in welfare and policy-counterfactual models.
  • For cross-country and development economics

    • The "Generative AI Divide" implies different aggregate productivity and distributional outcomes across countries; models should account for infrastructure, human capital, regulatory capacity, and adoption asymmetries.
    • Research and policy should prioritize capacity-building: digital infrastructure, training ecosystems, regulation that addresses bias and surveillance, and support for SMEs to adopt augmentation-focused AI.
  • For firm strategy and policy prescriptions

    • Augmentation-first deployment (design AI to complement workers) is likelier to yield sustainable productivity and preserve human capital—models of firm behavior should include managerial choice over automation vs. augmentation.
    • Policies encouraging transparent change management, worker participation, and subsidized upskilling can reduce insecurity-driven frictions and improve realized productivity.
    • Regulatory and governance frameworks need to be context-sensitive: universal rules may misfire where institutional capacity and labor-market structures differ.
  • Research gaps / priorities

    • Comparative empirical work across developing economies and sectors beyond knowledge-intensive settings.
    • Longer-term longitudinal studies on career trajectories, wages, and firm-level investment dynamics.
    • Meta-analytic synthesis once standardized effect measures accumulate.
    • Studies measuring how specific implementation choices (incentives, monitoring, governance) mediate productivity vs. wellbeing trade-offs.

Summary conclusion: Generative AI can raise aggregate and individual productivity substantially, but economic impacts depend critically on tasks, institutions, and deployment choices. AI economics research and policy should shift from asking whether generative AI matters to asking how, for whom, and under what governance regimes the gains are realized and equitably distributed.

Assessment

Paper Typereview_meta Evidence Strengthmedium — The paper synthesises 40 empirical and conceptual studies and reports consistent patterns (productivity gains, mixed perceptions), but it does not produce new causal estimates; included studies are heterogeneous with few strong causal designs reported, and no quantitative meta-analysis is presented to aggregate effects. Methods Rigormedium — The review follows PRISMA and uses a structured search (3,252 records screened plus hand searches), which indicates systematic procedure, but the search is limited to two databases (Google Scholar, Dimensions.ai), inclusion/exclusion criteria and quality appraisal details are not fully described, and no formal meta-analytic aggregation of effect sizes is reported, leaving risks of selection and publication bias. SampleSystematic review of 40 included studies (empirical and conceptual) published 2020–2025, identified from 3,252 database records (Google Scholar and Dimensions.ai) plus 8 hand-searched studies; the body comprises surveys, case studies, interviews, observational analyses and a small number of field/experimental designs across organisational and professional contexts, with heavier representation from developed-country settings. Themesproductivity human_ai_collab skills_training labor_markets adoption GeneralizabilityGeographic bias toward studies in developed economies; comparative evidence for developing countries is limited, Possible publication and database selection bias (search limited to Google Scholar and Dimensions.ai), Likely English-language bias (language restrictions not specified), Heterogeneity in study designs, outcomes and measures limits pooled inferences, Few longitudinal or strong causal studies among included research, Sector- and role-specific findings may not generalize across industries, Rapid evolution of generative AI tools (2020–2025) may limit temporal generalizability

Claims (9)

ClaimDirectionConfidenceOutcomeDetails
This systematic literature review synthesised findings from 40 empirical and conceptual studies published between 2020 and 2025 using the PRISMA framework (search across Google Scholar and Dimensions.ai), yielding 3,252 database records plus 8 hand-searched studies, of which 40 met the inclusion criteria. Other null_result high systematic review sample and search yield (records screened/included)
n=40
0.4
There is consistent evidence of productivity improvements from generative AI in workplace settings, driven by task automation, decision support, and knowledge augmentation. Organizational Efficiency positive high productivity improvements (via task automation, decision support, knowledge augmentation)
n=40
0.24
Generative AI adoption is associated with mixed employee perceptions: some studies report increased efficiency and higher job satisfaction. Worker Satisfaction positive high reported efficiency gains and job satisfaction
n=40
0.24
These positive perceptions coexist with employee concerns about skill obsolescence related to generative AI. Skill Obsolescence negative high concerns about skill obsolescence
n=40
0.24
Employees express concerns about role displacement (job loss or role changes) associated with generative AI adoption. Job Displacement negative high perceived risk of role displacement / job loss
n=40
0.24
Job insecurity emerges as a critical mediating factor influencing employee attitudes and behavioural responses to generative AI, including upskilling intentions and resistance to technological change. Skill Acquisition negative high upskilling intentions and resistance to technological change (mediated by job insecurity)
n=40
0.24
There is a significant research gap in comparative understanding of generative AI's impact across developed and developing economies; differences in infrastructure, labour markets, and skill distributions may lead to uneven outcomes. Inequality null_result high comparative evidence on generative AI impacts across developed vs. developing economies (absence/lack)
n=40
0.4
The effects of generative AI on work and organisations are heterogeneous and context-dependent, shaped by job roles, skill levels, and institutional environments. Task Allocation mixed high heterogeneity of AI effects across roles/skills/institutions
n=40
0.24
The review integrates fragmented literature into a cohesive framework and offers implications for managers and policymakers to pursue more balanced, inclusive, and context-sensitive AI adoption strategies. Governance And Regulation positive high guidance for managerial and policy decision-making regarding AI adoption
n=40
0.04

Notes