Evidence (2469 claims)
Adoption
5539 claims
Productivity
4793 claims
Governance
4333 claims
Human-AI Collaboration
3326 claims
Labor Markets
2657 claims
Innovation
2510 claims
Org Design
2469 claims
Skills & Training
2017 claims
Inequality
1378 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 402 | 112 | 67 | 480 | 1076 |
| Governance & Regulation | 402 | 192 | 122 | 62 | 790 |
| Research Productivity | 249 | 98 | 34 | 311 | 697 |
| Organizational Efficiency | 395 | 95 | 70 | 40 | 603 |
| Technology Adoption Rate | 321 | 126 | 73 | 39 | 564 |
| Firm Productivity | 306 | 39 | 70 | 12 | 432 |
| Output Quality | 256 | 66 | 25 | 28 | 375 |
| AI Safety & Ethics | 116 | 177 | 44 | 24 | 363 |
| Market Structure | 107 | 128 | 85 | 14 | 339 |
| Decision Quality | 177 | 76 | 38 | 20 | 315 |
| Fiscal & Macroeconomic | 89 | 58 | 33 | 22 | 209 |
| Employment Level | 77 | 34 | 80 | 9 | 202 |
| Skill Acquisition | 92 | 33 | 40 | 9 | 174 |
| Innovation Output | 120 | 12 | 23 | 12 | 168 |
| Firm Revenue | 98 | 34 | 22 | — | 154 |
| Consumer Welfare | 73 | 31 | 37 | 7 | 148 |
| Task Allocation | 84 | 16 | 33 | 7 | 140 |
| Inequality Measures | 25 | 77 | 32 | 5 | 139 |
| Regulatory Compliance | 54 | 63 | 13 | 3 | 133 |
| Error Rate | 44 | 51 | 6 | — | 101 |
| Task Completion Time | 88 | 5 | 4 | 3 | 100 |
| Training Effectiveness | 58 | 12 | 12 | 16 | 99 |
| Worker Satisfaction | 47 | 32 | 11 | 7 | 97 |
| Wages & Compensation | 53 | 15 | 20 | 5 | 93 |
| Team Performance | 47 | 12 | 15 | 7 | 82 |
| Automation Exposure | 24 | 22 | 9 | 6 | 62 |
| Job Displacement | 6 | 38 | 13 | — | 57 |
| Hiring & Recruitment | 41 | 4 | 6 | 3 | 54 |
| Developer Productivity | 34 | 4 | 3 | 1 | 42 |
| Social Protection | 22 | 10 | 6 | 2 | 40 |
| Creative Output | 16 | 7 | 5 | 1 | 29 |
| Labor Share of Income | 12 | 5 | 9 | — | 26 |
| Skill Obsolescence | 3 | 20 | 2 | — | 25 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
Org Design
Remove filter
Autonomous agents are likely to substitute for routine, structured cognitive tasks while complementing higher-level managerial and strategic tasks, accelerating task reallocation within firms.
Synthesis of prior literature (generative AI productivity findings) and observational deployment patterns from Alfred AI indicating substitution of routine tasks and continued human involvement in oversight/strategy.
Realized productivity gains from AI agents are materially constrained by governance complexity, model reliability limits (errors, hallucinations, edge cases), orchestration challenges across tools/data/human teams, and continued need for human-in-the-loop oversight.
Qualitative operational impacts and deployment observations from Alfred AI implementations, documented frictions in policies, safety constraints, error handling, and orchestration; evidence drawn from observational deployments and operational logs.
Effectiveness and safety of AI agents require structured guardrails and human-in-the-loop designs; AI agents function as scalable cognitive infrastructure only conditional on such governance.
Synthesis of deployment experience and analysis of constraints; recommendation grounded in observed model reliability issues, governance complexity, and oversight needs from the Alfred AI experiments.
Deployment of AI agents shifts demand toward roles focused on oversight, orchestration, prompt/agent engineering, and governance, creating new types of labor that may offset some direct labor reductions.
Authors' inference based on observed need for human oversight and orchestration in deployments; not quantitatively measured in the study (no headcount or labor-share data reported).
Labor-market consequences will involve reallocation effects: routine-task automation, rising returns to managerial and technical skills, and potential within-firm wage dispersion.
Synthesis of labor economics theory and prior empirical work on automation; book recommends matched employer-employee panel studies to trace these effects but does not report such new panel results.
AI’s effects vary by industry, task composition, and firm capabilities; high-data, standardized-task sectors see faster, deeper impacts.
Cross-sector examples and theoretical arguments about task routineness and data intensity; calls for heterogeneity-aware empirical designs (e.g., difference-in-differences with staggered adoption).
Automation of routine tasks raises demand for cognitive, interpersonal, and technical skills; firms face reskilling needs and changing task allocation between humans and machines.
Task-level analytic framework and literature review on automation effects; book recommends empirical approaches (e.g., occupation and job-task data) to quantify these changes but does not present a single large empirical estimate.
Managers shift from routine decision execution to tasks involving oversight, interpretation, strategic design, and ethical stewardship of AI systems.
Qualitative case studies and literature review of task-level research; suggested task-analytic methods rather than reporting a specific empirical task dataset.
AI complements some researcher tasks (idea generation, analysis, writing) and substitutes others (routine editing, literature searches), changing skill demand and training priorities.
Stated under Labor Market Effects. Supported conceptually and likely by task-level studies or surveys; abstract doesn't cite specific empirical evidence or measurement details.
Impacts of AI adoption are broad, affecting individual researcher productivity, team workflows, and institutional outcomes in scholarly communication and digital scholarship.
Key Points summary. Basis likely includes mixed-methods evidence (surveys/interviews at individual and team levels, case studies, platform usage data) synthesized in the paper; abstract lacks detail on scope and samples.
Investment in governance and training is a necessary cost to realize sustained returns from generative AI; these costs influence adoption timing and the distribution of benefits.
Conceptual argument from the review supported by case examples and economic reasoning about complementary investments.
There is a risk of wage polarization: increased returns to AI‑complementary skills and potential downward pressure on wages for automatable tasks.
Theoretical synthesis drawing on economic models of skill‑biased technological change and early empirical observations; no definitive causal wage studies reported.
Generative AI will drive occupational reallocation by substituting routine cognitive tasks while complementing higher‑order cognitive and monitoring skills.
Theoretical labor economics arguments synthesized with early empirical examples; no large‑scale causal labor market study provided in the review.
TGAIF clarifies where GenAI acts as a complement (augmenting consultant capability) versus where it risks substitution.
Conceptual distinction and mapping presented in the TGAIF derived from practitioner accounts; theoretical/qualitative, not empirically quantified across tasks.
TGAIF implies reallocation of work away from GenAI‑suitable subtasks (routine synthesis, drafting, summarization) toward tasks where human judgment and client interaction add most value.
Based on authors' inductive analysis of practitioner interviews describing which subtasks firms consider suitable for GenAI and which require human oversight; qualitative, not quantitatively tracked reallocation.
Aligning consulting tasks with generative-AI capabilities via a Task–GenAI Fit (TGAIF) framework can unlock substantial efficiency gains while containing key risks (notably hallucinations and loss of skill retention).
Inductive framework developed from qualitative, interpretive interviews with practitioners at leading German management‑consulting firms. The abstract does not report sample size, interview protocol, or quantitative validation; evidence is based on practitioner reports and the authors' synthesis.
DAR implies changes to labor and contracting: reversible AI leadership reshapes task boundaries, demand for oversight skills, and should be reflected in contracts and procurement with explicit authority-reversal rules and audit obligations.
Theoretical/ normative argument in implications section; no empirical labor or contract data included.
AI substitutes for routine coding tasks but complements higher-order tasks such as system architecture, integration, and orchestration.
Interpretation from qualitative evidence at Netlight where practitioners used AI for routine chores while retaining control of higher-order design tasks; no quantitative task-time displacement data presented.
Human roles are shifting toward oversight, curation, specification, and orchestration of multiple AI components and tools.
Synthesized from practitioner descriptions and changing task allocations observed in the Netlight fieldwork (interviews/observations); no longitudinal measurement of role changes reported.
Short-run consumer gains from faster, cheaper service can be undermined by trust losses from hallucinations or perceived deception, reducing long-term consumer surplus.
Conceptual welfare analysis and cited case examples in the literature; no longitudinal consumer-surplus measurement provided in this review.
Conventional productivity metrics (e.g., handle time) may misstate value because they do not capture multi-dimensional impacts like quality and trust.
Conceptual critique and synthesis of measurement challenges discussed in the literature; no empirical measurement study presented in this review.
There is potential for substantial cost savings and throughput gains in repetitive, high-volume interactions, but these are offset by costs for integration, monitoring, and error remediation.
Industry case examples and conceptual cost–benefit reasoning aggregated in the review; the paper contains no new quantitative cost estimates or sample-based measurements.
Generative AI will substitute for routine service tasks while complementing skilled workers for escalations and complex problem solving, shifting labor demand toward supervisory and relationship-focused roles.
Economic and labor-market analyses synthesized in the review; projections are inferential and based on heterogeneous secondary sources, not primary labor-market experiments.
Full automation of customer service is suboptimal because persistent risks (hallucinations, contextual errors, lack of genuine empathy, integration complexity) remain; hybrid human–AI systems achieve the best outcomes.
Synthesis of documented failure modes and practitioner case examples from the literature; no primary experimental data or controlled trials in this review. Inference based on heterogeneous empirical reports and conceptual analyses.
Welfare effects of democratized access to AI-assisted ideation are ambiguous: access could democratize innovation but also amplify low-quality outputs and misinformation absent proper curation.
Theoretical discussion and empirical examples of misinformation/low-quality outputs from LLMs cited in the review; no comprehensive welfare accounting provided.
Net gains in innovation from increased idea volume depend on complementary human capacity for curation and development; raw increases in ideas do not automatically translate into higher-quality innovation.
Synthesis noting studies where idea quantity rose but downstream quality or successful development did not necessarily increase; review highlights heterogeneity across workflows and dependence on human integration.
The most effective deployment model is a 'cognitive co-pilot' in which AI expands and challenges the idea space while humans provide curation, strategic evaluation, and experiential judgment.
Prescriptive conclusion drawn from synthesis of studies where human-AI collaboration (human curation/selection) produced better downstream outcomes than AI-alone outputs; evidence heterogenous and largely short-term.
Generative AI functions as a dual-purpose cognitive tool: a high-volume catalyst for divergent idea generation and a structured assistant for decomposing complex problems.
Nano-review / synthesis of existing empirical literature on LLM-assisted creativity and problem-solving, drawing on experimental ideation tasks, design/ideation studies, and applied case evidence; no original dataset or new experiments in this paper.
Net value from generative AI is contingent: gains are largest where breadth of ideas and rapid iteration matter, and smaller or riskier where deep domain expertise, tacit knowledge, or high-stakes judgments are required.
Synthesis of heterogeneous empirical results showing task-dependent benefits; argument grounded in observed differences across lab and field contexts and documented limitations in domain-specific performance.
Data-driven HRM reinforces skill-biased technological change: routine HR tasks are being substituted by automation while demand rises for analytical and interpersonal skills.
Theoretical implication and synthesis across studies in the review noting automation of routine tasks and increased demand for analytic/interpersonal skills.
Adoption will be heterogeneous and distributional effects will follow: organizational readiness, regulatory environments, and industry structure will drive uneven adoption and competitive impacts.
Review finds varying adoption patterns in empirical and practitioner literature and synthesizes theoretical reasons for heterogeneity; empirical causal estimates are noted as scarce.
One-off AI features typically produce limited returns unless organizations build complementary human and process capabilities and adapt governance and incentives.
Interpretive synthesis of case studies and practitioner guidance showing short-lived or limited benefits from isolated feature deployments without complementary investments.
Governance reduces downside risk (compliance fines, outages) but raises implementation costs; economic assessments must weigh risk-adjusted returns.
Conceptual economic argument in the paper; supported by reasoning and practitioner experience but not by empirical cost–benefit studies within the article.
When evaluating GenAI investments, firms should treat prompt-fraud controls and monitoring as persistent operating costs rather than one-time setup costs.
Practical recommendation informed by conceptual cost and governance analysis; not supported by longitudinal cost studies in the paper.
Smaller firms or departments using shadow AI may realize productivity gains but face outsized fraud exposure due to weaker controls.
Theoretical trade-off analysis in the implications section; no empirical firm-level comparisons or experiments presented.
Safer scaling of automation may increase substitution of routine ERP/CRM tasks while governance and oversight roles create complementary high-skill positions (e.g., compliance engineers, auditors, prompt engineers).
Labor-market implications presented as theoretical reasoning based on how governance and automation interact; informed by practitioner observation but not empirically tested in the paper.
Decentralized governance can foster a more pluralistic ecosystem but may produce fragmentation and underinvestment in public‑goods data infrastructure.
Inferential implication based on U.S. texts showing plural institutional actors and literature on decentralized governance trade‑offs; not empirically measured in this study.
Decentralized, rights‑based regimes (e.g., U.S.) may preserve individual and institutional controls that can increase transactional frictions but support market entry via clearer procedural safeguards.
Inferential implication from the U.S. policy texts' emphasis on rights, transparency, and procedural safeguards; based on coded document content rather than observed market outcomes.
Centralized, sovereignty‑oriented regimes (e.g., China) may enable large, state‑facilitated data aggregation projects that lower data costs for favored actors but restrict cross‑border flows and outsider access.
Inferential implication drawn from the Chinese policy texts' developmentalist and techno‑sovereignty framing together with literature on state‑led data aggregation (no empirical measurement of outcomes in this study).
Openness and security are better understood as co‑evolving, layered institutional processes rather than strict, mutually exclusive binaries.
Conceptual synthesis grounded in the document coding results and an extension of modular coordination theory developed in the paper.
AI-enabled macro and fiscal models can improve policy testing and contingency planning but require transparency, validation, and safeguards against overreliance.
Conceptual argument and illustrative examples; no empirical trials or model performance metrics reported.
AI shifts the locus of economic governance from static rules to living systems that anticipate shocks and adapt in real time.
Policy-analytic framing and scenario-based reasoning within the book; supported by illustrative examples rather than empirical measurement.
Federated systems can lower barriers for advertisers and publishers who previously lacked aggregated data, but they also create coordination and infrastructure costs that may favor organizations able to invest in shared infrastructures or consortium governance.
Economic analysis and policy discussion outlining effects on entry, competition, and coordination costs. Evidence is conceptual; no empirical market-entry case studies provided.
Land-transfer effects on AGTFP are positive but constrained: institutional frictions limit the contribution of land transfer to green transformation.
Mediation results indicating a positive but limited indirect effect via land transfer/scale expansion, supplemented by discussion of institutional barriers in the paper.
The AI-as-advisor approach has limitations: people frequently ignore accurate advice, rely too much on inaccurate advice, and their decision-making skills may deteriorate over time.
Paper asserts these limitations in motivation/background and/or derives them from observed behavior in experiments (stated in abstract as known problems with AI-as-advisor).
Our findings surface practical limits on the complexity people can manage in human-AI negotiation.
Synthesis claim based on the empirical study varying number of issues and observed decline in performance beyond three issues; presented as a conceptual/practical implication of the results.
Beyond an environment-specific optimum, scaling further degrades institutional fitness because trust erosion and cost penalties outweigh marginal capability gains.
Analytical argument from the Institutional Scaling Law together with illustrative examples and discussion of mechanisms (trust erosion, cost penalties) in the paper.
Traditional ex ante regulatory approaches struggle to keep pace with AI development, exacerbating the 'pacing problem' and the Collingridge dilemma.
Theoretical/legal literature review and conceptual argument presented in the paper (no empirical sample or quantitative data reported in the abstract).
Low internal conflict or unanimity can be diagnostic of variance depletion (i.e., exclusion) rather than healthy integration, so governance systems should treat low conflict as a potential red flag until heterogeneity integration is verified.
Interpretive policy implication derived from the model's demonstration that exclusionary processes can produce deceptively low observed disagreement while increasing fragility; this recommendation is based on theoretical reasoning without empirical validation in the paper.
The paper identifies five structural challenges arising from the memory governance gap: memory silos across agent workflows; governance fragmentation across teams and tools; unstructured memories unusable by downstream systems; redundant context delivery in autonomous multi-step executions; and silent quality degradation without feedback loops.
Qualitative analysis and problem framing presented in the paper (authors' identification of five specific challenges).