Evidence (11677 claims)
Adoption
7395 claims
Productivity
6507 claims
Governance
5921 claims
Human-AI Collaboration
5192 claims
Org Design
3497 claims
Innovation
3492 claims
Labor Markets
3231 claims
Skills & Training
2608 claims
Inequality
1842 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 609 | 159 | 77 | 738 | 1617 |
| Governance & Regulation | 671 | 334 | 160 | 99 | 1285 |
| Organizational Efficiency | 626 | 147 | 105 | 70 | 955 |
| Technology Adoption Rate | 502 | 176 | 98 | 78 | 861 |
| Research Productivity | 349 | 109 | 48 | 322 | 838 |
| Output Quality | 391 | 121 | 45 | 40 | 597 |
| Firm Productivity | 385 | 46 | 85 | 17 | 539 |
| Decision Quality | 277 | 145 | 63 | 34 | 526 |
| AI Safety & Ethics | 189 | 244 | 59 | 30 | 526 |
| Market Structure | 152 | 154 | 109 | 20 | 440 |
| Task Allocation | 158 | 50 | 56 | 26 | 295 |
| Innovation Output | 178 | 23 | 38 | 17 | 257 |
| Skill Acquisition | 137 | 52 | 50 | 13 | 252 |
| Fiscal & Macroeconomic | 120 | 64 | 38 | 23 | 252 |
| Employment Level | 93 | 46 | 96 | 12 | 249 |
| Firm Revenue | 130 | 43 | 26 | 3 | 202 |
| Consumer Welfare | 99 | 51 | 40 | 11 | 201 |
| Inequality Measures | 36 | 106 | 40 | 6 | 188 |
| Task Completion Time | 134 | 18 | 6 | 5 | 163 |
| Worker Satisfaction | 79 | 54 | 16 | 11 | 160 |
| Error Rate | 64 | 79 | 8 | 1 | 152 |
| Regulatory Compliance | 69 | 66 | 14 | 3 | 152 |
| Training Effectiveness | 82 | 16 | 13 | 18 | 131 |
| Wages & Compensation | 70 | 25 | 22 | 6 | 123 |
| Team Performance | 74 | 16 | 21 | 9 | 121 |
| Automation Exposure | 41 | 48 | 19 | 9 | 120 |
| Job Displacement | 11 | 71 | 16 | 1 | 99 |
| Developer Productivity | 71 | 14 | 9 | 3 | 98 |
| Hiring & Recruitment | 49 | 7 | 8 | 3 | 67 |
| Social Protection | 26 | 14 | 8 | 2 | 50 |
| Creative Output | 26 | 14 | 6 | 2 | 49 |
| Skill Obsolescence | 5 | 37 | 5 | 1 | 48 |
| Labor Share of Income | 12 | 13 | 12 | — | 37 |
| Worker Turnover | 11 | 12 | — | 3 | 26 |
| Industry | — | — | — | 1 | 1 |
AI tools assist across the full research lifecycle: idea generation, study design, literature review and synthesis, data management and analysis, writing/editing, publishing, communication, and compliance.
Key point asserted in the paper. Implied support comes from aggregated reports and studies of tool functionality and user reports (literature review, surveys, case studies). No specific sample or usage statistics provided in the abstract.
AI is becoming an integrated research productivity layer in universities that speeds and improves the entire scholarly workflow — from idea generation through analysis to dissemination — by lowering cognitive and technical burdens, which boosts research quality and institutional research performance.
Statement presented as the paper's main finding. Abstract summarizes "recent evidence" but does not specify original data or methods; likely based on literature synthesis (empirical studies, survey/interview work, case reports) rather than a single original dataset. No sample size, measurement definitions, or identification strategy provided in the abstract.
First‑mover adoption and superior governance can create persistent competitive advantages for firms deploying generative AI effectively.
Theoretical reasoning and case examples from industry reports included in the synthesis; absence of broad causal evidence noted.
Scale and data advantages associated with generative AI adoption may reinforce winner‑take‑all dynamics, favoring large firms that can exploit data and integration economies.
Conceptual argument and industry observations synthesized in the review; no comprehensive market concentration empirical analysis presented.
Realizing sustainable economic value from generative AI requires robust governance, AI literacy, and human‑centric augmentation strategies (AI as assistant, not replacement).
Normative conclusion based on conceptual synthesis of empirical patterns and theoretical arguments in the review.
Generative AI has potential to improve the quality of information processing and the speed of decision‑making.
Conceptual arguments plus early case examples and small empirical studies reported in the literature synthesis; no broad causal estimates provided.
Short‑term deployments of generative AI produce efficiency gains such as time savings and faster turnaround.
Early empirical studies and industry reports summarized in the review; reported case examples of tool deployments (no unified sample size reported).
Generative AI produces measurable gains in operational efficiency and strategic insight.
Synthesized findings and illustrative case examples from early empirical studies and industry reports; authors note lack of large-scale causal evidence.
Generative AI enables scalable personalized communication with customers, employees, and partners.
Aggregation of industry use cases and early empirical reports discussed in the conceptual synthesis (no large-scale causal studies reported).
Generative AI enhances decision support by synthesizing information, surfacing options, and generating explanations for decision‑makers.
Critical literature synthesis and early case examples from industry reports and small studies cited in the review; theoretical evaluation of decision workflows.
Generative AI automates routine administrative workflows and parts of analytical pipelines.
Nano review / conceptual synthesis aggregating early empirical studies, industry reports, and case examples; no original primary dataset reported.
Short-run: measurable productivity gains for many coding tasks imply higher effective output per developer.
Controlled experiments and benchmark tasks that report time savings and/or increased task throughput with LLM assistance; studies often in lab/microtask settings with varying sample sizes.
Organizations will need to build processes and tools (automated testing, static analysis, code review augmented for AI outputs) to realize net benefits safely.
Qualitative case studies and practitioner reports documenting emerging organizational practices and recommendations; derived from observed failure modes and security/IP risks.
The highest value arises when human developers verify, adapt, and integrate AI suggestions—human–AI complementarity.
User studies and controlled experiments showing improved outcomes when humans validate and edit AI outputs; qualitative interviews and case studies reporting effective human-in-the-loop workflows.
These tools lower initial barriers for novices by giving example code, explanations, and templates, potentially accelerating onboarding.
User studies, observational analyses, and qualitative interviews reporting that novices use LLM outputs as examples and templates; evidence primarily short-term and context-dependent.
LLMs are most effective when used interactively as assistants rather than as autonomous code authors.
User studies, observational analyses, and controlled comparisons showing better outcomes for interactive, iterative prompting and verification versus one-shot autonomous code generation; heterogeneous study designs (mostly short-term lab or microtask settings).
LLMs can speed up many programming tasks (boilerplate, code completion, documentation, simple debugging) and change how developers iterate.
Synthesis of controlled experiments and benchmark tasks comparing developer speed/accuracy with and without LLM assistance, supplemented by user studies and observational analyses; sample sizes and tasks vary across studies (typically lab/microtask settings, often tens to low hundreds of participants).
Token taxes incentivize more efficient model designs (fewer tokens per task) and may shift competition toward lightweight models or on-device solutions.
Mechanism-based economic reasoning about price incentives included in the paper; no empirical or simulation evidence provided.
Agent-based models (ABMs) are needed to simulate micro-to-macro dynamics of token taxes because standard representative-agent or DSGE models may miss heterogeneity, network effects, and path dependence.
Methodological argument in the paper advocating ABMs; no ABM results included (proposal only).
Black-box token verification (tamper-evident consumption tokens or receipts tied to API calls) can prove taxable consumption without full model inspection.
Technical proposal for cryptographic/ledgered receipts described in the paper; no prototype, security analysis, or empirical tests provided.
A staged audit pipeline—black-box token verification, norm-based tax rates, then white-box audits—provides a feasible path to design and evaluate token taxes.
Proposed enforcement architecture described in the paper (conceptual design); no deployment or simulation results presented.
Token taxes can be enforced using existing compute-governance and commercial billing infrastructure (API billing, cloud metering, hardware telemetry, attestation).
Technical architecture discussion proposing use of existing billing and telemetry systems; no implementation or pilot data provided.
Compared with robot- or FLOP-based taxes, token taxes better capture where AI-generated value is realized.
Analytic comparison in the paper arguing tokens map to user-facing consumption while FLOP/robot taxes map to inputs; conceptual reasoning rather than empirical test.
The framework enables scenario testing for policies and shocks (e.g., lockdowns, targeted interventions, information campaigns) where human judgment and adaptation matter.
Paper reports experiments across policy regimes and discusses use cases for testing timing, targeting, and communication strategies; however, concrete policy evaluation examples and quantitative policy results are not detailed in the summary.
Experiments run with multiple LLM backends (proprietary and open-source) show qualitatively consistent dynamics, indicating framework stability to model choice.
Cross-backend comparisons and robustness checks reported in the paper; several LLMs used though the exact models and counts are not specified in the summary.
Behavioral changes in the simulation emerge endogenously from cognitive reasoning rather than from parameterized switches, producing context-sensitive, heterogeneous responses.
Description of agent heterogeneity (differences in perceptions, priorities, and local conditions) and use of CoT reasoning per agent; reported emergent, diverse responses in experiments. (Degree of heterogeneity and quantitative heterogeneity metrics not provided in summary.)
LLM-driven agents embedded in a Perception–Deliberation–Action (PDA) loop produce endogenous, human-like adaptive behaviors via Chain-of-Thought reasoning.
Multi-agent simulation where each agent is implemented as an LLM-driven cognitive unit running the PDA loop each timestep; agents use Chain-of-Thought (CoT) prompts/internal reasoning to make decisions. (Exact simulation sample size / population not specified in summary.)
Task‑based, dynamic exposure measures and real‑time data enable earlier detection of displacement risks and reallocation needs than static, occupation‑level extrapolations.
Conceptual argument and proposed architecture; no empirical timing comparison or lead-time statistics provided.
LLMs can be used to score task automation/augmentation plausibility and to detect emergent tasks.
Methodological proposal describing use of LLMs for semantic mapping/scoring of tasks; no empirical validation or accuracy metrics for LLM task scoring provided in the paper.
Modeling nonlinearity (threshold adoption, network spillovers, complementarities) and path dependence in adoption dynamics is necessary rather than relying on linear extrapolation.
Theoretical argument and model suggestions (S‑curve diffusion, agent-based models) in the paper; no empirical comparison demonstrating superior performance provided.
Applying causal inference methods (difference‑in‑differences, synthetic controls, instrumental variables, structural counterfactuals) can distinguish automation (task substitution) from augmentation (productivity/role change) and estimate net employment effects.
Methodological recommendation with examples of applicable identification strategies; no specific empirical applications or results reported in the paper.
Integrating multiple data streams (CPS, LEHD/LODES, UI wage records, administrative microdata, job ads, occupational manuals, enterprise adoption surveys) yields richer gross‑flows and skills measurement than using single data sources.
Proposed data-integration strategy and references to candidate datasets; no empirical demonstration or quantified improvement in measurement presented.
A dynamic Occupational AI Exposure Score (OAIES) can quantify exposure at the task level using LLMs, job‑task matrices (e.g., O*NET), and real‑time job ad / workplace data to capture evolving capability of AI systems.
Methodological description of OAIES construction (mapping tasks to occupations, LLM scoring, weighting by time use/criticality); no empirical implementation or validation data presented in the paper.
Measurement and forecasting should move away from occupation-level forecasts toward task-level, continuously updated indicators linked to real-world adoption measures (firm purchases, API usage, procurement).
Recommendation in the paper motivated by rapid changes in AI capabilities and limitations of static indices; evidence basis is methodological argument and examples of richer adoption measures rather than a quantified evaluation of forecast improvements.
Policy should prioritise flexible reskilling and retraining programs targeted at high-risk tasks and low-skilled workers, informed by task-level exposure maps.
Policy implication recommended by the paper drawing on distributional findings (higher displacement risk for low-skilled tasks) and the availability of task-level exposure indices; evidence basis combines empirical pattern synthesis and normative recommendation rather than an RCT or program evaluation.
Think tanks and international organisations are emphasising scenario planning with differing adoption initial conditions to inform reskilling and labour-market policy.
References to policy and scenario work by organisations named in the paper (TBI, IPPR, IMF, TBI 2024; IPPR 2024; Korinek 2023); evidence basis is published scenario reports and policy papers rather than experimental data.
Practical measures (task selection, oversight, verification, governance) enable responsible deployment of GenAI that balances firm-level goals with individual consultants' skill development.
Recommendations synthesized from interviews with practitioners and the TGAIF framework; presented as practice guidance rather than experimentally tested interventions.
The Task–GenAI Fit (TGAIF) framework maps task characteristics to GenAI capabilities to guide decisions about when and how to use GenAI effectively in consulting processes.
Framework inductively derived from interview data in the study; authors present mapping logic based on task features and reported GenAI capabilities. Evidence is conceptual and qualitative rather than empirically validated.
Generative AI offers efficiency and scaling opportunities in consulting.
Reported repeatedly in practitioner interviews summarized by the authors; qualitative impressions rather than measured productivity gains. No quantitative sample-size or effect-size reported.
A closed interaction loop—MLLM ingesting multimodal inputs (visual, machine feedback, user actions) and outputting structured commands and AR overlays—reduces user cognitive load during machine operation.
System architecture described in the paper plus empirical finding of reduced subjective workload in the CMM case study; supports the claim that the interaction loop contributes to cognitive-load reduction. (Causal attribution to loop structure is inferred rather than directly isolated experimentally.)
An iterative, scenario-refined prompt engineering structure enables the LLM (ChatGPT in this study) to generate task-specific, contextualized guidance that aligns with real-time user actions and machine state.
System design and methods: authors describe developing and refining a prompt structure across multiple machine-operation scenarios and using ChatGPT as the generative engine to produce stepwise instructions and contextual overlay content. Evidence is methodological and qualitative within the paper's development process.
Participants reported lower perceived workload and improved usability when using the AR-MLLM system.
Subjective workload/usability questionnaires were administered in the CMM case study; authors report reduced reported workload under AR-MLLM guidance. (Questionnaire instrument, scales, and sample size not specified in the summary.)
Participants completed assigned CMM tasks faster when using the AR-MLLM system compared to baseline/traditional training.
Task execution time was recorded in the CMM case study; authors report statistically meaningful reductions in completion time with AR-MLLM guidance versus baseline. (Summary does not give numerical effect sizes or sample size.)
The AR-MLLM system achieved high measurement/feature-activity accuracy (participants performed correct measurements under AR-MLLM guidance).
Measurement/feature activity correctness was measured in the CMM case study; authors report high measurement accuracy under the AR-MLLM condition. (Exact rates and sample size not provided in the summary.)
The AR-MLLM system achieved high task-recognition accuracy (the system correctly identified the current task/step).
Measured task recognition accuracy in the CMM case study; authors report 'high' recognition accuracy for the system. (Exact numeric accuracy and sample size not specified in the summary.)
An AR + multimodal LLM (AR-MLLM) training system can substantially improve training and execution in complex machine operations (demonstrated on a Coordinate Measuring Machine).
Case-study experiment in the paper where human participants performed CMM measurement tasks both with and without the AR-MLLM system; metrics collected included task recognition accuracy, measurement activity correctness, task completion time, and subjective workload/usability. (Participant sample size not specified in the provided summary.)
AI methods such as transfer learning, active learning, and Bayesian approaches improve data efficiency and uncertainty quantification in drug discovery and preclinical modeling.
Methodological literature and exemplar studies summarized in the review describing these approaches; heterogeneous examples, no quantitative synthesis.
Clear regulatory alignment (e.g., preparation of credibility plans and qualified digital endpoints) reduces regulatory uncertainty, de-risks investment, and raises adoption rates of AI tools.
Policy and regulatory framework analysis in the review; references to regulatory guidance and qualification processes (narrative, forward-looking).
Economic value from AI adoption concentrates with data-rich firms and platforms that own large, high-quality datasets and validation pipelines.
Economic analysis and theoretical arguments in the paper (narrative), supported by observed market patterns cited in the literature; no formal empirical valuation provided.
Adopting equity-by-design (including diverse, non‑European datasets and subgroup evaluation) reduces model bias and improves global generalizability of AI models.
Recommendations and examples in the review; draws on literature documenting subgroup performance differences and bias remediation strategies (narrative evidence).