Evidence (4793 claims)
Adoption
5539 claims
Productivity
4793 claims
Governance
4333 claims
Human-AI Collaboration
3326 claims
Labor Markets
2657 claims
Innovation
2510 claims
Org Design
2469 claims
Skills & Training
2017 claims
Inequality
1378 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 402 | 112 | 67 | 480 | 1076 |
| Governance & Regulation | 402 | 192 | 122 | 62 | 790 |
| Research Productivity | 249 | 98 | 34 | 311 | 697 |
| Organizational Efficiency | 395 | 95 | 70 | 40 | 603 |
| Technology Adoption Rate | 321 | 126 | 73 | 39 | 564 |
| Firm Productivity | 306 | 39 | 70 | 12 | 432 |
| Output Quality | 256 | 66 | 25 | 28 | 375 |
| AI Safety & Ethics | 116 | 177 | 44 | 24 | 363 |
| Market Structure | 107 | 128 | 85 | 14 | 339 |
| Decision Quality | 177 | 76 | 38 | 20 | 315 |
| Fiscal & Macroeconomic | 89 | 58 | 33 | 22 | 209 |
| Employment Level | 77 | 34 | 80 | 9 | 202 |
| Skill Acquisition | 92 | 33 | 40 | 9 | 174 |
| Innovation Output | 120 | 12 | 23 | 12 | 168 |
| Firm Revenue | 98 | 34 | 22 | — | 154 |
| Consumer Welfare | 73 | 31 | 37 | 7 | 148 |
| Task Allocation | 84 | 16 | 33 | 7 | 140 |
| Inequality Measures | 25 | 77 | 32 | 5 | 139 |
| Regulatory Compliance | 54 | 63 | 13 | 3 | 133 |
| Error Rate | 44 | 51 | 6 | — | 101 |
| Task Completion Time | 88 | 5 | 4 | 3 | 100 |
| Training Effectiveness | 58 | 12 | 12 | 16 | 99 |
| Worker Satisfaction | 47 | 32 | 11 | 7 | 97 |
| Wages & Compensation | 53 | 15 | 20 | 5 | 93 |
| Team Performance | 47 | 12 | 15 | 7 | 82 |
| Automation Exposure | 24 | 22 | 9 | 6 | 62 |
| Job Displacement | 6 | 38 | 13 | — | 57 |
| Hiring & Recruitment | 41 | 4 | 6 | 3 | 54 |
| Developer Productivity | 34 | 4 | 3 | 1 | 42 |
| Social Protection | 22 | 10 | 6 | 2 | 40 |
| Creative Output | 16 | 7 | 5 | 1 | 29 |
| Labor Share of Income | 12 | 5 | 9 | — | 26 |
| Skill Obsolescence | 3 | 20 | 2 | — | 25 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
Productivity
Remove filter
We present a specialized, self-hosted 8B-parameter model designed for a conversational bot in CriQ, a sister app to Dream11 that answers user queries about cricket statistics.
Stated implementation detail in the paper describing the model architecture and deployment target (CriQ conversational bot). No experimental sample size reported for this statement.
Those extended-model equilibria also show increasing concentration consistent with power-law-like distributions (i.e., winner-take-most / superstar effects).
Theoretical model combining quality heterogeneity and reinforcement dynamics that yields equilibrium distributions with heavy tails; argument and formalization presented in the paper; no empirical testing reported.
Even as the number of producers increases and average attention per producer falls, total output expands (production scales elastically).
Same formal theoretical model (analytical result): production scales elastically in the model despite finite attention; no empirical validation provided.
If you can prove the value and the effort behind API token spending (agent memory), you can resell it.
Normative/operational claim within the paper's proposal; presented as an implication of verifiable provenance and market layering, with no empirical proof or transactional data.
Enabling timely memory transfer reduces repeated exploration.
Argument in the paper asserting that shared/tradable memory decreases redundant exploration; no experimental or observational data provided.
Together, clawgang and meowtrade transform one-shot API token spending into reusable and tradable assets.
High-level systems argument in the paper; no empirical measurements of reuse or tradability presented.
Meowtrade is a market layer for listing, transferring, and governing certified memory artifacts.
Design proposal described in the paper; no pilot deployment, user adoption metrics, or experimental data provided.
Clawgang binds memory to verifiable computational provenance.
System/design claim describing the proposed mechanism (clawgang) in the paper; no implementation results or empirical validation reported.
Agent memory can serve as an economic commodity in the agent economy, if buyers can verify that it is authentic, effort-backed, and produced in a compatible execution context.
Conceptual argument in the paper's proposal; no empirical evaluation, sample size, or experiments reported.
Economic theory can be used to generate structured synthetic data that improves foundation-model predictions when the theory implies observable patterns in the data.
General conclusion drawn from the paper's experimental findings: improvement in model predictions after fine-tuning on theory-derived synthetic data.
Fine-tuning on GARP-consistent synthetic data substantially improves prediction relative to zero-shot Chronos-2 at all forecast horizons we study.
Empirical results comparing fine-tuned Chronos-2 to zero-shot Chronos-2 across multiple forecast horizons on the authors' experimental panel (no numeric metrics or sample sizes given in the excerpt).
The fine-tuned model serves as a rationality-constrained forecasting prior: it learns price-quantity relations from GARP-consistent synthetic histories and then uses those relations to predict the choices of real consumers.
Empirical approach described in paper: model fine-tuned on synthetic GARP-consistent histories and then evaluated on real consumer choice data (supports claim that model transfers learned relations to predicting real choices).
GARP is a simple condition to check that allows us to generate time series from a large class of utilities efficiently.
Methodological argument in the paper: authors use GARP as a constructive condition to generate synthetic time series from many utility functions (no numeric efficiency metrics provided in the excerpt).
Teaching them basic economic logic improves how they predict demand using an experimental panel.
Reported experimental results in the paper: fine-tuning models on synthetic, economics-consistent data and evaluating on an experimental panel of consumer demand (no numeric sample size or metrics provided in the excerpt).
AI adoption and the associated improved governance lead to higher total factor productivity (TFP).
Empirical analysis showing a positive association between firm-level AI application index and measures of total factor productivity in the 2010–2023 Chinese A-share panel.
AI adoption and the associated improved governance lead to a lower cost of debt financing for firms.
Empirical tests linking firm-level AI application and governance improvements to measures of debt financing costs (e.g., interest rates on debt, financing spreads) in the Chinese A-share firm sample.
The governance risk-mitigation effects of AI operate through enhancing external monitoring.
Mechanism analyses showing that AI adoption is associated with measures of stronger external monitoring (e.g., analyst coverage, media scrutiny, regulator activity) in the firm-year panel, linking that channel to reduced misconduct.
The governance risk-mitigation effects of AI operate through strengthening internal control capacity.
Mechanism analyses showing that higher AI application is associated with improved internal control measures (as reported by firms or regulatory/financial-control indicators) in the dataset of Chinese A-share firms.
The governance risk-mitigation effects of AI operate through lowering agency costs.
Mechanism analyses reported by authors linking AI adoption to reductions in measures interpreted as agency costs (e.g., agency-cost proxies, corporate governance metrics) in the same firm-year panel.
AI application significantly reduces the monetary amount of penalties associated with executive misconduct.
Regression analyses on monetary penalty data for Chinese A-share firms (2010–2023) showing a statistically significant negative relationship between firm AI application index and penalty amounts.
AI application significantly reduces the frequency (number) of violations by executives.
Empirical frequency/regression analyses on the firm-year panel of Chinese A-share firms using the AI application index; authors report robust reductions in the number/frequency of violations conditional on AI adoption.
AI application significantly reduces the incidence of executive misconduct.
Empirical analysis on Chinese A-share listed firms (2010–2023) using the constructed firm-level AI application index; reported significant negative association between AI application and whether a firm experiences executive misconduct (incidence).
Using Chinese A-share firms listed in Shanghai and Shenzhen from 2010 to 2023, we construct a firm-level AI application index and examine whether and how AI adoption mitigates executive misconduct.
Authors report building a firm-level AI application index and applying it to Chinese A-share listed firms (Shanghai and Shenzhen) over 2010–2023 to study links between AI adoption and executive misconduct (method: panel analysis using firm-year observations).
Adoption of AI can reduce procurement costs by 15.7%.
Field survey data (n=326) and regression analysis; authors report a 15.7% reduction in procurement costs associated with AI adoption.
Adoption of AI can shorten the procurement decision-making cycle by 21.3%.
Field survey data (n=326) analyzed (authors report a 21.3% reduction in procurement decision-making cycle associated with AI adoption); method described as questionnaire surveys and multiple linear regression.
Supplier AI capability positively drives AI adoption in procurement (β = 0.28, p < 0.01).
Same questionnaire survey (n=326) and multiple linear regression analysis; reported coefficient β=0.28 with p<0.01.
Perceived usefulness positively drives AI adoption in procurement (β = 0.32, p < 0.01).
Questionnaire survey of 326 procurement managers/supply chain managers in SMEs (Yangtze River Delta and Pearl River Delta) analyzed using multiple linear regression; reported coefficient β=0.32 with p<0.01.
The paper provides recommendations for designing strategic indicators to drive adoption, foster innovation, and objectively assess whether digital tools are delivering top-line impact.
Descriptive claim about the content of the perspective article (the authors state they provide these recommendations); the excerpt itself summarizes this contribution.
The shift from expert-driven computer-aided drug design (CADD) to semiautonomous AI necessitates a new framework of impact-oriented KPIs.
Stated by the EFMC2 community authors as a normative conclusion in the perspective piece; based on the characterisation of a technological shift rather than on presented empirical tests in the excerpt.
Harnessing AI's potential requires moving beyond measuring technical model performance (e.g., predictive accuracy) to measuring strategic impact.
Authors argue this as a conceptual requirement for realizing AI's benefits in R&D; presented as a recommendation rather than supported by quantified empirical evidence in the excerpt.
Preliminary analyses suggest that 'AI-native' companies may be outpacing traditional peers.
Explicitly stated in the paper as based on preliminary analyses; the excerpt provides no details on the analyses, metrics, or sample sizes.
The broad introduction of AI into the R&D landscape over the last years holds the promise to lift pharmaceutical R&D out of its productivity problem.
Framed as an expectation/promise in the paper; based on recent broad adoption trends of AI in R&D (no specific empirical evaluation or sample size reported in the excerpt).
The visualization preserved human control.
Reported result from the within-subjects experiment (N=32) indicating that using the visualization did not reduce human control/agency in the negotiation process.
In the same within-subjects experiment (N=32), the visualization improved efficiency.
Within-subjects experiment (N=32) reported in the paper; the authors state the visualization improved efficiency (likely measured as time, number of rounds, or steps to reach agreement).
In a within-subjects experiment (N=32), the uncertainty-based visualization improved human outcomes.
Within-subjects user experiment reported in the paper with N=32 participants comparing performance with and without the visualization.
We introduce a novel uncertainty-based visualization driven by Bayesian estimation of agreement probability that shows how the space of mutually acceptable agreements narrows as negotiation progresses, helping users identify promising options.
Design and implementation of a visualization technique described in the paper; the visualization is driven by Bayesian estimation of agreement probability and is presented as a tool to reveal the shrinking feasible agreement space during negotiation.
Generative AI can autonomously produce novel content, including text, images, models, and scenarios.
General technical/descriptive claim stated in the paper's background/introduction; not an empirically tested claim within the provided excerpt.
Generative AI facilitates the synthesis of structured and unstructured information from diverse sources, enabling managers to explore multiple decision pathways, identify potential risks, and optimize strategic choices.
Descriptive/functional claim made in the paper's introduction and conceptual framing; the empirical component (survey + SEM) is described generally but no specific measures or effect sizes for information synthesis or these capabilities are provided in the excerpt.
Generative AI augments human creativity by producing innovative solutions and scenario-planning alternatives that may not emerge through conventional analytical approaches.
Stated in the conceptual/argumentative portion of the paper; may be supported by survey items but no explicit empirical measure or effect size for creativity is provided in the provided text.
Decision quality and strategic agility positively influence organizational performance.
Reported SEM results from the paper linking the constructs (decision quality and strategic agility) to organizational performance using survey data from senior managers and AI adoption specialists; method = SmartPLS.
Generative AI adoption significantly enhances strategic agility.
Same empirical source as above: survey of senior managers/decision-makers/AI adoption specialists; tested via Structural Equation Modeling (SmartPLS) as reported in the paper.
Generative AI adoption significantly enhances decision quality.
Empirical analysis reported in the paper: survey data collected from senior managers, decision-makers, and AI adoption specialists across multiple industries; relationships assessed using Structural Equation Modeling (SmartPLS). No numeric sample size or effect estimate reported in the provided text.
Human-like presentations increased perceived usefulness and agency in certain tasks.
Experimental manipulation of the human-likeness of AI presentation in the study's three tasks; the abstract reports increased perceived usefulness and agency for human-like presentations in some tasks. No sample sizes, task specifics, or effect magnitudes reported in abstract.
A single dissent within a panel reduced pressure to conform.
Experimental manipulation of within-panel consensus (introducing a single dissent) in the study's three tasks; abstract reports that a single dissent lowered conformity pressure. No numerical data provided in abstract.
Accuracy improved for small panels relative to a single AI.
Reported experimental result from the paper's study: participants completed three tasks and received advice from AI panels; panel size was manipulated (small panels vs single AI). The abstract states this accuracy improvement for small panels. (Sample size and exact tasks not reported in abstract.)
Policy should address not only the aftermath of AI labor displacement but also the competitive incentives that drive it.
Normative implication drawn from the model's findings; recommendation in the paper's conclusion based on theoretical results.
Only a Pigouvian automation tax can eliminate the excess automation in the model.
Theoretical welfare analysis demonstrating that a properly set Pigouvian tax that internalizes the demand externality restores the socially optimal level of automation in the model; analytical result, no empirical sample.
The paper proposes a dual-nudge governance architecture leveraging the DHDE to redistribute cross-prefectural flows and reduce economic leakage.
Policy/design proposal presented by the authors as an outcome of the DHDE adaptation and analysis (conceptual/proposed intervention).
The AI-driven decision support system achieves out-of-sample predictive performance of 68% (R^2 = 0.683).
Model performance metric reported in the paper (out-of-sample R^2 value); presumably from held-out validation or cross-validation on the datasets.
The AI-driven decision support system achieves in-sample explanatory power of 81% (R^2 = 0.810).
Model performance metric reported in the paper (in-sample R^2 value); derived from applying the DSS to the supplied datasets.