Evidence (7953 claims)
Adoption
5539 claims
Productivity
4793 claims
Governance
4333 claims
Human-AI Collaboration
3326 claims
Labor Markets
2657 claims
Innovation
2510 claims
Org Design
2469 claims
Skills & Training
2017 claims
Inequality
1378 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 402 | 112 | 67 | 480 | 1076 |
| Governance & Regulation | 402 | 192 | 122 | 62 | 790 |
| Research Productivity | 249 | 98 | 34 | 311 | 697 |
| Organizational Efficiency | 395 | 95 | 70 | 40 | 603 |
| Technology Adoption Rate | 321 | 126 | 73 | 39 | 564 |
| Firm Productivity | 306 | 39 | 70 | 12 | 432 |
| Output Quality | 256 | 66 | 25 | 28 | 375 |
| AI Safety & Ethics | 116 | 177 | 44 | 24 | 363 |
| Market Structure | 107 | 128 | 85 | 14 | 339 |
| Decision Quality | 177 | 76 | 38 | 20 | 315 |
| Fiscal & Macroeconomic | 89 | 58 | 33 | 22 | 209 |
| Employment Level | 77 | 34 | 80 | 9 | 202 |
| Skill Acquisition | 92 | 33 | 40 | 9 | 174 |
| Innovation Output | 120 | 12 | 23 | 12 | 168 |
| Firm Revenue | 98 | 34 | 22 | — | 154 |
| Consumer Welfare | 73 | 31 | 37 | 7 | 148 |
| Task Allocation | 84 | 16 | 33 | 7 | 140 |
| Inequality Measures | 25 | 77 | 32 | 5 | 139 |
| Regulatory Compliance | 54 | 63 | 13 | 3 | 133 |
| Error Rate | 44 | 51 | 6 | — | 101 |
| Task Completion Time | 88 | 5 | 4 | 3 | 100 |
| Training Effectiveness | 58 | 12 | 12 | 16 | 99 |
| Worker Satisfaction | 47 | 32 | 11 | 7 | 97 |
| Wages & Compensation | 53 | 15 | 20 | 5 | 93 |
| Team Performance | 47 | 12 | 15 | 7 | 82 |
| Automation Exposure | 24 | 22 | 9 | 6 | 62 |
| Job Displacement | 6 | 38 | 13 | — | 57 |
| Hiring & Recruitment | 41 | 4 | 6 | 3 | 54 |
| Developer Productivity | 34 | 4 | 3 | 1 | 42 |
| Social Protection | 22 | 10 | 6 | 2 | 40 |
| Creative Output | 16 | 7 | 5 | 1 | 29 |
| Labor Share of Income | 12 | 5 | 9 | — | 26 |
| Skill Obsolescence | 3 | 20 | 2 | — | 25 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
The literature on EEG XAI covers tasks including seizure detection, sleep staging, brain–computer interfaces (BCI), cognitive/emotional state recognition, and diagnostic/supportive tools.
Descriptive review of topical coverage across surveyed papers; specific task categories enumerated in the review.
Limitation: the study analyzes national‑level formal policy texts only and does not measure enforcement, implementation outcomes, or public reactions.
Author‑stated limitations in the paper specifying scope restricted to formal policy documents and absence of empirical enforcement/compliance data.
The paper uses qualitative content analysis, coding documents against the four analytical dimensions to generate a comparative typology of coordination approaches.
Method description: manual qualitative coding of the 36 documents into the specified dimensions, producing the typology distinguishing Chinese and U.S. approaches.
The study's empirical basis comprises 36 national‑level policy documents (18 from China; 18 from the United States) focused on scientific data governance.
Author‑reported dataset and sampling description in the Data & Methods section.
The comparative analysis is organized across four dimensions: coordination objectives, institutional actors, governance mechanisms, and stakeholder legitimacy.
Methodological design reported in the paper; documents were coded against these four analytic categories.
Child-specific surveillance across human, animal, and environmental domains is sparse, limiting understanding of pediatric One Health risks.
Authors' methodological assessment based on literature search and review; explicit limitation stated that standardized child-focused surveillance data are lacking and heterogeneous across sectors.
The legal arguments create some uncertainty about scope and enforcement timelines; economic actors will respond to expected enforcement probabilities and expected sanctions, so clarity from regulators or courts will shape the ultimate economic effects.
Doctrinal acknowledgement of legal uncertainty combined with standard economic modeling of regulatory expectations; no empirical modeling in the Article.
The paper is primarily legal/policy scholarship rather than an empirical assessment of the prevalence or magnitude of discrimination in EdTech; it does not provide econometric estimates of harm.
Explicit limitation noted in the Article (self‑reported).
The Article's evidence consists of illustrative case law and statutory text rather than empirical datasets; it builds doctrinal chains, hypotheticals, and applications of statutory language to modern procurement and EdTech deployment models.
Explicit description of evidence and limits in the Article (self‑reported).
Methodologically, the paper uses doctrinal legal analysis and policy argumentation — close reading of federal civil‑rights statutes, administrative guidance, and judicial decisions interpreting 'recipient' and 'federal financial assistance.'
Explicit methodological statement in the Article (self‑reported).
The legal argument is grounded in statutory interpretation and precedent about the scope of 'recipient' and how federal financial assistance flows and influence should be understood.
Doctrinal analysis of statutes, administrative guidance, and judicial decisions cited and discussed in the Article.
The authors recommend empirical approaches for future work including randomized controlled trials in labs, before-after adoption studies, and collection of microdata on instrument usage, model versions, and provenance to measure impacts.
Explicit methodological recommendations in the Measurement and empirical research agenda section; these are proposals rather than executed studies.
There is a need for rigorous evaluation metrics and benchmarks for safety, reproducibility, and empirical studies quantifying productivity or scientific impact of LLM-driven instrument control.
Identified research gaps and recommended empirical research agenda described by the authors; these are recommendations rather than empirical findings.
The evidence presented consists mainly of qualitative arguments drawn from documented advances and discussion of prototypes; no controlled experimental evaluation is presented.
Authors' own description in the Data & Methods section about the nature of evidence supporting their perspective.
This paper is a conceptual perspective/review rather than an original empirical study.
Explicit statement in the Data & Methods section that the contribution is a perspective synthesizing literature and illustrative examples with no controlled experimental evaluation.
Modern microscopes are increasingly software-driven and data-intensive, while existing ML tools for microscopy are task-specific and fragmented.
Synthesis of recent literature on optical microscopes, detectors, and task-specific ML for image analysis referenced in the perspective (descriptive claim; no new empirical data collected).
Techno‑economic assessments (TEA) and life‑cycle analyses (LCA) are necessary research tools to compare bio‑routes to incumbent chemical synthesis on cost and emissions, and current literature is incomplete in this regard.
Review notes the presence of some TEA/LCA studies but highlights gaps and heterogeneity in methods and results across case studies; many processes lack published TEA/LCA at commercial scales.
Empirical grounding for behavioral-genetic claims and the Four Shell Model comes from the Agora-12 program dataset consisting of 720 agents producing 24,923 decision points.
Reported dataset and experimental sample: Agora-12 program (n = 720 agents; 24,923 decisions) used in analyses and validations.
Robustness checks include city and year fixed effects and heterogeneous-effect examinations by digital infrastructure level.
Reported robustness analyses in the paper: models controlling for city and time fixed effects and tests of heterogeneity by digital infrastructure purported to support the main findings (sample: 280 cities, 2008–2021).
The study's identification strategy treats the Demonstration Zone designation as a quasi-natural experiment using a staggered, multi-period DID across 280 prefecture-level cities (2008–2021).
Stated research design: multi-period difference-in-differences exploiting variation in timing of designation; sample comprises 280 prefecture-level cities over 2008–2021; results include city and time fixed effects.
The employment increase occurred without a corresponding increase in counts of formal cultural enterprises.
Secondary outcome analysis in the same DID framework on formal enterprise counts in the cultural sector using the 280-city panel (2008–2021); reported null effect on number of formal cultural enterprises.
Dataset composition: 261 publicly traded U.S. financial firms matched to CFPB complaint records, monthly observations covering 2018–2023.
Data description in the paper: CFPB complaint records matched to 261 firms with monthly panel from 2018 through 2023 used in all reported analyses.
The paper does not make strong causal claims; causal interpretation is limited and future work should address endogeneity and reverse causality (e.g., with event studies or instrumental variables).
Authors explicitly note limitations on causal interpretation and recommend methods (event studies, IVs, natural experiments) for future causal identification.
Fixed-effects panel path models are used to control for firm-level heterogeneity and to estimate direct and mediated relationships between complaint features and abnormal returns.
Econometric approach described: panel path models with firm fixed effects (monthly firm–level data for 261 firms, 2018–2023) to parse direct/mediated associations between complaint measures and returns.
Econometric approach relies on cross-country panel regressions and interaction terms to assess direct effects and complementarities; identification is associative (panel variation + controls) rather than claiming causal identification using instruments or natural experiments.
Paper describes use of panel regressions with interaction terms and emphasizes that identification comes from panel variation and covariate controls, without detailing stronger causal identification strategies.
Models control for key macroeconomic covariates (e.g., GDP per capita, trade openness, human capital, institutional quality) to isolate technology effects.
Paper documents inclusion of macro controls in regression models to reduce omitted-variable bias.
Dependent variable is a composite national Sustainable Development Goal (SDG) performance index (aggregate/summary measure).
Paper specifies the dependent variable as an aggregate SDG performance measure used in the panel regressions.
Unit of analysis is country-year observations for G20 members covering 2015–2023.
Paper states sample and scope as a cross-country panel of G20 economies from 2015–2023 (panel dataset). (Up to 20 countries × 9 years = up to 180 country-year observations, depending on coverage).
Analyses were conducted as intent-to-treat comparisons across arms, with hypothesis tests reported (including p-values) and principal stratification used for mechanism decomposition.
Methods statement: intent-to-treat comparisons, reported p-values for score differences, and use of principal stratification for separating total effect into adoption and effectiveness channels in the randomized trial (n = 164).
The primary outcomes analyzed were LLM adoption (use), exam score (grade points), and answer length.
Study’s stated primary outcomes in methods: adoption indicator, exam score on an issue-spotting exam, and answer length (measured). Sample size n = 164.
The study used a randomized controlled design with three arms: no LLM access, optional LLM access, and optional LLM access plus brief training.
Study methods description: randomized assignment of 164 law students to three experimental conditions as listed.
The intervention consisted of roughly a ten-minute training focused on how to use the LLM effectively.
Study description of the intervention in the randomized experiment (three-arm design with one arm receiving ~10-minute targeted training).
Findings are estimated for Chinese cities and require replication in other institutional contexts to assess external validity.
Scope statement in the paper — primary empirical sample limited to 274 Chinese cities; authors note generalizability limits and call for replication elsewhere.
The paper’s AI exposure index — capturing automation and service-sector transformation — is important for robust measurement in empirical work on AI’s macro and environmental effects.
Methodological claim justified by the paper's construction of the index and its use in the main and robustness regressions; robustness checks reported using alternative index specifications.
The paper constructs an AI exposure index that captures both industrial automation (robots) and AI-enabled transformation of service-sector jobs/tasks.
Methodological construction described in the paper combining measures of industrial robot adoption (sectoral push) and AI-driven changes in service-sector job/task content.
The study uses a panel of 274 Chinese cities from 2007–2021 as the primary empirical sample.
Descriptive dataset information reported in the paper — city-level panel covering 274 cities and the years 2007 through 2021.
The paper's empirical approach is primarily qualitative and interpretive: a systematic literature review plus comparative qualitative case studies, using policy documents, public diplomacy examples, development initiatives, technology export and standards behaviour, and secondary empirical studies as evidence.
Methods section of the paper explicitly states the approach and evidence types; sample of four comparative cases (US, China, EU, Russia) is specified.
The paper demonstrates different mixes and institutional practices of smart power in practice by applying the framework to the United States, China, the European Union, and Russia.
Explicit comparative qualitative case studies of four major international actors (sample size: four cases) using policy documents, public diplomacy examples, and development/technology initiatives as illustrative evidence.
Empirical validation of the book’s proposals would require complementary case studies, model documentation, and outcome measurements.
Author/reviewer recommendation in the blurb about methodological limitations and next steps; not an empirical finding.
The book is predominantly conceptual and policy-analytic and uses illustrative case vignettes rather than presenting a single empirical study.
Explicit methodological description in the Data & Methods blurb: synthesis of technical ideas, governance requirements, and illustrative vignettes; no empirical sample or experimental protocol described.
The research program is grounded in 12 years of forensic legal research spanning 2014–2026.
Author-stated research timeline and methodology (2014–2026 forensic legal research).
The protocol is underpinned by a forensic audit of approximately 4,200 specialized texts (legal doctrine, regulation, standards, technical literature).
Stated corpus and audit in the Methods section: ~4,200 texts reviewed as part of the forensic audit.
The protocol systematizes arguments for 16 projected rulings at Mexico’s Supreme Court (SCJN) to anchor the proposed rights and rules in constitutional practice.
Doctrinal projection and constitutional strategy section of the compendium describing 16 projected SCJN rulings (method: legal projection/modeling).
The compendium’s findings and recommendations are based on a forensic audit of approximately 4,200 specialized texts covering doctrine, jurisprudence, regulation and technical literature.
Stated methodological claim in the compendium: forensic corpus audit of ~4,200 texts (sample size reported).
The evidence base is qualitative: the study uses conceptual framework synthesis, comparative analysis of multi-sector implementations, and case examples rather than randomized or large-sample empirical evaluation.
Methods and limitations section of the paper explicitly describing the evidence base and methods (qualitative synthesis, pattern extraction, cross-case lessons).
The paper presents a deployment pattern intended to be adapted by sector and regulatory context rather than a one-size-fits-all blueprint.
Explicit statement in the paper and the described pattern design; based on qualitative pattern extraction and prescriptive guidance.
Partial least squares structural equation modeling (PLS-SEM) was used to test hypothesized direct, mediated, and moderated paths.
Methods/analysis section states PLS-SEM was the statistical approach to estimate paths, mediation, and moderation effects.
The study employed a 2 × 2 between-subjects experimental design manipulating (1) identity disclosure (transparent vs. nondisclosed) and (2) conversational tone (empathetic/personalized vs. generic).
Explicit description of experimental factors and design in the methods (2 × 2 between-subjects).
Stimuli (chatbot dialogues) were standardized and pretested using a large-language-model (LLM) workflow to ensure consistent experimental stimuli across conditions.
Methods section describing stimuli creation: LLM-generated dialogues were produced and pretested to standardize messages across the 2 × 2 conditions.
Methodological claim: combining fixed-effects panel estimation, mediation analysis, and panel threshold models is an effective multi-method approach to (a) estimate average effects, (b) unpack causal channels, and (c) detect nonlinear stage-dependent impacts.
The paper's applied methodology: fixed-effects panel regressions, mediation framework, and panel threshold modeling on the 2012–2022 provincial panel.