Evidence (2340 claims)
Adoption
5267 claims
Productivity
4560 claims
Governance
4137 claims
Human-AI Collaboration
3103 claims
Labor Markets
2506 claims
Innovation
2354 claims
Org Design
2340 claims
Skills & Training
1945 claims
Inequality
1322 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 378 | 106 | 59 | 455 | 1007 |
| Governance & Regulation | 379 | 176 | 116 | 58 | 739 |
| Research Productivity | 240 | 96 | 34 | 294 | 668 |
| Organizational Efficiency | 370 | 82 | 63 | 35 | 553 |
| Technology Adoption Rate | 296 | 118 | 66 | 29 | 513 |
| Firm Productivity | 277 | 34 | 68 | 10 | 394 |
| AI Safety & Ethics | 117 | 177 | 44 | 24 | 364 |
| Output Quality | 244 | 61 | 23 | 26 | 354 |
| Market Structure | 107 | 123 | 85 | 14 | 334 |
| Decision Quality | 168 | 74 | 37 | 19 | 301 |
| Fiscal & Macroeconomic | 75 | 52 | 32 | 21 | 187 |
| Employment Level | 70 | 32 | 74 | 8 | 186 |
| Skill Acquisition | 89 | 32 | 39 | 9 | 169 |
| Firm Revenue | 96 | 34 | 22 | — | 152 |
| Innovation Output | 106 | 12 | 21 | 11 | 151 |
| Consumer Welfare | 70 | 30 | 37 | 7 | 144 |
| Regulatory Compliance | 52 | 61 | 13 | 3 | 129 |
| Inequality Measures | 24 | 68 | 31 | 4 | 127 |
| Task Allocation | 75 | 11 | 29 | 6 | 121 |
| Training Effectiveness | 55 | 12 | 12 | 16 | 96 |
| Error Rate | 42 | 48 | 6 | — | 96 |
| Worker Satisfaction | 45 | 32 | 11 | 6 | 94 |
| Task Completion Time | 78 | 5 | 4 | 2 | 89 |
| Wages & Compensation | 46 | 13 | 19 | 5 | 83 |
| Team Performance | 44 | 9 | 15 | 7 | 76 |
| Hiring & Recruitment | 39 | 4 | 6 | 3 | 52 |
| Automation Exposure | 18 | 17 | 9 | 5 | 50 |
| Job Displacement | 5 | 31 | 12 | — | 48 |
| Social Protection | 21 | 10 | 6 | 2 | 39 |
| Developer Productivity | 29 | 3 | 3 | 1 | 36 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
| Skill Obsolescence | 3 | 19 | 2 | — | 24 |
| Creative Output | 15 | 5 | 3 | 1 | 24 |
| Labor Share of Income | 10 | 4 | 9 | — | 23 |
Org Design
Remove filter
Commonly used data types in AI-driven drug discovery include biochemical/binding assay data, protein structural data, HTS results, ADME/Tox and PK datasets, omics/phenotypic readouts, and scientific literature/patents.
Cataloguing of data sources used across studies and company pipelines described in the paper.
AI became widely adopted in pharmaceutical discovery during the 2010s, driven by greater compute, larger datasets, and advances in deep learning.
Historical overview and trend analysis in the paper referencing increased compute availability, growth in public and proprietary datasets, and the rise of deep-learning publications and tools over the 2010s.
Current evidence is illustrative rather than systematic; there is a lack of long-run, quantitative measures of AI’s effect on late-stage clinical outcomes in the literature reviewed.
Explicit methodological statement in the paper: study is an expert/opinion synthesis and narrative review with no new causal econometric estimates or primary experimental data.
The paper identifies three core mechanisms underlying calibrated trust and complementarity: (1) calibrated trust balancing reliance and oversight, (2) complementarity–trust interaction for optimal performance, and (3) dynamic feedback loops producing reinforcing learning cycles.
Explicit identification of mechanisms claimed in the paper's synthesis; this is a descriptive claim about the paper's content rather than an empirical finding—no sample or empirical test reported in the abstract.
The authors surveyed workers and developers on a representative sample of 171 tasks and used language models (LMs) to scale ratings to 10,131 computer-assisted tasks across all U.S. occupations.
Study methodology reported in the paper: surveys of 'workers and developers' on 171 tasks, plus LM-based scaling to 10,131 tasks (coverage claims across U.S. occupations).
The study uses a game-theoretic model involving a foundation model provider and two competing downstream firms to analyze how policy interventions affect consumer surplus in the AI supply chain.
Methodological description in the paper: a formal game-theoretic model with one upstream provider and two downstream competing firms; equilibrium analysis and comparative statics are performed on model outcomes (prices, qualities, profits, consumer surplus).
Foi realizada etnografia organizacional orientada ao SCF, com roteiro e triangulação de evidências.
Método qualitativo divulgado no resumo: etnografia organizacional com roteiro e triangulação; o resumo não fornece número de organizações, duração ou amostragem.
Foi construído e validado um instrumento psicométrico (escala SCF-30) e calculado um índice 0–100, com modelagem por Equações Estruturais (SEM) e testes de confiabilidade/validade.
Descrição metodológica explícita no resumo: construção e validação da escala SCF-30, uso de SEM e testes de confiabilidade e validade. O resumo não detalha estatísticas, amostra ou resultados numéricos.
O SCF é operacionalizado por três vetores centrais: Percepção de Complexidade (PC), Aversão ao Risco Institucional (AR) e Inércia Cultural (IC).
Estrutura conceitual e operacional apresentada no artigo; especificação explícita dos três vetores como componentes do construto SCF.
Degree, betweenness, and eigenvector centrality metrics were used to identify structural vulnerabilities and leverage points in the construction supply chain network.
Paper reports calculation of degree, betweenness, and eigenvector centrality to outline vulnerabilities; specific metrics and interpretations are reported (e.g., degree centrality value for brokers).
Thematic coding translated reported interactions into nodes and edges of a complex network and grouped challenges into thematic categories.
Methods described: thematic coding applied to interview data to create network structure and to generate challenge categories (six main categories, 16 open codes reported).
This study combines empirical, semi-structured interviews with social network analytics to map construction supply chain relationships and vulnerabilities.
Methods reported in the paper: use of semi-structured interviews plus social network analysis (thematic coding to create nodes/edges, calculation of network metrics). Sample size not specified in the abstract.
Extensive experiments were conducted using both synthetic and real hospital datasets to evaluate the framework.
Statement in the paper indicating experiments on synthetic and real datasets; exact sizes, sources, and composition of these datasets are not provided in the excerpt.
Coordination is treated as a structural property of the coupled dynamics (agents + incentives + persistent environment) rather than as the solution to a centralized global optimization objective or purely agent-centric learning problem.
Conceptual framing supported by the formal dynamical model and theorems showing properties of the closed-loop dynamics that do not rely on an underlying global objective.
The persistent environment component of the model stores accumulated coordination signals, and a distributed incentive field transmits those signals locally to adaptive agents, which update their states in response.
Model construction and definitions in the paper describing (i) an environmental state variable with persistent dynamics that accumulates signals, (ii) a spatially/distributed incentive field mapping environmental memory to local agent inputs, and (iii) adaptive update rules for agents.
The paper formalizes agents, incentives, and the environment as a recursively closed feedback architecture (i.e., a coupled dynamical system in which agents adapt to incentive signals that themselves depend on a persistent environmental memory produced by agent actions).
Mathematical model and definitions presented in the paper (formal system specification of agent states, incentive field, and persistent environment; no empirical data).
In a field experiment on the DiagnosUs medical crowdsourcing platform, the authors held the true prevalence in the unlabeled stream fixed at 20% (blasts) while varying the prevalence of positives in the gold-standard feedback stream (20% vs. 50%) and the response interface (binary labels vs. elicited probabilities).
Field experiment conducted on the DiagnosUs platform with experimental manipulations: (i) true prevalence in unlabeled stream fixed at 20% blasts, (ii) feedback-stream prevalence manipulated to 20% vs 50%, (iii) response interface manipulated between binary labels and elicited probabilities. (Sample size and number of workers not specified in the provided excerpt.)
The framework was evaluated on 2,847 queries across 15 task categories.
Paper reports an evaluation dataset consisting of 2,847 queries spanning 15 task categories; used as the sample for reported empirical results.
Non-text processing paths use SLM-assisted modality decomposition.
Paper reports that non-text queries are decomposed using SLM-assisted modality decomposition; described as the non-text routing approach in the framework.
For text-only queries, the framework uses learned routing via RouteLLM.
Paper states text-only routing is handled by a learned model named RouteLLM; presented as part of the system architecture.
A central Supervisor dynamically decomposes user queries, delegates subtasks to modality-appropriate tools (e.g., object detection, OCR, speech transcription), and synthesizes results through adaptive routing strategies rather than predetermined decision trees.
Methodological description in the paper of a Supervisor component that performs dynamic decomposition, delegation to modality-appropriate tools (examples given), and adaptive routing; supported by the framework's implementation details.
We present an agentic AI framework for autonomous multimodal query processing that coordinates specialized tools across text, image, audio, video, and document modalities.
Paper describes the framework design and components (Supervisor, modality-specific tools) and states support for text, image, audio, video, and document modalities; no external benchmark cited for this capability beyond the paper's own implementation.
The essay reviews seven books from the past dozen years by social scientists examining the economic impact of artificial intelligence (AI).
Qualitative book-review performed by the author; sample size explicitly stated as seven books published within the last ~12 years; method = synthesis/assessment of those seven books.
The study is limited by the scope of available industry data and the generalisability of case study findings.
Explicit limitation reported in the paper summary stating constraints related to industry data availability and generalisability of case studies.
The research adopts a mixed-method approach, combining theoretical analysis with empirical insights, and uses data gathered from the 'AI-driven transformation' Scopus database.
Explicit methodological statement in the paper summary: mixed-method design and Scopus database as the data source. (No further methodological details or sample counts provided in the summary.)
The experimental sample underlying the statistical tests comprised 20 observations (implied by ANOVA degrees of freedom: df between = 1, df within = 18).
Interpretation of the reported one-way ANOVA degrees of freedom (F(1,18) for multiple outcomes) indicating total N = 20 observations.
Field experiments at the Al‐Ra'id Research Station in Baghdad during the 2025 season compared conventional diesel‐based irrigation with AI‐assisted irrigation using soil moisture sensors, IoT controllers, and predictive weather algorithms.
Reported field experiment design in the paper (Al‐Ra'id Research Station, Baghdad, 2025 season) specifying two treatments: conventional diesel irrigation vs AI-assisted irrigation using soil moisture sensors, IoT controllers, and predictive weather algorithms.
By integrating dynamic capabilities theory with a micro foundations perspective, the study proposes a conditional model that reframes the essential challenge from technology adoption to organizational adaptation.
Model/theory construction presented in the paper (conceptual integration). This is a methodological/theoretical claim about the paper's contribution; no empirical validation provided.
This study identifies three types of AI triggers that target routines, cognitive frameworks, and resource allocation.
Proposed taxonomy / typology presented in the paper (theoretical classification). The claim is descriptive of the paper's contribution rather than empirically validated.
The study treats AI-agent populations as a system in which four key variables governing collective behaviour can be independently toggled: nature (innate LLM diversity), nurture (individual reinforcement learning), culture (emergent tribe formation), and resource scarcity.
Study design described in the paper (experimental setup allowing independent manipulation of the four variables: model diversity, individual RL, emergent tribe formation, and resource scarcity).
The study integrates Fuzzy Best Worst Method (BWM), PROMETHEE II, and DEMATEL (Fuzzy BWM-PROMETHEE II-DEMATEL) as a three-stage MCDM framework for prioritization and causal analysis of barriers.
Methodology explicitly described in paper: literature survey + expert knowledge feeding into integrated Fuzzy BWM, PROMETHEE II, and Fuzzy DEMATEL analyses.
This study investigates the barriers to the adoption of Industry 4.0 (I4.0) in the Thai automotive industry to inform firms and policymakers.
Stated research aim in paper; approach based on literature survey and expert knowledge; three-stage multi-criteria decision-making (MCDM) model used. (Sample size of experts / respondents not specified in the provided text.)
The paper's findings are based on a combination of literature review, data analysis, and an empirical study involving HR professionals.
Methodological description given in the paper's summary (no further methodological details, sample size, instruments, or statistical methods provided in the summary).
The study draws extensively on contemporary literature in sustainable supply chain management, healthcare procurement, and ESG governance.
Methodological claim about the paper's research approach: literature review/synthesis across the cited domains (bibliographic evidence within the paper).
A complete evaluation methodology is specified, including baselines and an ablation design.
Paper claims to specify evaluation methodology with baselines and ablation; details presumably in the methods section.
The paper formalizes two testable hypotheses on security coverage and latency overhead.
Explicit statement in the paper that two testable hypotheses are formalized (security coverage and latency overhead); no experimental results shown in the abstract.
We conducted preregistered experiments in two tasks (a sentiment-analysis task and a geography-guessing task) to study whether user characteristics influence the effectiveness of AI explanations.
Preregistered experimental studies described in the paper; two distinct tasks (sentiment-analysis and geography-guessing). (Sample sizes and additional procedural details are not provided in the excerpt.)
The framework is depicted across organization areas with primary focus on strategic management and workforce decision-making and secondary focus on finance, operations, and marketing.
Descriptive claim based on the conceptual framework and its mapping to organizational domains within the paper. No empirical application or case studies reported.
This paper outlines a Human–AI Collaborative Decision Analytics Framework integrating five overlapping layers: data, AI analytics, business analytics interpretation, human judgment, and feedback learning.
Presentation of a conceptual framework developed by the authors (conceptual/modeling contribution). No empirical validation reported.
The results presented in the paper are based on a literature recherche, an analysis of individual tasks across different occupations (conducted within Erasmus+ projects), and discussions with trainers/educators.
Methodological statement from the paper; indicates the types of evidence used. The abstract does not provide numbers for analyzed tasks, the number of occupations, details of Erasmus+ projects, or counts of trainers/educators consulted.
The paper identifies key research gaps and proposes a future research agenda focused on human–AI interaction, organizational governance, and ethical accountability.
Conclusions/recommendations from the conceptual meta-analysis (paper-generated research agenda; no empirical testing reported in abstract).
This study presents a conceptual meta-analysis of interdisciplinary literature on AI-augmented decision-making in organizations.
Methodological statement of the paper (the paper itself is a conceptual meta-analysis); no primary empirical sample reported in the abstract.
A Job Digital Intensity Index (JDII) was constructed to capture how digitally intensive jobs are overall, based on the range of digital tasks performed.
Methodological construction described in the report using ESJS digital task items to form a composite JDII.
Deterministic automated verifiers provide objective pass/fail checks for task success.
Methods section: verifiers are deterministic and automated, enabling objective evaluation of whether an agent's trajectory accomplished the task.
Scale of experiments: seven agent–model configurations and 7,308 execution trajectories were used to compute pass rates and deltas.
Reported experimental scale in Methods: 7 agent–model configurations and a total of 7,308 agent execution traces collected and analyzed across tasks/conditions.
Each task was evaluated under three conditions: (1) no Skills, (2) curated (human-authored) Skills, and (3) self-authored (model-generated) Skills.
Experimental protocol described in Methods: three-arm evaluation per task across the SkillsBench benchmark.
SkillsBench benchmark: evaluates 86 tasks spanning 11 domains with deterministic, automated verifiers.
Dataset and benchmark description in the paper: SkillsBench contains 86 tasks across 11 domains and uses deterministic pass/fail verifiers for objective evaluation.
Framing claim: Ideological contests typically produce opposing normative visions (e.g., collectivized command economies vs. market democracies), which makes the development of Western economic theories that portray markets and democracy as dysfunctional puzzling.
Framing and motivation provided in the paper's introduction and background sections; synthesis of conventional expectations about ideological contest outcomes.
The paper uses a qualitative case‑study approach (archival and textual analysis, contextualization, interpretive synthesis) rather than attempting exhaustive quantitative causal identification.
Explicit methods description in the paper: in‑depth historical/institutional examination, archival/textual work, and interpretive synthesis.
Calibration via Method of Simulated Moments (MSM) matches six empirical moments to discipline mechanism magnitudes.
Model calibration procedure reported in the paper: MSM matching six chosen empirical moments that summarize key pre/post-AI patterns (paper states six moments were used).