Evidence (2340 claims)
Adoption
5267 claims
Productivity
4560 claims
Governance
4137 claims
Human-AI Collaboration
3103 claims
Labor Markets
2506 claims
Innovation
2354 claims
Org Design
2340 claims
Skills & Training
1945 claims
Inequality
1322 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 378 | 106 | 59 | 455 | 1007 |
| Governance & Regulation | 379 | 176 | 116 | 58 | 739 |
| Research Productivity | 240 | 96 | 34 | 294 | 668 |
| Organizational Efficiency | 370 | 82 | 63 | 35 | 553 |
| Technology Adoption Rate | 296 | 118 | 66 | 29 | 513 |
| Firm Productivity | 277 | 34 | 68 | 10 | 394 |
| AI Safety & Ethics | 117 | 177 | 44 | 24 | 364 |
| Output Quality | 244 | 61 | 23 | 26 | 354 |
| Market Structure | 107 | 123 | 85 | 14 | 334 |
| Decision Quality | 168 | 74 | 37 | 19 | 301 |
| Fiscal & Macroeconomic | 75 | 52 | 32 | 21 | 187 |
| Employment Level | 70 | 32 | 74 | 8 | 186 |
| Skill Acquisition | 89 | 32 | 39 | 9 | 169 |
| Firm Revenue | 96 | 34 | 22 | — | 152 |
| Innovation Output | 106 | 12 | 21 | 11 | 151 |
| Consumer Welfare | 70 | 30 | 37 | 7 | 144 |
| Regulatory Compliance | 52 | 61 | 13 | 3 | 129 |
| Inequality Measures | 24 | 68 | 31 | 4 | 127 |
| Task Allocation | 75 | 11 | 29 | 6 | 121 |
| Training Effectiveness | 55 | 12 | 12 | 16 | 96 |
| Error Rate | 42 | 48 | 6 | — | 96 |
| Worker Satisfaction | 45 | 32 | 11 | 6 | 94 |
| Task Completion Time | 78 | 5 | 4 | 2 | 89 |
| Wages & Compensation | 46 | 13 | 19 | 5 | 83 |
| Team Performance | 44 | 9 | 15 | 7 | 76 |
| Hiring & Recruitment | 39 | 4 | 6 | 3 | 52 |
| Automation Exposure | 18 | 17 | 9 | 5 | 50 |
| Job Displacement | 5 | 31 | 12 | — | 48 |
| Social Protection | 21 | 10 | 6 | 2 | 39 |
| Developer Productivity | 29 | 3 | 3 | 1 | 36 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
| Skill Obsolescence | 3 | 19 | 2 | — | 24 |
| Creative Output | 15 | 5 | 3 | 1 | 24 |
| Labor Share of Income | 10 | 4 | 9 | — | 23 |
Org Design
Remove filter
LLMs are more likely to complement human tacit skills than to replace explicit rule‑following jobs; value accrues to workers and firms that integrate model outputs with human judgment and tacit expertise.
Labor‑economics style argument and theoretical reasoning; no empirical labor market analysis provided.
Commoditization via rule extraction is limited; firms that can harness and deploy tacit LLM capabilities will retain economic rents.
Theoretical economic argument based on non‑rule‑encodability; no empirical firm‑level data included.
The highest‑value attributes of LLMs may be inherently non‑decomposable into simple, auditable rules, which increases the value of proprietary, black‑box models and strengthens economies of scale and scope for large model providers.
Economic reasoning and theoretical implications drawn from the central thesis; no empirical market analyses provided.
Some LLM capabilities are tacit, practice‑derived, or 'insight'‑like, akin to the Chinese concept of Wu (sudden insight through practiced skill).
Philosophical framing and analogy to the concept of tacit knowledge (Wu); argumentative rather than empirical support.
The economically valuable capabilities of large language models are precisely those that cannot be fully encoded as a complete, human‑readable set of discrete rules.
Formal, conceptual argument (proof by contradiction) plus qualitative historical case analysis comparing expert systems and LLMs; no new empirical datasets or experiments reported.
The proposed algorithm's performance is robust to heterogeneous populations in the synthetic experiments (i.e., it continues to find core alternatives under varying degrees of population heterogeneity).
Empirical robustness checks reported in the experiments where population heterogeneity is varied and performance (core-attainment frequency) is evaluated.
The authors compare their sampling algorithm against classical social-choice rules and LLM-based heuristics and report superior core-attainment frequency for their method.
Experimental comparisons described in the paper between the proposed algorithm and baseline methods (classical social-choice rules, LLM-based heuristics) on the synthetic dataset; results summarized in the experiments section.
On a synthetic text-preference dataset, the proposed algorithm reliably finds alternatives that lie in the proportional veto core.
Empirical experiments reported in the paper using a synthetic dataset of text preferences; evaluation metric reported as frequency (proportion) of runs where the returned alternative is in the proportional veto core.
Standardized runtime governance frameworks could lower per-deployment compliance engineering costs and increase diffusion of agentic systems.
Theoretical argument that standardization reduces transaction/engineering costs; suggested market dynamics; no empirical implementation evidence.
A market will develop for third-party governance tools, auditors, and insurers providing policy evaluators, risk calibration, and certification services.
Economic argument and analogy to existing markets (governance-as-a-service, insurance); no empirical evidence presented.
Adopting this approach shifts required skills and organizational roles away from lengthy parametric modeling toward data engineering, controller integration, and monitoring.
Authors' discussion of practical/organizational implications (qualitative); argument based on removal of model-building step and increased emphasis on data infrastructure and online operations.
DeePC outperforms baseline controllers (e.g., fixed-time and standard adaptive schemes) in the simulated experiments.
Comparative simulation experiments reported in the paper where DeePC-controlled signals achieve superior system-level metrics relative to baseline controllers.
The method was validated on a very large, high-fidelity microscopic closed-loop simulator of Zürich; the paper reports this as the largest such closed-loop urban-traffic simulation in the literature.
Authors' description of the experimental environment: city-scale microscopic simulator of Zürich with controller in the loop; explicit statement in the paper claiming it is the largest closed-loop urban-traffic simulation reported in the literature.
Regularization and the use of measured Hankel/data matrices make the method more robust to measurement noise and limited data.
Method description includes regularization terms in the DeePC optimization and use of Hankel matrices built from measured trajectories; simulation experiments show continued performance under noisy / limited-data conditions.
DeePC handles sparse or limited traffic measurements better than many machine-learning methods.
Claims in the paper supported by experiments and methodological notes: use of Hankel structures and regularization in DeePC to operate with limited/sparse sensing; comparative statements versus generic ML methods (qualitative and simulation evidence).
The DeePC-based approach avoids the expensive, time-consuming model-building step required by model-based control methods.
Methodological argument and demonstration that controller uses historical input–output trajectories directly rather than requiring separate parametric model identification; supported by simulation implementation that bypasses model identification.
Policy instruments that can support shorter workweeks include tax incentives for firms that maintain pay while reducing hours, regulatory transition frameworks, and conditionality on AI subsidies or public procurement tied to job-preservation or reduced hours.
Policy-analytic argument drawing on standard policy toolkits and selected prior examples; no new policy pilot results presented.
Shorter workweeks help sustain consumer purchasing power by reducing aggregate labor supply and thereby distributing automation gains more equitably.
Theoretical labour-supply reasoning plus historical case studies of work-time reductions; argumentual and normative rather than demonstrated with new macroeconomic empirical tests in AI-rich settings.
A gradual, policy-driven reduction in the standard workweek can absorb labor displaced by automation, help maintain employment levels, and preserve wages per hour.
Synthesis of prior empirical findings on work-hour reductions and historical precedents (e.g., six-day to five-day transition); no new randomized or large-scale contemporary trials presented.
Firms use layoffs strategically to signal efficiency and boost short-term stock prices, even when automation is not fully substitutive.
Organizational- and finance-literature synthesis on signaling and market reactions to cost-cutting; historical/case examples referenced rather than new econometric estimates.
Policymakers should prioritize retraining programs, strengthened social protection, and redistributive policies to mitigate automation-induced unemployment and inequality.
Policy recommendation based on the author's synthesis of risks and expert judgment; not based on an empirical intervention study in the paper.
There has been progress in software import substitution, contributing to partial technological sovereignty in Russia.
Use of statistics on software import substitution (authors reference national statistics but do not report detailed numbers or methodology).
Digitalization enables management optimization (improved management processes and decision-making) in Russian enterprises and public administration.
Qualitative analysis of policy documents and expert assessment by the author; no empirical evaluation or quantified effect sizes provided.
Digitalization has produced measurable labor productivity growth in segments of the Russian economy.
Author's interpretation drawing on national statistics and strategic documents; statistical details (period, sectors, sample sizes) not specified in the paper.
Policy/managerial implication: organizational structures and incentives (e.g., TMT diversity, ESOPs) are effective levers to sustain managerial attention to employee welfare and mitigate the negative effects of deep AI penetration on ECSR.
Inference from empirical moderator results (TMT diversity and ESOP interactions) combined with theoretical ABV/dual-agent argument; paper includes managerial and policy recommendations based on these findings.
Employee stock ownership (ESOP) moderates the relationship by flattening and right-shifting the inverted U, aligning employee incentives and preserving employee-focused attention as AI adoption deepens.
Interaction terms between AI (and AI^2) and ESOP presence/level show mitigated negative effects of high AI adoption on ECSR and a later turning point; based on panel regressions with controls and robustness checks on the 2,575-firm sample.
Top management team (TMT) functional diversity moderates the AI–ECSR curve by flattening it and right-shifting the peak, delaying and mitigating negative attention shifts from employees to AI.
Interaction of AI (and AI^2) with a TMT functional diversity measure in panel regressions indicates a less pronounced inverted U and a later turning point for firms with more diverse TMTs; analysis uses the main panel (2,575 firms, 2013–2023) with robustness checks.
At low-to-moderate levels of AI adoption, AI increases managerial attention to employees and raises ECSR (human attention gain mechanism).
Positive slope of the estimated AI–ECSR relationship at lower AI values implied by the significant linear AI term in the quadratic panel model; theoretical framing via an attention-based view (ABV) and dual-agent model; empirical results interpreted as consistent with increased managerial attention and higher ECSR at low-to-moderate AI adoption. (Sample: 2,575 firms, 2013–2023.)
AI-mediated collaboration will create new organizational roles and governance structures, such as AI mediators and verification/oversight roles.
Conceptual discussion of organizational implications and illustrative role examples; no organizational case studies with sample sizes reported.
Autonomous AI agents can automate routine coordination tasks, follow-up, and some task execution, thereby reducing human coordination overhead.
Paper uses conceptual mapping of agent capabilities to coordination/execution functions and provides illustrative case scenarios; no experimental or field data presented.
Multimodal systems (integrating text, speech, images, video) broaden communication channels and thus can improve the range and fidelity of mediated communication.
Conceptual argument and illustrative examples in the paper describing how multimodal integration maps to communication functions; no empirical validation reported.
Multilingual language models reduce language barriers by translating and normalizing meaning across languages.
Conceptual synthesis of capabilities (multilingual LMs) and mapping to coordination function (translation/normalization); supported in paper by illustrative examples rather than empirical testing.
Trust in AI should be conceptualized as a socio-technical, team-level mechanism (trust calibration) that mediates between AI design/enablers and downstream collaboration and performance, rather than an individual-level stable attitude.
Theoretical synthesis combining findings from the thematic analysis of 40 interviews with socio-technical systems theory (STS) and adaptive structuration theory (AST) to propose an initial and revised conceptual model linking enablers → trust-calibration practices → collaboration dynamics → performance.
Five enablers support effective trust calibration: transparency/explainability, clear role definitions, good user experience (UX), supportive cultural norms, and timely system feedback.
Synthesized from recurring themes in the interview data (N=40) where respondents identified these factors as facilitating appropriate reliance on AI in project settings; coded and aggregated through thematic analysis.
Performance and reward structures must be redesigned to value oversight, hypothesis testing, escalation and governance behaviours that mitigate model risk but may not immediately increase output.
Managerial recommendation derived from the framework and organizational reward literature; no empirical evaluation provided.
Firms need new metrics to decompose value created by humans, AI, and their interaction (to distinguish complementarities versus substitution).
Analytic implication derived from the framework and literature on productivity measurement; presented as a recommendation for empirical work rather than tested evidence.
Symbiarchic leadership is a practical, HR‑oriented framework for leading integrated human–AI “cyber teams,” specifying four linked leadership practices that make AI a co‑actor in knowledge work while preserving human judgement, accountability and organizational legitimacy.
Paper's central proposition based on theoretical synthesis of academic literature on human–AI collaboration, hybrid teams and digital‑era leadership plus illustrative practitioner examples; no original empirical data or experiments.
Policies improving data sharing, standardization, and model transparency would increase overall welfare by reducing duplication and improving model performance.
Policy argumentation in the paper drawing on economic theory and examples where shared datasets/standards improved research productivity.
Organizations that tightly integrate AI teams with experimental groups achieve higher productivity.
Case studies and internal metrics cited in the paper showing improved throughput and candidate progression in integrated teams versus siloed approaches.
Value accrues to firms that control high-quality data, integrated platforms, and wet-lab validation—data and experimental capacity are strategic assets.
Market and organizational analysis in the paper citing examples of firms leveraging proprietary data/platforms and wet-lab capabilities to advance candidates more effectively.
AI reduces time and cost in early-stage discovery (discovery-to-candidate), lowering per-candidate screening and design costs.
Reported case studies and cost/time comparisons in the paper showing faster candidate identification and reduced experimental burden in early stages; aggregated industry claims summarized.
Several AI-guided molecules have entered clinical trials and show encouraging early-phase indicators.
Industry reports and trial registries summarized in the paper reporting multiple AI-guided programs reaching Phase I/II; company disclosures and early-phase biomarker or safety readouts referenced.
Firms with superior proprietary data and integration capability gain competitive advantage, increasing firm-level heterogeneity in AI returns.
Narrative analysis of market structure implications and examples; no cross-firm empirical heterogeneity study included.
Returns to complementary investments (data infrastructure, experiment automation, cross-disciplinary teams) increase as AI becomes more central to discovery workflows.
Synthesis of adoption lessons and case examples emphasizing complementary capital; no quantitative ROI estimates provided.
Embedding AI into organizational processes, decision-making, and wet-lab validation is crucial to capturing its value.
Narrative review of adoption and integration lessons from large biopharma experience and illustrative case studies.
Successful AI adoption requires investment in data, talent, and workflows rather than reliance on bolt-on point solutions.
Thematic analysis of adoption-level lessons and industry case examples indicating organizational and infrastructural requirements for realized value.
AI has produced genuine early-stage breakthroughs in drug discovery, accelerating hit identification and early design cycles.
Narrative expert synthesis and thematic analysis of industry experience over the first decade of AI adoption, illustrated by early-case successes and firm-reported accelerations; no new primary experimental data or causal econometric estimates provided.
Dynamic feedback loops create reinforcing organisational learning cycles.
Theoretical assertion from the paper's synthesis indicating learning dynamics as part of the model; described conceptually without empirical quantification in the abstract.
Complementarity–trust interaction determines optimal performance when high capability utilisation combines with appropriate trust levels.
Mechanistic claim from the TCM‑CI derived via systematic review/synthesis of existing studies; no primary experimental or field sample reported in the abstract to validate this interaction effect.
Calibrated trust maximises collective intelligence by balancing appropriate reliance with necessary oversight.
Core mechanism asserted by the paper based on synthesis of prior research in human–AI interaction and trust literature; presented as a conceptual mechanism rather than tested empirically in the abstract.