The Commonplace
Home Dashboard Papers Evidence Digests 🎲

Evidence (3103 claims)

Adoption
5267 claims
Productivity
4560 claims
Governance
4137 claims
Human-AI Collaboration
3103 claims
Labor Markets
2506 claims
Innovation
2354 claims
Org Design
2340 claims
Skills & Training
1945 claims
Inequality
1322 claims

Evidence Matrix

Claim counts by outcome category and direction of finding.

Outcome Positive Negative Mixed Null Total
Other 378 106 59 455 1007
Governance & Regulation 379 176 116 58 739
Research Productivity 240 96 34 294 668
Organizational Efficiency 370 82 63 35 553
Technology Adoption Rate 296 118 66 29 513
Firm Productivity 277 34 68 10 394
AI Safety & Ethics 117 177 44 24 364
Output Quality 244 61 23 26 354
Market Structure 107 123 85 14 334
Decision Quality 168 74 37 19 301
Fiscal & Macroeconomic 75 52 32 21 187
Employment Level 70 32 74 8 186
Skill Acquisition 89 32 39 9 169
Firm Revenue 96 34 22 152
Innovation Output 106 12 21 11 151
Consumer Welfare 70 30 37 7 144
Regulatory Compliance 52 61 13 3 129
Inequality Measures 24 68 31 4 127
Task Allocation 75 11 29 6 121
Training Effectiveness 55 12 12 16 96
Error Rate 42 48 6 96
Worker Satisfaction 45 32 11 6 94
Task Completion Time 78 5 4 2 89
Wages & Compensation 46 13 19 5 83
Team Performance 44 9 15 7 76
Hiring & Recruitment 39 4 6 3 52
Automation Exposure 18 17 9 5 50
Job Displacement 5 31 12 48
Social Protection 21 10 6 2 39
Developer Productivity 29 3 3 1 36
Worker Turnover 10 12 3 25
Skill Obsolescence 3 19 2 24
Creative Output 15 5 3 1 24
Labor Share of Income 10 4 9 23
Clear
Human Ai Collab Remove filter
At the firm level, cost-effective automation captures approximately 11% of computer-vision-exposed labor compensation.
Calibration and implementation in computer vision; reported firm-level estimate from the framework.
high positive Economics of Human and AI Collaboration: When is Partial Aut... share of computer-vision-exposed labor compensation captured by cost-effective a...
Scale of deployment is a key determinant: AI-as-a-Service and AI agents spread fixed costs across users, sharply expanding economically viable tasks.
Modeling and calibration arguments showing fixed-cost spreading effects increase set of tasks for which automation is cost-effective; qualitative and quantitative comparisons in implementation.
high positive Economics of Human and AI Collaboration: When is Partial Aut... number/coverage of economically viable tasks (adoption potential) as a function ...
Because higher accuracy is disproportionately costly (convex cost), full automation is often not cost-minimizing; partial automation, where firms retain human workers for residual tasks, frequently emerges as the equilibrium.
Theoretical model combined with calibration (scaling laws + task mappings); equilibrium outcomes reported from the framework implementation.
high positive Economics of Human and AI Collaboration: When is Partial Aut... prevalence of partial automation vs full automation as cost-minimizing choices
We model automation intensity as a continuous choice in which firms minimize costs by selecting an AI accuracy level, from no automation through partial human-AI collaboration to full automation.
The paper develops a theoretical framework / model that treats automation intensity as a continuous decision variable; described as the central modeling approach.
high positive Economics of Human and AI Collaboration: When is Partial Aut... degree of automation (accuracy level chosen by firms)
The results (conceptual/model results) support corporate GenAI policies, leadership development programs, and HR assessment of leader readiness for GenAI-enabled delegation and communication.
Practical implications and recommendations section arguing policy and HR applications based on the conceptual model.
high positive LEADER EMOTIONAL INTELLIGENCE IN THE GENERATIVE AI ERA: “HUM... policy and HR adoption/application
The article introduces an EI-driven trust-calibration framework as an explanatory mechanism showing when generative AI improves leadership effectiveness and when it amplifies managerial errors.
Novel theoretical framework developed in the paper synthesizing EI, trust calibration, and psychological safety to explain boundary conditions of AI in leadership.
high positive LEADER EMOTIONAL INTELLIGENCE IN THE GENERATIVE AI ERA: “HUM... leadership effectiveness (and amplification of managerial errors)
The paper provides an operationalization toolkit including measures: GenAI use intensity; delegation quality indices (clarity, boundaries, success criteria); communication quality indices (empathy, tone, transparency); psychological safety markers; and behavioral trust-calibration measures.
Operationalization section in the paper listing suggested indices and markers for empirical measurement.
high positive LEADER EMOTIONAL INTELLIGENCE IN THE GENERATIVE AI ERA: “HUM... measurement constructs for empirical studies (e.g., GenAI use intensity, delegat...
As a follow-up validation path, the paper proposes a two-wave time-lag design and 180° assessment (leader + subordinates) to reduce common-method bias.
Methodological proposal in the paper describing longitudinal and multi-rater validation approaches.
high positive LEADER EMOTIONAL INTELLIGENCE IN THE GENERATIVE AI ERA: “HUM... robustness/validity of empirical findings (reduction of common-method bias)
The paper proposes a 'Package B' rapid empirical design: a randomized online experiment manipulating access to generative AI in core managerial tasks (decision, delegation, team communication), combined with EI measurement and trust-calibration indicators.
Methodology section proposing the rapid randomized online experiment design as the primary empirical test.
high positive LEADER EMOTIONAL INTELLIGENCE IN THE GENERATIVE AI ERA: “HUM... experimental test of human–AI leadership effects
Emotional intelligence strengthens the positive impact of generative AI on managerial outcomes when trust is properly calibrated and psychological safety is maintained.
Conceptual model and integrative argument combining EI, trust-calibration, and psychological safety; supported by proposed empirical test design.
high positive LEADER EMOTIONAL INTELLIGENCE IN THE GENERATIVE AI ERA: “HUM... managerial outcomes (e.g., decision quality)
The paper conceptualizes human–AI leadership as an integrated managerial competence.
Conceptual modeling presented in the paper integrating EI theory, psychological safety, and trust calibration (theoretical synthesis).
high positive LEADER EMOTIONAL INTELLIGENCE IN THE GENERATIVE AI ERA: “HUM... human–AI leadership competence (integrated managerial competence)
In the user study, AI-expanded 5W3H prompts increase user satisfaction from 3.16 to 4.04.
Reported pre/post or baseline vs AI-expanded satisfaction scores in the N=50 user study with numeric scores 3.16 and 4.04.
high positive Structured Intent as a Protocol-Like Communication Layer: Cr... user satisfaction (rating scale)
In the user study, AI-expanded 5W3H prompts reduce interaction rounds by 60 percent.
Reported comparison in the N=50 user study between baseline interaction rounds and rounds after AI-assisted 5W3H expansion; percentage reduction reported as 60%.
high positive Structured Intent as a Protocol-Like Communication Layer: Cr... interaction rounds (number of back-and-forth interactions to reach goal)
A weak-model compensation pattern was observed: the lowest-baseline model (Gemini) shows a much larger D-A gain (+1.006) than the strongest model (Claude, +0.217).
Model-level comparison of D-A gain (difference between structured and unstructured conditions) across three models (Claude, GPT-4o, Gemini) on the evaluated outputs; reported gains for Gemini and Claude.
high positive Structured Intent as a Protocol-Like Communication Layer: Cr... D-A gain (improvement in goal-alignment score from structured prompting)
The strongest structured conditions reduce cross-language sigma from 0.470 to about 0.020.
Reported numeric comparison of sigma (variance) between unstructured baseline and strongest structured prompting conditions across evaluated outputs.
high positive Structured Intent as a Protocol-Like Communication Layer: Cr... cross-language sigma (standard deviation of scores across languages)
Structured prompting substantially reduces cross-language score variance relative to unstructured baselines.
Empirical comparison across 3,240 outputs evaluated by DeepSeek-V3, comparing structured vs. unstructured prompting across three languages.
high positive Structured Intent as a Protocol-Like Communication Layer: Cr... cross-language score variance (sigma)
Prior work showed that PPS (Prompt Protocol Specification), a 5W3H-based structured intent framework, improves goal alignment in Chinese and generalizes to English and Japanese.
Statement referring to prior work (not new experiments in this paper); no sample size or methods provided in this text excerpt.
high positive Structured Intent as a Protocol-Like Communication Layer: Cr... goal alignment (language generalization)
Large language model (LLM) use can improve observable output and short-term task performance.
Paper synthesizes empirical findings from human–AI interaction studies, learning-research experiments, and model-evaluation work indicating improved produced outputs and short-term task performance when humans use LLMs; no single pooled sample size or unified effect estimate is reported in the paper.
high positive Beyond the Steeper Curve: AI-Mediated Metacognitive Decoupli... observable output quality and short-term task performance
Frontier models (Claude Haiku 4.5, GPT-5-chat, GPT-5-mini) achieve statistically indistinguishable semantic closeness scores above 4.6 out of 5.0.
Reported semantic closeness scores from the LLM-as-Judge evaluation on the 15-proposal dataset; the paper states frontier models scored above 4.6/5.0 and were statistically indistinguishable from each other.
high positive Can Commercial LLMs Be Parliamentary Political Companions? C... semantic closeness score (LLM-as-Judge)
Autor et al. (2024) show that the majority of current employment is in job specialties that did not exist in 1940, with new task creation driven by augmentation-type innovations.
Citation reported in the paper summarizing Autor et al. (2024); no sample size provided in excerpt.
high positive NBER WORKING PAPER SERIES share of employment in new job specialties (post-1940) and driver of new task cr...
Firms may not sufficiently account for non-monetary aspects of technological progress (well-being, safety, quality of work); a planner would include such considerations in steering technological progress.
Normative conclusion based on theoretical analysis comparing firm objective functions (profits) vs social planner objectives (including non-monetary utility).
high positive NBER WORKING PAPER SERIES attention to non-monetary aspects / inclusion in technological steering
The planner can raise social welfare by focusing technological progress on making goods cheaper that are disproportionately consumed by relatively poorer agents, thereby raising their real income.
Extension of the baseline model to multiple goods showing distributional gains via composition of price changes (real income channel).
high positive NBER WORKING PAPER SERIES real income of poorer agents / social welfare
When capital and labor are gross complements, a planner concerned with workers' welfare would favor capital-augmenting innovations to raise wages.
Analytical result from the model analyzing factor-augmenting technological progress and complementarity between capital and labor.
high positive NBER WORKING PAPER SERIES wages
A planner with sufficient welfare weight on workers will impose positive robot taxes, with the tax rate increasing in the planner's concern for workers' welfare.
Application of the baseline model to robot taxation; analytical derivation of optimal robot tax under planner preferences.
high positive NBER WORKING PAPER SERIES optimal robot tax rate
As labor's economic value diminishes, steering progress focuses increasingly on enhancing human well-being (non-monetary aspects) rather than labor productivity.
Theoretical discussion and model results in the paper showing planner's shifting objective when labor is devalued.
high positive NBER WORKING PAPER SERIES focus of technological steering (monetary productivity vs non-monetary well-bein...
The welfare benefits of steering technology are greater the less efficient social safety nets are.
Analytical result from the paper's theoretical model comparing a planner who can/cannot perform transfers and evaluating steering as second-best when redistribution is costly.
high positive NBER WORKING PAPER SERIES welfare benefits of steering technological progress
These household-level non-market productivity gains (ChatGPT making productive online tasks more efficient and freeing time for leisure) are economically large and likely constitute a substantial share of the overall economic impact of generative AI.
Combination of empirical IV estimates showing leisure increases and productivity-unchanged productive time, plus model-implied efficiency gains; authors' interpretation and welfare discussion in paper.
high positive https://arxiv.org/pdf/2603.03144 household non-market productivity and welfare (implied aggregate economic impact...
Mapping the empirical time-reallocation into a quantitative household time-allocation model implies generative AI approximately doubles the efficiency of productive online tasks for adopters; preferred calibration implies efficiency gains of 76%–176%.
Quantitative time-allocation model adapted from Aguiar et al. (2021); model uses empirical IV estimates for time reallocation and Engel curve elasticities estimated via IV (local precipitation shocks). Authors report implied efficiency gains of 76%–176% and state 'approximately doubles' efficiency.
high positive https://arxiv.org/pdf/2603.03144 efficiency (productivity) of productive digital tasks
Households predominantly utilize ChatGPT in the context of productive online activities (education, job search, informational research) rather than during leisure browsing, as inferred from the browsing context around ChatGPT use.
High-frequency analysis comparing 30-minute browsing intervals around ChatGPT visits to intervals of demographically similar non-users; LLM-based inference of website purpose; observed co-occurrence with productive-site categories.
high positive https://arxiv.org/pdf/2603.03144 context/purpose of ChatGPT use (productive vs leisure)
ChatGPT adoption increases the leisure share of browsing duration by about 30 percentage points.
IV long-difference estimates from Comscore browsing data with LLM-based site classification; authors report a ~30 percentage point increase in leisure share after adoption.
high positive https://arxiv.org/pdf/2603.03144 leisure share of total browsing duration
In long-difference IV estimates, ChatGPT adoption raises total leisure browsing time by roughly 150 log points.
IV long-difference estimates using pre-ChatGPT exposure as instrument; reported effect described as 'roughly 150 log points' increase in total leisure browsing time.
high positive https://arxiv.org/pdf/2603.03144 total leisure browsing time (log change)
A household's pre-ChatGPT ex-ante exposure (based on 2021 browsing composition) strongly predicts subsequent ChatGPT adoption: a 1 SD higher exposure predicts a 2.5 percentage point higher rate of having used ChatGPT by December 2024.
Constructed 'exposure' measure by aggregating site-level overlap with chatbot capabilities over household 2021 browsing; predictive regression (household-level) linking 1 SD change in exposure to 2.5pp higher adoption by Dec 2024 (statistic reported in paper).
high positive https://arxiv.org/pdf/2603.03144 probability / rate of ChatGPT adoption by Dec 2024
ChatGPT adoption among private households has been rapid following release, but adoption is far from uniform.
Descriptive adoption patterns measured from Comscore browsing data over time (pre- and post-Nov 30, 2022) on the household panel (2021–2024); time-series of observed ChatGPT site visits and adoption rates.
high positive https://arxiv.org/pdf/2603.03144 ChatGPT adoption rate over time
Task-level analyses show that activities expanded in AI-enabled projects—particularly ideation and experimentation—are increasingly compatible with large language model capabilities, suggesting potential for future productivity gains as these technologies mature.
Task-level classification mapping tasks described in proposals to LLM-relevant capabilities using LLM-based classification; finding that tasks expanded in AI-enabled projects cluster on ideation and experimentation, which align with current LLM strengths.
high positive Artificial Intelligence in Science: Returns, Reallocation, a... frequency/expansion of specific task categories (ideation, experimentation) and ...
AI-enabled projects undertake a broader set of tasks.
Task-level analysis of proposal descriptions (task inventories) classifying tasks via keyword extraction and LLMs, showing AI-enabled proposals list a wider variety of activities than non-AI proposals.
high positive Artificial Intelligence in Science: Returns, Reallocation, a... breadth/variety of tasks undertaken in projects
AI-enabled projects involve larger teams.
Comparison of team structure in proposals (team size) between AI-enabled and non-AI projects using the same comprehensive proposal dataset and LLM-based classification of AI presence.
high positive Artificial Intelligence in Science: Returns, Reallocation, a... team size / team structure
AI-enabled projects reallocate resources toward human capital (i.e., shift budget allocations toward labor / human capital).
Analysis of detailed budget allocations in the proposal dataset, comparing projects identified as AI-enabled versus non-AI projects using keyword extraction and LLM classification to identify AI presence and role.
high positive Artificial Intelligence in Science: Returns, Reallocation, a... budget allocation share toward human capital (labor share)
In the short run, AI adoption is associated with modest improvements in scientific outcomes concentrated in the upper tail.
Observational analysis linking identified AI presence in a comprehensive dataset of research proposals (funded and unfunded) to subsequent publication outcomes; AI presence identified via keyword extraction combined with large language model (LLM) classification; publication outcomes measured after proposal submission.
high positive Artificial Intelligence in Science: Returns, Reallocation, a... subsequent publication outcomes (scientific outcomes)
The experience-centered learning mechanism proactively recalls rewarded trajectories at inference time.
Specific technical/design claim about Synergy's learning mechanism; asserted in paper as a mechanism feature rather than demonstrated with quantified results in the provided text.
high positive Synergy: A Next-Generation General-Purpose Agent for Open Ag... agent learning behavior (recall of rewarded trajectories during inference)
Synergy grounds collaboration in session-native orchestration, repository-backed workspaces, and social communication; identity in typed memory, notes, agenda, skills, and persistent social relationships; and evolution in an experience-centered learning mechanism that proactively recalls rewarded trajectories at inference time.
Detailed design claims describing Synergy's mechanisms and intended grounding for collaboration, identity, and evolution; presented as architectural description, no experimental evaluation provided in the excerpt.
high positive Synergy: A Next-Generation General-Purpose Agent for Open Ag... architectural features supporting collaboration, identity, and learning (session...
We present Synergy, a general-purpose agent architecture and runtime harness for persistent, collaborative, and evolving agents on Open Agentic Web.
Paper's contribution statement indicating the authors propose an architecture named Synergy; this is a systems/design claim rather than an empirical result in the provided text.
high positive Synergy: A Next-Generation General-Purpose Agent for Open Ag... existence of an architecture (Synergy) designed for persistent, collaborative, e...
The next generation of agents must become Agentic Citizens, defined by three requirements: Agentic-Web-Native Collaboration, participation in open collaboration networks rather than only closed internal orchestration; Agent Identity and Personhood, continuity as a social entity rather than a resettable function call; and Lifelong Evolution, improvement across task performance, communication, and collaboration over time.
Normative/design prescription from the authors; conceptual argument for three requirements rather than empirical validation.
high positive Synergy: A Next-Generation General-Purpose Agent for Open Ag... agent design properties (collaboration, identity/personhood, lifelong evolution)
As the internet prepares to host billions of such entities, it is shifting toward what we call Open Agentic Web, a decentralized digital ecosystem in which agents from different users, organizations, and runtimes can discover one another, negotiate task boundaries, and delegate work across open technical and social surfaces at scale.
Conceptual claim / framing by the authors describing a projected/ongoing shift; no empirical measurement of 'billions' or of ecosystem properties provided in the excerpt.
high positive Synergy: A Next-Generation General-Purpose Agent for Open Ag... emergence of an Open Agentic Web (decentralized agent ecosystem, discovery, nego...
Embodied agents are spreading across smartphones, vehicles, and robots.
Author observation/claim in the paper's opening; no empirical study, metrics, or examples quantified in the provided text.
high positive Synergy: A Next-Generation General-Purpose Agent for Open Ag... deployment/penetration of embodied agents across device categories
Open-source frameworks such as OpenClaw are putting personal agents in the hands of millions.
Author assertion naming OpenClaw and a numeric adoption claim; no supporting empirical data or citation contained in the provided text.
high positive Synergy: A Next-Generation General-Purpose Agent for Open Ag... adoption of personal agents (number of users)
AI agents are rapidly expanding in both capability and population: they now write code, operate computers across platforms, manage cloud infrastructure, and make purchasing decisions.
Author assertion in paper's introduction / high-level observation; no empirical study, dataset, or experiment reported in the provided text.
high positive Synergy: A Next-Generation General-Purpose Agent for Open Ag... agents' capabilities and population growth (write code, operate computers, manag...
Education and workforce development should shift focus from rote knowledge accumulation to cultivating skills in human-AI collaboration, creative problem-solving, and the design of novel economic domains.
Normative policy recommendation derived from the paper's framework and analysis of anticipated labor market changes (no empirical evaluation or trial data reported in the abstract).
high positive AI Civilization and the Transformation of Work educational focus / skill composition
Human-AI co-evolution will significantly increase individual productivity and open new frontiers of economic activity.
Projected outcome based on combined analysis of AI capabilities, historical patterns, and platform growth; the abstract does not report empirical measurement or sample sizes for this projection.
high positive AI Civilization and the Transformation of Work individual productivity and emergence of new economic activities
AI-driven productivity augmentation dramatically lowers the barriers to creating economic value, enabling the decentralized generation of employment.
Argument supported by paper's analysis of contemporary labor market dynamics and the growth of digital platforms; no quantified empirical estimates or sample sizes provided in the abstract.
high positive AI Civilization and the Transformation of Work barriers to entry for value creation / individual productivity
The transition to an AI-civilization will fundamentally restructure the mechanisms of employment creation from a centralized model (few organizations creating jobs for the many) to a decentralized ecosystem where individuals are empowered to generate their own employment opportunities.
Central thesis of the paper, motivated by theoretical argumentation and synthesis of contemporary data on labor markets and digital platforms (no empirical test or sample sizes specified in the abstract).
high positive AI Civilization and the Transformation of Work structure/mechanism of employment creation (centralized vs decentralized)