The Commonplace
Home Dashboard Papers Evidence Digests 🎲

Evidence (3103 claims)

Adoption
5267 claims
Productivity
4560 claims
Governance
4137 claims
Human-AI Collaboration
3103 claims
Labor Markets
2506 claims
Innovation
2354 claims
Org Design
2340 claims
Skills & Training
1945 claims
Inequality
1322 claims

Evidence Matrix

Claim counts by outcome category and direction of finding.

Outcome Positive Negative Mixed Null Total
Other 378 106 59 455 1007
Governance & Regulation 379 176 116 58 739
Research Productivity 240 96 34 294 668
Organizational Efficiency 370 82 63 35 553
Technology Adoption Rate 296 118 66 29 513
Firm Productivity 277 34 68 10 394
AI Safety & Ethics 117 177 44 24 364
Output Quality 244 61 23 26 354
Market Structure 107 123 85 14 334
Decision Quality 168 74 37 19 301
Fiscal & Macroeconomic 75 52 32 21 187
Employment Level 70 32 74 8 186
Skill Acquisition 89 32 39 9 169
Firm Revenue 96 34 22 152
Innovation Output 106 12 21 11 151
Consumer Welfare 70 30 37 7 144
Regulatory Compliance 52 61 13 3 129
Inequality Measures 24 68 31 4 127
Task Allocation 75 11 29 6 121
Training Effectiveness 55 12 12 16 96
Error Rate 42 48 6 96
Worker Satisfaction 45 32 11 6 94
Task Completion Time 78 5 4 2 89
Wages & Compensation 46 13 19 5 83
Team Performance 44 9 15 7 76
Hiring & Recruitment 39 4 6 3 52
Automation Exposure 18 17 9 5 50
Job Displacement 5 31 12 48
Social Protection 21 10 6 2 39
Developer Productivity 29 3 3 1 36
Worker Turnover 10 12 3 25
Skill Obsolescence 3 19 2 24
Creative Output 15 5 3 1 24
Labor Share of Income 10 4 9 23
Clear
Human Ai Collab Remove filter
For stronger causal evidence, recommended empirical methods include difference-in-differences on adopting firms vs. controls, matched samples, and randomized pilots for particular tools, supplemented by qualitative interviews.
Methodological recommendations stated in the paper (not an empirical finding); no implementation/sample reported in the abstract.
high null result Role of Artificial Intelligence in the Accounting Sector validity of causal inference on AI impacts (identification quality)
Actionable research priorities include running larger-scale field trials linking game use to observed land-use and economic outcomes, developing validation protocols for game-backed models against empirical on-farm data, studying heterogeneity of impacts, and designing incentive mechanisms that leverage game-demonstrated profitability co-benefits.
Synthesis-driven recommendations based on identified evidence gaps—specifically the predominance of small-scale/qualitative studies and lack of long-term/causal evidence.
high null result Serious games and decision support tools: Supporting farmer ... Observed land-use change, economic outcomes, validated model performance, hetero...
Rigorous economic evaluation (RCTs, quasi-experiments) is needed to quantify how game-enhanced DSTs affect investment, land-use choices, emissions outcomes, and farm incomes.
Chapter recommendation grounded in observed gaps: the literature lacks sufficiently rigorous causal impact evaluations; current evidence is largely qualitative or observational.
high null result Serious games and decision support tools: Supporting farmer ... Investment decisions, land-use change, emissions (measured GHG outcomes), farm i...
The paper's evidence is policy‑oriented, qualitative and analytical; it does not report causal estimates from new field data and produces testable propositions and an empirical agenda instead.
Explicit methods statement in the paper: structured desk review, corridor process mapping, governance gap analysis; absence of field experiments or causal quantitative analysis.
high null result Training as corridor governance: TVET alignment, skills reco... absence of new causal effect estimates in the study
No new laboratory measurements or datasets are reported in the paper; the approach is methodological and conceptual rather than empirical.
Methods section and explicit statements within the paper noting absence of new data; verifiable by reading the paper.
high null result XChronos and Conscious Transhumanism: A Philosophical Framew... presence/absence of original empirical data or datasets in the paper
These operators are presented as conceptual/theoretical bridges rather than immediately quantifiable laboratory units.
Explicit methodological statement in the paper emphasizing interpretive/theoretical intent; no empirical operationalization reported.
high null result XChronos and Conscious Transhumanism: A Philosophical Framew... operationalizability (current lack of direct quantification) of Chronons/Hexachr...
The literature is heterogeneous (different LLM families/sizes, prompting techniques, participant persona modeling, environments, and evaluation protocols), which impedes general conclusions about when LLMs reliably mimic humans.
Review notes wide variation across study designs and methods in the 182 studies; inability to produce a single performance estimate motivated unified conceptual framing.
high null result Synthetic Participants Generated by Large Language Models: A... methodological heterogeneity across studies (variance in models, prompts, evalua...
Evaluations reporting outcomes predominantly relied on learner surveys, knowledge/skill tests, or self‑reported behavior change measures.
Methods of evaluation extracted from the included studies: most used surveys, tests, or self-report measures to assess Kirkpatrick‑Barr levels 1–3.
high null result Assessing the effectiveness of artificial intelligence educa... evaluation methods (surveys, tests, self-report behavior change)
The study used a cross-sectional quantitative survey (purposive sampling) of pharmaceutical-sector employees in Karnataka, India (N = 350) and analyzed relationships using PLS-SEM (SmartPLS 4.0).
Study design and methods as reported in the paper summary: cross-sectional survey, purposive sampling, N = 350, analysis via Partial Least Squares Structural Equation Modeling (SmartPLS 4.0).
high null result AI-driven stress management and performance optimization: A ... study design / methodological characteristics
Policy recommendations include: invest in open metadata standards; fund pilot programs to evaluate ROI (earnings, placement, employer satisfaction); require model governance and periodic external audits for AI-assisted curriculum tools; and support smaller providers via shared infrastructure or accreditation hubs.
Explicit policy recommendations in paper (prescriptive).
high null result Curriculum engineering: organisation, orientation, and manag... implementation of open metadata standards, number and outcomes of funded pilots,...
Careful attention is needed to validity/reliability of assessments and to selection bias in employment outcome measurement.
Paper's methodological caveat (prescriptive); no empirical bias analysis provided.
high null result Curriculum engineering: organisation, orientation, and manag... assessment validity/reliability metrics; selection bias indicators in outcome me...
Suggested evaluation metrics include placement rates, wage premiums, competency attainment, compliance scores, cost per qualification, and update latency.
Paper's recommended evaluation metrics (prescriptive).
high null result Curriculum engineering: organisation, orientation, and manag... placement rates, wage premiums, competency attainment, compliance scores, cost p...
Implementation requires integration with information systems for documentation, versioning, metadata, and audit trails, and benefits from continuous monitoring dashboards.
Paper's technical implementation recommendations (prescriptive).
high null result Curriculum engineering: organisation, orientation, and manag... IT integration level: documentation/versioning/metadata/audit trail availability...
Recommended analysis methods are qualitative (semi-structured interviews, focus groups, document review) and quantitative (surveys, competency mapping, statistical analysis of outcomes), plus systematic audit methods including traceability checks.
Paper's methods section (methodological specification).
high null result Curriculum engineering: organisation, orientation, and manag... use of specified qualitative, quantitative, and audit methods
Data inputs for the framework should include competency taxonomies, labor-market signals, regulatory requirements, learner assessment results, and stakeholder interviews.
Paper's data-input specification (descriptive).
high null result Curriculum engineering: organisation, orientation, and manag... presence and use of specified data inputs
Management principles emphasised are transparency, traceability of outcomes, IT integration for documentation, and continuous monitoring/evaluation.
Explicit management principles in paper (prescriptive).
high null result Curriculum engineering: organisation, orientation, and manag... degree of adherence to transparency, traceability, IT integration, continuous mo...
Research and audit should emphasise validity, reliability, and compliance using mixed methods (qualitative interviews/focus groups; quantitative surveys/statistics) and systematic curriculum audits.
Recommended research & audit approach in paper (methodological guidance).
high null result Curriculum engineering: organisation, orientation, and manag... application of mixed-methods and systematic audits to assess validity/reliabilit...
Tools recommended include logigrams (visual decision/compliance flows) and algorigram (algorithmic step-flows for planning, assessment, audit).
Tool definitions and recommendations in paper (descriptive).
high null result Curriculum engineering: organisation, orientation, and manag... adoption of logigrams and algorigrams in curricula tooling
Core components of the framework are inputs (learner needs, industry requirements, regulatory standards), processes (curriculum mapping, competency alignment, career assessment), and outputs (structured lesson plans, compliance-ready frameworks, career-path documentation).
Framework component list provided in paper (descriptive).
high null result Curriculum engineering: organisation, orientation, and manag... presence and completeness of inputs/processes/outputs in implementation
Scope of the program includes curriculum design, organisational management, career-alignment, and audit/compliance processes.
Explicit scope statement in paper (descriptive).
high null result Curriculum engineering: organisation, orientation, and manag... inclusion of specified scope elements in program design
The framework foregrounds logical modelling (logigrams, algorigrams) and mixed-methods data analysis to support design, auditability, and alignment with industry and regulatory standards.
Paper's methodological design and tool recommendations (conceptual). No empirical implementation data reported.
high null result Curriculum engineering: organisation, orientation, and manag... use of logical modelling tools and mixed-methods analysis in curriculum design
The program offers a comprehensive curriculum-engineering framework linking organizational orientation, management systems, lesson planning, and career assessment into traceable, compliance-ready curriculum products.
Paper's program description and framework specification (conceptual); no empirical evaluation or sample size reported.
high null result Curriculum engineering: organisation, orientation, and manag... availability of traceable, compliance-ready curriculum products (framework prese...
The paper calls for subsequent quantitative validation (using task-based, matched employer-employee, and provider-level panel data) to estimate causal impacts on productivity, health outcomes, wages, and employment composition across the three interaction levels.
Stated research agenda and measurement recommendations in the paper's discussion section.
high null result Toward human+ medical professionals: navigating AI integrati... need for causal estimates of productivity, health outcomes, wages, employment co...
The study is qualitative and small-sample (four case) and therefore interpretive and illustrative rather than statistically generalizable.
Explicit methodological statement in the paper: design = qualitative multiple case study, sample = four AI healthcare applications.
high null result Toward human+ medical professionals: navigating AI integrati... generalizability/external validity
The study identifies a three-level taxonomy of human–AI interaction in healthcare: AI-assisted, AI-augmented, and AI-automated.
Conceptual taxonomy derived from multiple qualitative case studies (n=4) using cross-case comparison and Bolton et al. (2018)'s three-dimensional service-innovation framework.
high null result Toward human+ medical professionals: navigating AI integrati... classification of AI–human interaction (taxonomic mapping)
Few longitudinal or randomized studies were found, which limits the evidence base for causal claims about digital transformation's effect on productivity.
Review recorded a limited number of longitudinal analyses and quasi-experimental designs among the 145 studies; randomized studies were scarce or absent.
high null result Digital transformation and its relationship with work produc... presence/absence of longitudinal/randomized designs relevant to causal inference
Measurement heterogeneity across studies includes self-reported productivity, output-per-worker metrics, and process efficiency indicators.
Extraction of productivity indicators from included studies (detailed in Methods/Extraction fields) showed multiple distinct measurement approaches.
high null result Digital transformation and its relationship with work produc... types of productivity measures used in studies
There is a lack of standardized instruments and inconsistent controls for confounding factors across studies, limiting causal inference about the effect of digital transformation on productivity.
Review extraction documented varied instruments/measures and inconsistent adjustment for confounders across the included studies; few randomized or robust longitudinal designs were found.
high null result Digital transformation and its relationship with work produc... quality of causal inference (control for confounding, presence of randomized/lon...
Heterogeneous definitions of 'digital transformation' and a variety of productivity measurement approaches prevented a formal quantitative meta-analysis.
Extraction found wide variation in how digital transformation and productivity were defined and measured across the 145 studies (self-reported productivity, output per worker, process efficiency metrics, etc.), leading authors to forgo meta-analysis.
high null result Digital transformation and its relationship with work produc... feasibility of quantitative meta-analysis / cross-study comparability
535 records were identified across Scopus, Web of Science, ScienceDirect, IEEE Xplore, and Google Scholar, of which 145 met PRISMA 2020 inclusion criteria.
Search and screening procedure documented in the review: initial database searches yielded 535 records → duplicates removed → screening → full-text evaluation → 145 included studies.
high null result Digital transformation and its relationship with work produc... study selection counts (records identified and studies included)
There are few large-scale randomized controlled trials (RCTs) showing direct patient outcome improvements from GenAI CDS; high-quality real-world and longitudinal studies are limited but essential.
Evidence-maturity statement in the paper summarizing the literature; the paper explicitly notes scarcity of large RCTs and longitudinal evaluations.
high null result GenAI and clinical decision making in general practice number of large-scale RCTs reporting patient outcome improvements; availability ...
The paper's empirical scope is primarily conceptual/theoretical and literature‑based rather than an empirical case study or large‑scale data experiment; it emphasizes the need for future empirical validation.
Explicit methodological description within the paper stating reliance on literature review and conceptual development; absence of empirical sample or case study.
high null result A Review of Manufacturing Operations Research Integration in... presence/absence of empirical validation within the study
This work is a conceptual framework and design proposal synthesizing methods from recommender systems and HRI rather than a report of novel empirical experiments.
Explicit statement in the Data & Methods section of the paper.
high null result Reimagining Social Robots as Recommender Systems: Foundation... presence/absence of original empirical experiments (absence)
The abstract does not report the study sample size, sectoral scope, or country/context—limiting assessment of external validity and generalizability.
Observation of reporting in the paper's abstract (absence of sample size, sectoral/country context information in the abstract as provided).
high null result Reimagining Stakeholder Engagement Through Generative AI: A ... Completeness of methodological reporting (sample/context disclosure)
The study used a two-stage mixed-methods design: a qualitative exploratory phase to surface determinants of trust and inertia, followed by a quantitative phase to validate the conceptual framework.
Methods description in the paper: explicit two-stage mixed-methods approach (qualitative then quantitative) used to identify and test determinants of initial trust and inertia toward GAICS.
high null result Reimagining Stakeholder Engagement Through Generative AI: A ... Study design / methodological approach
The study has potential selection and ecological-validity constraints because it was conducted at two institutions across six courses, limiting generalizability.
Authors note limitations regarding sample scope (two institutions, six courses) and the ecological validity of the experimental tasks/settings.
high null result Expanding the lens: multi-institutional evidence on student ... external validity/generalizability (limitation)
The study employed a multi-method approach combining experimental quantitative analysis (descriptives, GLM, non-parametric robustness checks) with qualitative topic-based coding of open-ended survey responses.
Methods description: randomized/experimental assignment; quantitative analyses using GLM and non-parametric tests; qualitative topic-based coding of student responses; sample N = 254 across six courses at two institutions.
high null result Expanding the lens: multi-institutional evidence on student ... study methodology (mixed-methods design)
The study did not directly measure accessibility or impacts on students with disabilities, though qualitative results suggest possible intersections with inclusive and multimodal learning design.
Limitation stated by authors: no direct measurement of accessibility outcomes; qualitative responses hinted at potential relevance to inclusive design but no empirical measurement of disability-related impacts.
high null result Expanding the lens: multi-institutional evidence on student ... accessibility/disability-related educational outcomes (not measured)
The study focused on short-term, knowledge-based tasks and did not measure long-term learning or retention.
Authors explicitly note as a limitation that the experimental tasks were short-term and knowledge-based and that long-term retention was not measured.
high null result Expanding the lens: multi-institutional evidence on student ... long-term learning/retention (not measured)
Empirical generalization across all climate-AI systems is constrained by heterogeneous data availability and proprietary models, limiting the ability to produce universal quantitative claims.
Stated methodological limitation in the paper, noting heterogeneous data and the proprietary nature of some models restrict broad generalization.
high null result The Rise of AI in Weather and Climate Information and its Im... Extent of empirical generalizability across climate-AI systems
The paper does not provide granular quantitative estimates of the economic cost of infrastructural asymmetries in climate-AI.
Explicit limitation stated by the authors in the Methods/Limitations section.
high null result The Rise of AI in Weather and Climate Information and its Im... Absence of quantified economic cost estimates in the paper
There is a need for empirical research quantifying earnings dispersion, labor substitution effects, and the welfare impacts of GenAI-driven content economies over time.
Explicit research recommendation made in the paper based on gaps identified during analysis of the 377 videos (study is qualitative and does not measure these outcomes).
high null result Monetizing Generative AI: YouTubers' Collective Knowledge on... absence of quantitative measures in current study / identified need for future m...
The analysis identifies ten shared use cases that creators present as pathways to income using GenAI.
Coding of the 377-video corpus resulted in a catalog of ten use cases (as reported in the paper).
high null result Monetizing Generative AI: YouTubers' Collective Knowledge on... count and identification of distinct use-case categories (ten)
Because the sample is small and purposive and the design is qualitative, insights are rich but not statistically representative or quantified across the broader research landscape.
Authors' stated study limitations in the paper acknowledging small purposive sample (n=16) and qualitative design.
high null result RCTs & Human Uplift Studies: Methodological Challenges and P... representativeness and generalizability of study findings
The study's data come from semi-structured interviews with 16 expert practitioners across biosecurity, cybersecurity, education, and labor.
Study methods reported in the paper: qualitative data source explicitly stated as 16 semi-structured interviews across listed domains.
high null result RCTs & Human Uplift Studies: Methodological Challenges and P... sample size and domain coverage of interviews
The authors released their code and data for reproducibility at https://github.com/blocksecteam/ReEVMBench/.
Statement in the paper indicating public release of code and dataset at the provided GitHub URL.
high null result Re-Evaluating EVMBench: Are AI Agents Ready for Smart Contra... code_and_data_availability (repository_link)
Crystallization Efficiency (CE) is defined as Useful_Crystallized_Knowledge / (Human_Effort × Time).
Operational formalism and metric definitions presented in the paper (explicit formula provided). This is a proposed metric, not an empirically validated measure.
high null result Nurture-First Agent Development: Building Domain-Expert AI A... Crystallization Efficiency as defined
The paper proposes operational patterns (Dual-Workspace Pattern separating live interaction workspace and persistent knowledge workspace) and a Spiral Development Model (iterative interaction → crystallization → validation → redeployment).
Operational framework section describing patterns and workflows; illustrated in the case study implementation.
high null result Nurture-First Agent Development: Building Domain-Expert AI A... existence and application of dual-workspace and spiral development workflows
The Knowledge Crystallization Cycle formalizes operations (extract, synthesize, validate, integrate) and proposes efficiency and quality metrics including Crystallization Efficiency (CE), Fidelity, Reuse Rate, and Freshness/Volatility Score.
Operational formalism section of the paper presenting metric definitions and proposed calculations (e.g., CE = Useful_Crystallized_Knowledge / (Human_Effort × Time)). These are proposed metrics, not validated at scale.
high null result Nurture-First Agent Development: Building Domain-Expert AI A... Crystallization Efficiency and related proposed metrics
The paper introduces a Three-Layer Cognitive Architecture that organizes agent knowledge by volatility and degree of personalization (stable/core knowledge; institutionalized heuristics/patterns; volatile/session-level tacit details).
Architectural specification presented in the paper (conceptual design document). No experimental validation beyond the illustrative case study.
high null result Nurture-First Agent Development: Building Domain-Expert AI A... categorization of knowledge artifacts into three volatility/personalization laye...