The Commonplace
Home Dashboard Papers Evidence Syntheses Digests 🎲

Evidence (4793 claims)

Adoption
5539 claims
Productivity
4793 claims
Governance
4333 claims
Human-AI Collaboration
3326 claims
Labor Markets
2657 claims
Innovation
2510 claims
Org Design
2469 claims
Skills & Training
2017 claims
Inequality
1378 claims

Evidence Matrix

Claim counts by outcome category and direction of finding.

Outcome Positive Negative Mixed Null Total
Other 402 112 67 480 1076
Governance & Regulation 402 192 122 62 790
Research Productivity 249 98 34 311 697
Organizational Efficiency 395 95 70 40 603
Technology Adoption Rate 321 126 73 39 564
Firm Productivity 306 39 70 12 432
Output Quality 256 66 25 28 375
AI Safety & Ethics 116 177 44 24 363
Market Structure 107 128 85 14 339
Decision Quality 177 76 38 20 315
Fiscal & Macroeconomic 89 58 33 22 209
Employment Level 77 34 80 9 202
Skill Acquisition 92 33 40 9 174
Innovation Output 120 12 23 12 168
Firm Revenue 98 34 22 154
Consumer Welfare 73 31 37 7 148
Task Allocation 84 16 33 7 140
Inequality Measures 25 77 32 5 139
Regulatory Compliance 54 63 13 3 133
Error Rate 44 51 6 101
Task Completion Time 88 5 4 3 100
Training Effectiveness 58 12 12 16 99
Worker Satisfaction 47 32 11 7 97
Wages & Compensation 53 15 20 5 93
Team Performance 47 12 15 7 82
Automation Exposure 24 22 9 6 62
Job Displacement 6 38 13 57
Hiring & Recruitment 41 4 6 3 54
Developer Productivity 34 4 3 1 42
Social Protection 22 10 6 2 40
Creative Output 16 7 5 1 29
Labor Share of Income 12 5 9 26
Skill Obsolescence 3 20 2 25
Worker Turnover 10 12 3 25
Clear
Productivity Remove filter
Fine-tuning a parameter-efficient 7B model (Qwen2.5-Coder-7B) via reinforcement learning in an OpenEnv-compatible environment yields near-state-of-the-art automated slide-generation: the tuned 7B model reaches 91.2% of Claude Opus 4.6’s quality.
Empirical evaluation on 48 diverse business briefs comparing six models; reported relative quality score of tuned Qwen2.5-Coder-7B = 91.2% of Claude Opus 4.6.
high positive Learning to Present: Inverse Specification Rewards for Agent... Relative slide-generation quality (percent of Claude Opus 4.6 quality) across 48...
Managing captures, traces, and replay sessions from a unified single design database ensures consistency across replay targets and sessions.
Method description emphasizes a single design database coordinating captures and replays across simulation and emulation for the demonstrator system. (Operational claim demonstrated in the implementation; no metrics on error reduction provided.)
high positive ODIN-Based CPU-GPU Architecture with Replay-Driven Simulatio... consistency of trace/replay data and configuration across targets
The captured traces can be deterministically replayed across different execution targets (software/hardware simulation and hardware emulation), reducing cross-platform setup complexity and discrepancies.
The same captured waveforms/traces were replayed on both simulation and emulation environments for the ODIN demonstrator; cross-target replay was part of the described method. (Demonstrated on the single reported system; no broad cross-toolchain study provided.)
high positive ODIN-Based CPU-GPU Architecture with Replay-Driven Simulatio... consistency of reproduced behavior across simulator and emulator targets
BATQuant significantly outperforms prior post-training quantization (PTQ) methods on MXFP4 microscaling floating-point formats under aggressive quantization.
Comparative experiments against rotation-based PTQ techniques and other existing PTQ baselines on the described multimodal and language tasks; improvements shown in benchmark metrics and recovery percentages in the paper's experimental section.
high positive BATQuant: Outlier-resilient MXFP4 Quantization via Learnable... Task-specific accuracy/quality metrics and percent recovery relative to full-pre...
BATQuant recovers up to 96.43% of full-precision performance under aggressive W4A4KV16 quantization on MLLMs and LLMs.
Empirical evaluation reported in the paper: experiments on multiple multimodal large language models (MLLMs) and standard LLMs using an aggressive W4A4KV16 quantization setup; performance reported as percentage of full-precision performance recovered (specific models, benchmark names, and exact sample sizes not enumerated in the summary).
high positive BATQuant: Outlier-resilient MXFP4 Quantization via Learnable... Percentage of full-precision performance recovered (model quality/accuracy on mu...
The paper provides concrete, regulation-inspired policy examples (e.g., content prohibition, sensitive data exfiltration) showing how they map into the Policy function.
Worked, illustrative examples included in the paper mapping regulatory constraints to the Policy(agent_id, partial_path, proposed_action, org_state) formalism.
high positive Runtime Governance for AI Agents: Policies on Paths representability of regulation-inspired policies in the formalism (yes/no; examp...
Runtime policy evaluation can intercept, score, log, allow/modify/block actions, and update organizational state as part of an agent's execution loop (reference implementation architecture).
Reference implementation design described in the paper (runtime policy evaluator hooks, logging, enforcement actions); architectural reasoning and pseudo-workflows provided; no production deployment data.
high positive Runtime Governance for AI Agents: Policies on Paths feasibility of integrating runtime policy evaluator into agent loops (architectu...
Policies can be formalized as deterministic functions p_violation = Policy(agent_id, partial_path, proposed_action, org_state) that return a probability or score of violation for a proposed next action.
Formal definition and mapping in the paper; worked examples showing how regulatory-style constraints map into this function; no large-scale empirical validation.
high positive Runtime Governance for AI Agents: Policies on Paths expressiveness of policy formalism (ability to represent targeted constraints)
Effective governance for agentic LLM systems requires treating the execution path as the central object and performing runtime evaluation of proposed next actions given the partial path.
Theoretical argument and formal proposal of runtime policy evaluator that takes (agent_id, partial_path, proposed_action, org_state) and returns a violation probability; reference architecture described; illustrative examples.
high positive Runtime Governance for AI Agents: Policies on Paths governance effectiveness for path-dependent policies (qualitative/coverage)
The surrogate-driven inverse-design pipeline transfers to physical hardware — designs produced by the CNN+GA pipeline were realized and validated experimentally.
Two fabricated prototypes implemented the optimized pixelated combiners and GaN HEMT Doherty PAs; measured performance metrics correspond to the designs, demonstrating transfer from surrogate-driven design to hardware.
high positive Deep Learning-Driven Black-Box Doherty Power Amplifier with ... consistency between surrogate-driven design outputs and measured prototype perfo...
Under a 20 MHz 5G-NR-like waveform (9 dB PAPR) with digital predistortion (DPD), each prototype reached average PAE greater than 51% while meeting ACLR ≤ −60.8 dBc.
Realistic waveform testing described: a 20 MHz 5G‑NR-like signal with 9 dB PAPR was applied to the prototypes, DPD was used, and measurements reported average PAE > 51% and ACLR ≤ −60.8 dBc for each prototype.
high positive Deep Learning-Driven Black-Box Doherty Power Amplifier with ... average power-added efficiency (PAE %) and adjacent channel leakage ratio (ACLR,...
Each prototype demonstrated drain efficiency greater than 52% at 9 dB back-off.
Back-off efficiency measurements reported for the fabricated prototypes showing drain efficiency > 52% at 9 dB back-off.
high positive Deep Learning-Driven Black-Box Doherty Power Amplifier with ... drain efficiency at 9 dB back-off (%)
Each prototype produced output power exceeding 44.1 dBm at 2.75 GHz.
Measured output power reported from RF characterization of the two fabricated prototypes; reported value > 44.1 dBm at the test frequency.
Each fabricated prototype achieved peak drain efficiency greater than 74%.
Measured RF characterization reported for the two prototypes showing peak drain efficiency > 74%; measurements conducted on fabricated hardware at 2.75 GHz.
high positive Deep Learning-Driven Black-Box Doherty Power Amplifier with ... peak drain efficiency (%)
A genetic-algorithm (GA) blackbox optimizer paired with the CNN surrogate can effectively search the discrete multi-port pixel layout space to synthesize output combiners for Doherty amplifiers.
Method description: CNN surrogate embedded in a blackbox Doherty framework and used within a GA to select pixelated combiner layouts; successful designs were produced and taken to fabrication.
high positive Deep Learning-Driven Black-Box Doherty Power Amplifier with ... ability of optimization stack to find feasible combiner layouts that meet system...
The parallel associative scan enables the reductions required by Newton-style updates across time steps, thereby enabling parallelism across sequence length.
Algorithmic construction and implementation details in the thesis showing how associative scan operations aggregate intermediate Jacobian/ update information across time; examples provided in implementation section.
high positive Unifying Optimization and Dynamics to Parallelize Sequential... practical parallelizability / ability to compute required reductions in parallel
The thesis proves linear convergence rates for a family of fixed-point/Newton-like solvers, with rates depending on approximation accuracy and stability properties of the chosen method.
Mathematical proofs and convergence theorems provided in the theoretical analysis section establishing linear rates under stated assumptions (bounds on approximation error, stability metrics).
high positive Unifying Optimization and Dynamics to Parallelize Sequential... convergence rate (linear) as a function of approximation error and stability mea...
Evaluation of dynamical systems can be cast as solving a system of nonlinear equations, enabling parallel solution methods.
Methodological framing and derivation in the thesis showing recurrent updates and Markov transitions can be represented as a global nonlinear root-finding problem; algorithmic constructions follow from this representation.
high positive Unifying Optimization and Dynamics to Parallelize Sequential... feasibility of parallel solution (existence of equivalent nonlinear-system formu...
Explicit enforcement of signal constraints in DeePC provides a safety/operational advantage over many pure learning approaches that do not explicitly enforce hard constraints.
Algorithmic formulation includes constraints in the optimization; paper contrasts this with unconstrained learning-based controllers and demonstrates constrained, feasible actuation in simulation.
high positive Data-driven generalized perimeter control: Zürich case study explicit constraint satisfaction and operational safety of signal timings
DeePC can compute traffic-light actuation sequences that respect hard operational and safety constraints (e.g., phasing, minimum/maximum green times).
Formulation of DeePC as a constrained optimization problem in the paper with explicit constraint terms for signal phasing and safety; implemented in simulation experiments where constraints are enforced in the controller optimization.
high positive Data-driven generalized perimeter control: Zürich case study constraint satisfaction / feasibility of computed actuation sequences
Reframing urban traffic dynamics with behavioral systems theory allows system evolution to be learned and predicted directly from measured input–output data (no explicit model identification).
Theoretical exposition in the paper showing that traffic trajectories can be represented as linear combinations of past measured trajectories via Hankel/data matrices; used as the basis for predictive control (DeePC).
high positive Data-driven generalized perimeter control: Zürich case study predictive capability from measured I/O trajectories (ability to forecast future...
Applying DeePC yields measurable improvements in system-level outcomes (reduced total travel time and CO2 emissions) in a very large, high-fidelity microscopic simulation of Zürich.
Simulation experiments in a city-scale, high-fidelity microscopic closed-loop simulator of Zürich comparing DeePC-controlled signals against baseline controllers (e.g., fixed-time or standard adaptive schemes); reported reductions in aggregated metrics (total travel time and CO2 emissions).
high positive Data-driven generalized perimeter control: Zürich case study total travel time; CO2 emissions
A model-free traffic control approach (DeePC) can steer urban traffic via dynamic traffic-light control without building explicit traffic models.
Algorithmic/theoretical development (behavioral systems theory + DeePC) and controller-in-loop experiments in a high-fidelity microscopic closed-loop simulator of Zürich demonstrating closed-loop control using only input–output trajectory data (Hankel matrices) rather than parametric model identification.
high positive Data-driven generalized perimeter control: Zürich case study ability to generate feasible control (traffic-light) actuation sequences and clo...
Traditional machine-learning baselines were included for comparison in the benchmarks.
Paper explicitly states that traditional ML baselines were used alongside TSFMs in benchmarking experiments. The summary does not list which baselines or their quantitative results.
high positive Bridging the High-Frequency Data Gap: A Millisecond-Resoluti... inclusion of traditional ML baseline models in comparative evaluation
The dataset sampling resolution is at the millisecond level, enabling forecasting horizons from 1 step (100 ms) up to 96 steps (9.6 s).
Paper states sampling resolution is millisecond-level and defines forecasting tasks spanning 1 to 96 steps (100 ms to 9.6 s). This is a methodological description rather than an experimental metric.
high positive Bridging the High-Frequency Data Gap: A Millisecond-Resoluti... supported forecast horizons (temporal prediction horizon: 100 ms–9.6 s)
Introduces a new millisecond-resolution dataset of wireless channel and traffic-condition measurements from an operational 5G deployment.
Paper describes collection of operational 5G telemetry at millisecond sampling resolution; dataset is presented as a novel domain addition to TSFM pretraining corpora. Exact number of records/sessions not specified in the provided summary.
high positive Bridging the High-Frequency Data Gap: A Millisecond-Resoluti... availability and characteristics of a millisecond-resolution 5G measurement data...
Historical transitions in standard work hours (e.g., six-day to five-day week) show that phased implementation, collective bargaining, and complementary policies can make work-time reductions feasible and economically beneficial.
Historical analyses and case studies of past industrialized-country workweek transitions cited in the synthesis; evidence drawn from historical institutional records and prior economic histories rather than a unified econometric analysis.
high positive A Shorter Workweek as a Policy Response to AI-Driven Labor D... feasibility and economic outcomes of phased work-time reductions (employment, pr...
The evaluation compared models on multiple metrics (accuracy, precision, recall, F1, AUC) across repeated trials and cross-company tests, and reported gains for AI methods across these metrics.
Evaluation protocol described: repeated trials, cross-validation, holdout sets, cross-company tests; reported performance improvements for AI models on the listed metrics.
high positive Adoption of AI-Based HR Analytics and Its Impact on Firm Pro... Classification evaluation metrics (accuracy, precision, recall, F1, AUC)
Ensemble methods and deep learning models show the largest and most consistent improvements in predictive performance relative to classic statistical models.
Aggregate results across repeated trials and evaluation metrics indicate Random Forests and Gradient Boosting (ensembles) and deep neural networks outperform linear/logistic regression and other baselines on the publicly available datasets used.
high positive Adoption of AI-Based HR Analytics and Its Impact on Firm Pro... Predictive performance (accuracy, F1, AUC, etc.)
Modern AI-driven prediction methods (especially ensemble models and deep neural networks) systematically outperform traditional statistical approaches at predicting job performance in publicly available workforce datasets.
Direct model comparison reported in the paper: baseline statistical models (linear/logistic regression) versus machine learning models (Random Forest, Gradient Boosting, SVM, deep neural networks) evaluated on multiple publicly available workforce datasets using cross-validation and holdout sets; performance reported on accuracy, precision, recall, F1, and AUC across repeated trials.
high positive Adoption of AI-Based HR Analytics and Its Impact on Firm Pro... Job performance prediction (classification performance metrics: accuracy, precis...
Research priorities include rigorous real-world trials assessing patient outcomes, cost-effectiveness, and labor impacts; comparative studies of integration strategies; measurement of long-run workforce effects; and development of standard metrics and monitoring frameworks.
Explicit recommendations from the narrative review based on identified gaps: scarcity of RCTs, economic analyses, and long-term workforce studies.
high positive Human-AI interaction and collaboration in radiology: from co... number and quality of real-world trials, existence of standardized monitoring fr...
Economists and researchers should measure organizational mediators (governance, mentoring practices, learning processes) alongside AI adoption and use empirical designs such as difference-in-differences with phased rollouts, randomized mentoring/training interventions, matched employer–employee panels, and IV exploiting exogenous shocks to innovation backing to identify causal effects.
Methodological recommendations and proposed empirical designs contained in the paper; no implementation or empirical results reported.
high positive Revolutionizing Human Resource Development: A Theoretical Fr... feasibility and validity of empirical identification strategies for causal effec...
The integrated framework links multi-level outcomes: micro (individual skills, task performance), meso (team coordination, workflows), and macro (organizational strategy, innovation, productivity) effects to adaptive structuration processes and affordance actualization.
Framework specification and theoretical mapping across levels in the conceptual paper; no empirical validation or sample.
high positive Revolutionizing Human Resource Development: A Theoretical Fr... individual skills and performance; team coordination and workflow quality; organ...
The paper develops a conceptual framework that integrates Adaptive Structuration Theory (AST) and Affordance Actualization Theory (AAT) to explain how effective human–AI collaboration can be structured within organizations.
Conceptual/theoretical synthesis and literature integration combining AST and AAT streams; no original empirical data or sample reported (theoretical development).
high positive Revolutionizing Human Resource Development: A Theoretical Fr... explanatory power / conceptual framework for human–AI collaboration
As the competition progressed, teams relied more on the AI for larger subtasks (increasing delegation and reliance).
Time-series instrumentation of AI interactions and participant behavior during the live CTF with 41 participants showing increased frequency and scope of delegated tasks later in the event.
high positive Understanding Human-AI Collaboration in Cybersecurity Compet... frequency of delegation and average scope/complexity of delegated tasks over com...
One autonomous agent finished second overall on the fresh challenge set.
Final ranking/scoreboard from benchmarking the four autonomous agents against the live CTF challenge set and human teams; agent achieved overall 2nd place.
high positive Understanding Human-AI Collaboration in Cybersecurity Compet... overall ranking (2nd place) on the challenge set
In a live onsite Capture-the-Flag (CTF) study (41 participants), human teams increasingly delegated larger subtasks to an instrumented AI as the competition progressed.
Empirical observation and instrumentation of AI interactions during a live, onsite CTF with 41 human participants/teams; delegation and task-size metrics tracked over time during the event.
high positive Understanding Human-AI Collaboration in Cybersecurity Compet... degree/size of subtasks delegated to the AI over time (delegation rate and subta...
Reward shaping at the assignment layer enables an explicit trade-off between diagnostic accuracy and human labor by incorporating penalties for human involvement.
Methodology section describing reward shaping and experimental comparisons showing different accuracy/human-effort trade-offs (results reported in paper; exact experimental details not provided in the summary).
high positive Hierarchical Reinforcement Learning Based Human-AI Online Di... diagnostic accuracy vs human effort (as controlled by reward shaping)
Masked reinforcement learning techniques constrain or mask action spaces, reducing exploration over huge symptom/action spaces.
Paper describes use of masked RL to limit action options during training and execution; used in both assignment and execution layers (methodological claim supported by algorithmic description and experiments).
high positive Hierarchical Reinforcement Learning Based Human-AI Online Di... action-space reduction / sample efficiency / learning stability (as applied to s...
The upper layer ('master') learns turn-by-turn human–machine assignment using masked reinforcement learning with reward shaping to balance accuracy and human cost.
Methodological description in the paper and empirical results from experiments using masked RL and reward-shaped objectives at the assignment layer (implementation and experimental setup reported; dataset/sample size not specified in summary).
high positive Hierarchical Reinforcement Learning Based Human-AI Online Di... assignment policy performance; human effort allocation; diagnostic accuracy unde...
Service empathy mediates the relationship between employee emotion and collaboration proficiency.
Mediation analysis conducted on the experimental sample (n = 861) showing that measured 'service empathy' accounts for (part of) the effect of employee emotion on collaboration proficiency.
high positive Adoption of AI partners in temporary tasks: exploring the ef... collaboration proficiency
The paper advances augmentation debates by articulating the leader’s practical role when decision lead‑agency shifts between humans and AI and by detailing systemic HR changes needed to sustain performance, legitimacy and well‑being.
Stated contribution of the conceptual synthesis comparing existing augmentation and leadership literatures and providing an HR‑focused framework; descriptive of the paper's intellectual contribution.
high positive Symbiarchic leadership: leading integrated human and AI cybe... clarity of leader role; specification of HR system changes
Core practice 4 — Embed governance: make accountability, bias testing, privacy safeguards, audit trails, escalation thresholds and human oversight explicit and routine.
Prescriptive governance practice grounded in literature on algorithmic accountability and risk management and in practitioner examples; presented without original empirical validation.
high positive Symbiarchic leadership: leading integrated human and AI cybe... bias incidence; privacy breaches; auditability and compliance metrics
Core practice 3 — Manage the human–AI relationship: build adoption, psychological safety and calibrated trust; address automation anxiety and misuse.
Framework recommendation synthesizing organizational‑psychology and technology adoption literature plus practitioner observations; not tested empirically in the paper.
high positive Symbiarchic leadership: leading integrated human and AI cybe... adoption rates; psychological safety; calibrated trust; misuse incidents
Core practice 2 — Treat AI outputs as hypotheses: require human sensemaking and validation rather than blind adoption of model outputs.
Prescriptive practice derived from reviewed research and practitioner cases emphasizing human oversight; presented as framework guidance rather than empirically validated intervention.
high positive Symbiarchic leadership: leading integrated human and AI cybe... decision quality; error rates; incidence of blind automation
Core practice 1 — Allocate work by comparative advantage: assign tasks to humans or AI based on relative strengths (e.g., speed, pattern detection, contextual judgement).
Conceptual component of the framework drawn from synthesis of empirical findings in prior human–AI and task allocation literature and practitioner examples; no new empirical testing in the paper.
high positive Symbiarchic leadership: leading integrated human and AI cybe... task assignment efficiency; productivity from task allocation
AI methods have improved molecular property prediction, protein structure modelling, ADME/Tox prediction, NLP-based extraction from literature, virtual screening, and generative chemistry, accelerating early-stage tasks.
Compilation of benchmarking results, method-comparison studies, and applied case studies cited in the paper across these specific application areas.
high positive Has AI Reshaped Drug Discovery, or Is There Still a Long Way... accuracy/quality of property and structure predictions, throughput/speed of virt...
AI has materially improved efficiency, decision-making, and early-stage productivity in drug discovery, especially in hit discovery, property prediction, and protein modelling.
Synthesis of published benchmarking studies and industry case studies reported in the paper (e.g., improvements in virtual screening throughput, property-prediction benchmarks, and protein-structure prediction results such as those from folding competitions and tool evaluations).
high positive Has AI Reshaped Drug Discovery, or Is There Still a Long Way... efficiency and productivity in early-stage drug discovery (hit discovery rate, t...
Molecule operates a marketplace for decentralized clinical and preclinical assets, focusing on tokenizing drug assets and enabling investors to finance development.
Case-study description based on Molecule's public materials and marketplace listings; demonstrates platform design and transactions rather than long-term outcomes.
high positive Decentralized Autonomous Organizations in the Pharmaceutical... number of assets tokenized, capital deployed via the marketplace
VitaDAO is a community-driven organization funding and acquiring IP for longevity-related research, emphasizing open science and community governance.
Detailed case-study description drawing on VitaDAO's public documentation, governance records, and whitepaper materials.
high positive Decentralized Autonomous Organizations in the Pharmaceutical... IP acquisitions by VitaDAO, funding rounds executed, degree of open-science publ...