The Commonplace
Home Dashboard Papers Evidence Syntheses Digests 🎲

Digests

2026-05-11 2026-05-04 2026-04-27 2026-04-20 2026-04-13 2026-04-06 2026-04-04 2026-04-04-before 2026-03-30 2026-03-23 2026-03-20 2026-03-18 2026-03-15

Executive Summary

  • The single biggest finding this week: preregistered randomized controlled trials (RCTs) find conversational AIs causally influence real user choices in the samples studied, tripling sponsored selections in one shopping experiment and increasing petition-signing and donation rates in large trials.
  • Main tension or surprise: these short-term persuasion and productivity gains coexist with evidence, in the contexts studied, that AI assistance is associated with reductions in independent skill, persistence, and collective diversity, and that standard transparency (for example, “Sponsored” labels) did not meaningfully blunt AI influence in the experiments.
  • Bottom line for a time-constrained reader: regulators, platform leaders, and managers should treat AI-mediated persuasion as a potentially potent, hard-to-detect market force and consider redesigning incentives, governance, and training to help preserve long-term human capabilities and competitive markets.

The Big Picture

This week’s papers suggest a common theme: conversational AI behaves less like a search box and more like an intermediary that can shape behavior, in the contexts studied, often with limited user awareness. Two large preregistered randomized controlled trials (RCTs) find large effects of conversational interfaces on real decisions, tripling sponsored choices in one shopping experiment and increasing petition-signing and donations, while simple labels did not neutralize the effects and attitudinal measures did not predict behavior in those trials.

The productivity evidence is similarly mixed. RCTs find assistance raises short-term performance, yet several randomized and field experiments observe reduced persistence and weaker unaided follow-on performance after even brief exposure. Experiments also find that group-level creativity can shrink absent deliberate incentives, and a retail field RCT finds mandated joint AI use lowers output, while modest reframes and rewards for originality can recover diversity and top-end quality. Meanwhile, panel and firm-level studies associate algorithmic advantage and data concentration with fewer entrants and higher markups, suggesting competition risks in AI-intensive sectors unless access and interoperability are addressed.

Bottom line: the evidence suggests conversational AI can already function as a powerful, opaque intermediary in markets and organizations. Treat deployment as a structural change rather than a feature rollout, and pair adoption with incentive design, interface governance, and competition policy aimed at protecting long-run human capability and market contestability.

Top Papers

  • Conversational AIs nearly triple sponsored-product choices and users rarely spot the steering, Francesco Salvi, Alejandro Cuevas, Manoel Horta Ribeiro (preregistered RCT, high evidence, established) - Two preregistered experiments (N=2,012) find conversational agents steer users toward randomly tagged sponsored products at nearly three times the rate of traditional search, with little detection and no meaningful mitigation from simple “Sponsored” labels, indicating chat interfaces can alter consumer choice architecture in commercially significant ways in the sample studied.

  • Conversational AI materially increases real-world political and charitable actions but attitudes don't predict behaviour, Kobi Hackenburg, Luke Hewitt, Caroline Wagner, Ben M. Tappin, Christopher Summerfield (preregistered RCT, high evidence, established) - Massive preregistered trials (N=14,779) find increases in petition signing and donations after AI conversations, yet attitudinal shifts did not explain behavior in these trials, underscoring that monitoring actions may be more informative than surveying opinions for civic safeguards.

  • Rewarding originality counters AI’s homogenizing effect on group creativity, Nathanael Jo, Manish Raghavan (preregistered RCT, high evidence, established) - In an interactive AI co-creation task, shifting incentives from raw quality to relative originality produces more diverse collective outputs without abandoning AI, suggesting a practical lever for managers concerned about sameness in AI-assisted work.

Also Notable

Emerging Patterns

AI-mediated persuasion and governance - The strongest causal evidence this week is from preregistered RCTs that find conversational agents move market and civic behavior by large margins, often without users detecting the nudge. Simple sponsorship labels did not blunt the effect in the shopping experiments, and separate audits suggest some commercial models lean toward provider-favoring recommendations. Together, the evidence points to an interface governance problem: persuasion is embedded in dialogue flow and ranking, not just in ad copy. Editorially, transparency alone looks insufficient; platforms may need incentive audits, default rules, and independent testing to align model behavior with user welfare.

Human-AI collaboration, productivity, and skill dynamics - Short-term productivity and quality gains appear consistently, yet multiple experiments warn of “capability atrophy” risks: lower persistence and worse unaided performance after brief AI help, and homogenized outputs in teams. The design frontier is promising—lightweight cognitive reframes, explanations in concurrent-assistance modes, and originality-weighted incentives can redirect how people use AI without banning tools. Mandates and rigid protocols are sometimes counterproductive, while optional, well-scaffolded use helps top performers most. The trajectory suggests organizations should shift from blanket rollout to targeted enablement with metrics that track independence, learning, and diversity.

Market structure, competition, and innovation in AI economies - Panel and firm studies associate algorithmic and data advantages with fewer entrants, higher concentration, and higher markups, while venture-capital complexity metrics highlight geographic and sectoral concentration. Startup evidence is bifurcated: big-data adopters face higher failure risk but surviving adopters grow faster and raise more capital, consistent with selection and winner-take-most dynamics. City-level digital pilots are associated with resilience gains via talent and clustering, indicating policy can shape local capability. The policy arc is logical but incomplete: data access, interoperability, and procurement/open standards are plausible levers, yet we still lack causal tests of which bundles restore contestability at scale.

Claims to Watch

  • Chatbots as powerful persuaders (established) - Large preregistered RCTs find conversational AIs substantially increase sponsored selections and real civic actions while users rarely notice steering. - Implication: Regulators and platforms should treat conversational flows as an advertising and mobilization channel that warrants audits, guardrails, and choice-architecture oversight.

  • Assistance boosts now, blunts later (established) - RCTs find brief AI help improves immediate performance but lowers persistence and harms subsequent unaided work. - Implication: Add independence and learning metrics to KPIs, and gate assistance in training and assessment contexts.

  • Incentives can counter homogenization (established) - In RCTs, rewarding originality relative to peers preserves collective diversity without suppressing AI use. - Implication: Calibrate performance management to include originality and dispersion targets in AI-enabled teams.

  • Algorithmic advantage and entry barriers (suggestive) - Panel evidence associates algorithmic advantage and data concentration with fewer entrants and higher concentration. - Implication: Consider data portability, interoperability mandates, and merger scrutiny in AI-intensive markets.

  • Explanations help only when AI stays on (suggestive) - In education settings, explainable interfaces aid concurrent AI-supported performance but do not transfer to later independent tasks. - Implication: Use explanations to improve real-time decisions, but invest separately in practice without AI for durable learning.

Methods Spotlight

  • ClawBench live-web evaluation with interception layer — ClawBench: Can AI Agents Complete Everyday Online Tasks? - A realism-first benchmark that executes 153 tasks across live sites while safely intercepting end actions, revealing capability gaps missed by sandbox tests and informing go/no-go decisions for automation.

  • DESIGN-AWARE benchmark for code fixes — Does Pass Rate Tell the Whole Story? Evaluating Design Constraint Compliance in LLM-based Issue Resolution - Links repository design constraints to automated checks, showing test pass rates overstate patch quality and providing a template for higher-fidelity QA in AI-assisted engineering.

  • Accountability incompleteness modeling — The Accountability Horizon: An Impossibility Theorem for Governing Human-Agent Collectives - A formal model that sets theoretical limits on accountability beyond an autonomy threshold, pushing governance toward architectural constraints rather than post hoc liability.

The Week Ahead

  • Stand up persuasion audits in conversational flows and test stronger defaults than labels, including opt-out, demotion of sponsored content in assistant answers, and conflict-of-interest checks.
  • Add persistence, unaided performance, and output-dispersion metrics to AI deployment dashboards; pilot originality-weighted incentives in creative and analytical teams.
  • Stress-test competition in AI-heavy lines of business by mapping data bottlenecks and planning for interoperability and data-portability remedies ahead of enforcement.
  • Validate agent readiness with live-environment benchmarks and design-aware QA before exposing customer-facing tasks to automation.
  • Scope governance to architectural moves: cap autonomy where accountability is infeasible, and trial separation-of-powers designs for agent actions tied to human principals.

Reading List