The Commonplace
Home Dashboard Papers Evidence Digests 🎲
← Papers

Consensus can mask fragility: organizations that suppress divergent reports create the illusion of agreement and miss accumulating tail risks until abrupt correction. Simple governance diagnostics—recorded dissent patterns, voting-anonymous gaps, and method/pipeline diversity—can distinguish healthy integration from dangerous exclusion.

Cohesion as Concentration: Exclusion-Driven Fragility in Financial Organizations
Foong Soon Cheong · March 16, 2026 · Journal of risk and financial management
openalex theoretical n/a evidence 7/10 relevance DOI Source PDF
A formal signal-aggregation model shows that apparent organizational cohesion can arise from exclusionary filtering rather than integrative debate, producing illusory precision, concentrated effective inputs, delayed recognition of tail regimes, and diagnostic observables to detect latent fragility.

Financial crises repeatedly reveal organizations that appear internally aligned while failing to recognize accumulating tail risks. This paper argues that cohesion is observationally ambiguous. It can arise from information integration, in which heterogeneous inputs are debated and synthesized, or from exclusion, in which variance is removed through conformity pressure, gatekeeping, and intolerance of dissent. This distinction is formalized using a signal aggregation model in which an organization maintains an anchor belief and achieves agreement through two exclusion channels: report shrinkage toward the anchor and a tolerance rule that discards reports deviating beyond a threshold. Relative to a full inclusion benchmark, exclusion based cohesion jointly produces state contingent bias that is small in normal regimes but grows sharply under displacement, illusory precision in which observed disagreement falls as tail regime estimation error rises, effective concentration of decision inputs below the nominal participant count, and, when the anchor updates from filtered aggregates, dynamic lock in with delayed regime recognition and abrupt correction. External inputs that bypass internal filtering shorten recognition delays. The model yields testable governance diagnostics linking latent fragility to observable patterns in recorded dissent, anonymous to formal voting gaps, scenario set diversity, pipeline and method concentration, and anchor lag. The central implication is that governance systems should treat low internal conflict and unanimity as potentially diagnostic of variance depletion and should monitor whether heterogeneity is integrated or excluded before stress reveals the difference.

Summary

Main Finding

Organizations that exhibit internal cohesion can hide two very different processes: (1) healthy information integration, or (2) exclusionary variance depletion (conformity, gatekeeping, intolerance of dissent). Using a formal signal-aggregation model, the paper shows that exclusionary cohesion produces fragile, state-contingent biases — small in normal times but large in displaced/tail regimes — along with "illusory precision", effective concentration of inputs, and dynamic lock‑in (delayed detection of regime shifts and abrupt correction). Low observed disagreement and unanimity can therefore signal latent fragility rather than robustness.

Key Points

  • Cohesion is observationally ambiguous:
    • Integration: heterogeneous inputs debated and synthesized, preserving variance.
    • Exclusion: disagreement removed via conformity pressure, report shrinkage toward an anchor, and discarding outlying reports (tolerance rule).
  • Model mechanisms:
    • Anchor belief maintained by organization; two exclusion channels formalized:
      • Report shrinkage: individual reports are pulled toward the anchor before aggregation.
      • Tolerance threshold: reports beyond a deviation threshold are discarded.
  • Main model consequences (relative to full inclusion benchmark):
    • State-contingent bias: bias is small in normal regimes but grows sharply when the environment displaces from typical regimes.
    • Illusory precision: observed disagreement falls even as true estimation error in tails rises (apparent confidence increases while accuracy declines).
    • Effective concentration: nominal participant counts overstate the independent information actually influencing decisions (variance depletion creates fewer effective inputs).
    • Dynamic lock-in: if the anchor updates from filtered aggregates, the system delays recognizing regime shifts and then corrects abruptly (overshoot/jerkiness).
    • External inputs that bypass internal filtering reduce recognition delays and break lock-in.
  • Testable governance diagnostics proposed:
    • Patterns in recorded dissent (frequency, amplitude) and differences between anonymous vs formal voting records.
    • Diversity of scenario sets and scenario-generation methods.
    • Concentration in pipelines and methods (e.g., few dominant teams/models).
    • Anchor lag: delay between true regime changes and organizational anchor update.
    • Metrics linking latent fragility to observable disagreement statistics, discarded-report rates, and effective participant counts.

Data & Methods

  • Formal model: a signal-aggregation framework where each participant/report supplies a noisy signal about the true state; the organization maintains an anchor and applies two filtering rules (shrinkage and tolerance/discarding) before aggregation.
  • Benchmark: full-inclusion aggregation (no shrinkage or discarding) used to compare bias, variance, and responsiveness.
  • Analytical results characterize how filtering transforms the distribution of aggregated estimates, produces state-dependent bias, and reduces observed disagreement while increasing tail estimation error.
  • Numerical examples / simulations (implied) illustrate dynamic behavior: delayed regime recognition, abrupt corrections, and how bypassing filters shortens delays.
  • Output: model-derived diagnostic statistics and proposed empirical metrics to detect exclusionary cohesion in real organizations (logs of dissent, variance of scenario sets, voting/anonymous-report gaps, pipeline concentration measures).

(Note: the supplied summary does not list specific datasets; empirical application would require collection of internal reports, vote records, scenario libraries, and metadata on pipelines/method diversity.)

Implications for AI Economics

  • For AI governance and risk assessment:
    • Unanimity or low internal conflict is not a reliable indicator of robust decision-making; it can be a leading indicator of hidden fragility.
    • Safety assessments, deployment approvals, and internal risk reviews should treat low disagreement with suspicion and test whether heterogeneity is being integrated or excluded.
  • Practical governance recommendations:
    • Preserve and surface raw/unfiltered inputs (anonymized if necessary) to measure true disagreement and avoid report shrinkage.
    • Require parallel, independent evaluation pipelines (reducing method/pipeline concentration) and track their correlations to estimate effective participant count.
    • Implement mechanisms for external, bypassing inputs (third‑party audits, red teams with independent reporting paths) to shorten regime recognition delays.
    • Monitor diagnostics: recorded dissent rates, anonymous vs formal voting gaps, fraction of discarded/outlier reports, scenario-set diversity, and anchor lag statistics.
    • Avoid aggregation rules that unduly shrink reports toward prevailing anchors; consider blind aggregation and aggregation rules that preserve variance.
  • For economic modeling of AI-related systemic risk:
    • Incorporate endogenous information-filtering and conformity mechanisms into models of organizational decision-making and technology deployment.
    • Evaluate how homogeneity across firms (e.g., common model architectures, shared training data, similar evaluation pipelines) can create systemic illusory precision and amplify tail risks.
    • Use the proposed diagnostics to empirically identify organizations or sectors at elevated tail-risk exposure and to design policy interventions (mandatory disclosure of dissent logs, diversity requirements for safety evaluations, independent oversight).
  • Empirical research directions:
    • Collect and analyze governance artifacts (internal reports, voting records, scenario libraries) to validate model predictions.
    • Measure effective input concentration (e.g., via inter-evaluator correlation, Herfindahl-style indices over pipelines/methods).
    • Test for illusory precision by relating observed disagreement to out-of-sample tail errors during realized regime shifts.

Assessment

Paper Typetheoretical Evidence Strengthn/a — This is a formal/theoretical paper that produces analytical results and testable diagnostics but does not present empirical estimation or causal inference from data. Methods Rigorhigh — Theoretical derivations are internally consistent, identify multiple channels and dynamic implications, and generate testable diagnostics; however, practical applicability depends on empirical mapping of model elements to organizational data and robustness to alternative behavioral assumptions. SampleNo empirical sample; uses an analytical signal aggregation model (and likely illustrative simulations or thought experiments) to generate theoretical predictions about organizational belief formation and failure modes. Themesgovernance org_design human_ai_collab IdentificationAnalytical signal-aggregation model that formalizes an organization as maintaining an anchor belief and achieving agreement through two exclusion channels (report shrinkage toward the anchor and a tolerance cutoff that discards outliers), with comparative counterfactuals to a full-inclusion benchmark to derive causal implications for bias, precision, and regime recognition. GeneralizabilityAbstract model may not capture full political, strategic, or incentive-driven behavior present in real organizations, Results depend on model assumptions (signal/noise distributions, specific anchor-update and filtering rules) that may not hold across contexts, Omitted features: heterogenous incentives, reputational effects, formal voting institutions, and asymmetric power that affect dissent expression, No cross-sector empirical validation; applicability across firm sizes, industries, and cultural contexts is uncertain, Translation to AI-specific settings (e.g., interaction of human judgment with model outputs) requires mapping model primitives to concrete AI workflows

Claims (10)

ClaimDirectionConfidenceOutcomeDetails
Organizational cohesion is observationally ambiguous: it can arise either from genuine information integration (debate and synthesis of heterogeneous inputs) or from exclusionary processes (conformity pressure, gatekeeping, intolerance of dissent). Organizational Efficiency mixed high source of observed cohesion (integration versus exclusion)
0.02
The paper formalizes the distinction using a signal-aggregation model in which an organization maintains an anchor belief and achieves agreement through two exclusion channels: (1) report shrinkage toward the anchor and (2) a tolerance rule that discards reports deviating beyond a threshold. Organizational Efficiency mixed high mechanisms producing agreement (report shrinkage, tolerance-based discarding)
0.02
Relative to a full-inclusion benchmark, exclusion-based cohesion produces state-contingent bias that is small in normal regimes but grows sharply under regime displacement (tail events). Decision Quality negative high estimation bias (especially under regime displacement/tail events)
state-contingent bias small in normal regimes, grows sharply in regime displacement
0.02
Exclusion-based cohesion induces 'illusory precision': observed disagreement can fall while actual estimation error in tail regimes rises (i.e., lower recorded variance despite higher true error). Decision Quality negative high observed disagreement (reported variance) versus true estimation error in tail regimes
illusory precision: lower observed variance despite higher true error in tail regimes
0.02
Exclusion leads to effective concentration of decision inputs: the effective number of independent inputs falls below the nominal participant count. Organizational Efficiency negative high effective number of independent decision inputs (information concentration)
effective number of independent inputs falls below nominal participant count
0.02
When the anchor belief is updated from internally filtered aggregates, the system can exhibit dynamic lock-in: delayed recognition of regime shifts followed by abrupt correction. Decision Quality negative high delay in regime recognition and magnitude/timing of corrective update
delayed recognition of regime shifts followed by abrupt correction
0.02
External inputs that bypass internal filtering shorten recognition delays (i.e., speed up detection of regime shifts). Decision Quality positive high time to recognize regime shift (recognition delay)
external bypass inputs shorten recognition delays
0.02
The model implies testable governance diagnostics linking latent fragility to observable patterns: recorded dissent (anonymous vs. formal voting gaps), scenario-set diversity, pipeline and method concentration, and anchor lag. Governance And Regulation positive medium observable diagnostics (recorded dissent patterns, voting gaps, scenario diversity, pipeline/method concentration, anchor lag) as indicators of latent fragility
0.01
Low internal conflict or unanimity can be diagnostic of variance depletion (i.e., exclusion) rather than healthy integration, so governance systems should treat low conflict as a potential red flag until heterogeneity integration is verified. Governance And Regulation negative medium internal conflict levels (observed dissent/unanimity) as indicator of variance depletion/exclusion
low conflict may indicate variance depletion/exclusion
0.01
Exclusion-based cohesion can produce state-contingent illusory precision together with effective input concentration and dynamic lock-in simultaneously—i.e., these phenomena co-occur under the model's parameter regimes. Decision Quality negative high co-occurrence of multiple adverse outcomes: tail bias, observed disagreement, effective input count, recognition delay
co-occurrence of tail bias, illusory precision, input concentration, lock-in under exclusion
0.02

Notes