The Commonplace
Home Dashboard Papers Evidence Digests 🎲
← Papers

As AI takes over routine speed and pattern-matching, success should be redefined around distinctly human capacities—empathy, judgement and imagination—which firms should cultivate and reward rather than measured productivity alone.

Deconstructing success: why being human still matters
Andrew Woods, Benton Jones · March 11, 2026 · Journal of Ethics in Entrepreneurship and Technology
openalex theoretical n/a evidence 7/10 relevance DOI Source PDF
The paper argues that as AI automates routine tasks, success should be redefined around distinctively human capacities—empathy, moral judgment, creativity and adaptive purpose—which become the strategic comparative advantage for individuals and organizations.

Purpose Artificial intelligence (AI) has redefined what it means to perform, achieve and succeed. Algorithms now surpass human capability in processing speed, pattern recognition and data-driven decision-making. However, as machines become increasingly intelligent, the question of what constitutes success in the human sense becomes increasingly important. The purpose of this paper is to provide a framework for evaluation of these intersecting concepts. Design/methodology/approach Drawing on leadership theory, emotional intelligence research and AI ethics, “Deconstructing success” involves dismantling productivity-based definitions and reconstructing a framework centered on adaptability and purpose. In an age of automation, being human is not a disadvantage; it is a defining strategic advantage. Findings This paper argues that the future of success will not depend on outpacing machines but on cultivating distinctly human capacities: empathy, discernment, imagination and moral reasoning. Originality/value This conceptual essay proposes the Human Excellence 2.0 model, positioning human consciousness and ethical awareness as the new frontier of achievement.

Summary

Main Finding

The paper argues that in an era of advanced AI, success should be redefined away from pure productivity metrics toward a model called Human Excellence 2.0, which treats distinctively human capacities — empathy, discernment, imagination and moral reasoning — as the strategic advantage. Rather than competing to outpace machines on speed or pattern-matching, individuals and organizations should cultivate adaptability and purpose as the criteria for achievement.

Key Points

  • Current productivity-based definitions of success (speed, output, efficiency) are inadequate when algorithms outperform humans on many tasks.
  • The paper synthesizes leadership theory, emotional intelligence research and AI ethics to critique productivity-centric frameworks.
  • Human Excellence 2.0 centers on capacities that are hard for current AI to replicate: empathy, moral reasoning, judgement/discernment, creativity/imagination, and adaptive purpose-driven thinking.
  • “Being human” becomes a comparative advantage: relational, ethical and imaginative skills complement automated systems rather than compete with them.
  • The framework is conceptual and normative — it seeks to reconstruct how success is evaluated rather than to test empirical claims.
  • Practical implication: organizations and institutions should reorient evaluation, hiring, training and leadership development toward these human-centered capacities.

Data & Methods

  • Design: conceptual essay / theoretical synthesis.
  • Sources: literature from leadership studies, emotional intelligence, and AI ethics; no primary empirical data are presented.
  • Methodological approach: deconstruction of productivity-based definitions and reconstruction of a normative framework for success focused on adaptability and purpose.
  • Limitations: absence of operational measures, empirical testing, or causal evidence; claims are theoretical and require empirical validation.

Implications for AI Economics

  • Labor-market comparative advantage

    • Tasks requiring empathy, moral judgement, and creative insight are likely to see increased relative demand and wage premia as routine tasks are automated.
    • The paper implies a shift in comparative advantage toward workers who cultivate nonroutinized social and cognitive skills.
  • Complementarity vs. substitution

    • AI will substitute for many routine cognitive and manual tasks, but will complement workers in roles emphasizing human capacities; firms should reassign AI to task bundles where it enhances human-led decision-making.
  • Human capital and education policy

    • Investment priorities should shift toward training in emotional intelligence, ethics, creativity and adaptive learning (lifelong learning, interdisciplinary curricula).
    • Credentialing and measurement systems will need to evolve to certify these capacities.
  • Measurement and productivity accounting

    • Standard productivity metrics and GDP accounting may undercount value created by empathy, moral leadership and imagination; new measurement approaches are required to capture these nonstandard outputs.
  • Firm strategy and organizational design

    • Hiring, promotion and performance evaluation should incorporate metrics for adaptability, ethical judgement and relational skills; role design should pair AI tools with human responsibilities that require discretion and purpose.
    • Incentive structures may need redesigning to value long-term, purpose-driven outcomes over short-term quantifiable outputs.
  • Distributional effects and inequality

    • Two-sided risk: those who acquire human-centric skills could capture significant returns, while those displaced from routinized occupations may face persistent disadvantage without effective retraining.
    • Policy interventions (retraining subsidies, transition supports) will shape distributional outcomes.
  • Wages and returns to skills

    • Expect changes in skill-biased technological change: increased returns to social, creative and moral skills rather than solely abstract cognitive or technical skills.
    • Empirical analysis is needed to quantify magnitude and timing.
  • Regulation and governance

    • Organizations and regulators will need standards for ethical oversight, accountability and certification of human roles that supervise or augment AI systems.
  • Research agenda for AI economists

    • Operationalize and measure Human Excellence 2.0 skills (surveys, performance metrics, behavioral assessments).
    • Task-level analyses (O*NET-type mappings) to identify occupations with high complementarity to AI.
    • Wage regressions and employer surveys to estimate returns to empathy/ethics/creativity across AI-exposed industries.
    • Field experiments and RCTs on training programs that teach these capacities and measure labor-market outcomes.
    • Structural and macro models to project impacts on growth, employment composition and welfare.
  • Cautions

    • The argument is normative and conceptual; empirical validation is required.
    • There is risk in romanticizing human skills — some “human” capacities could be partially automated or modeled by future AI.
    • Policy design must consider heterogeneity in access to training and the time lags for skill acquisition.

Overall, the Human Excellence 2.0 framework reframes economic questions about AI from “which tasks will be automated?” to “what uniquely human capacities should economies build and reward?” For AI economists, this suggests shifting empirical focus to measurement of social, ethical and creative skills, estimating their returns, and evaluating policies that foster complementarity between humans and machines.

Assessment

Paper Typetheoretical Evidence Strengthn/a — The paper is a conceptual and normative synthesis without primary empirical data, causal tests, or quantitative estimates, so it does not provide empirical evidence that can be graded. Methods Rigormedium — The argument draws on and integrates literature from leadership studies, emotional intelligence, and AI ethics in a coherent conceptual framework, but it lacks operationalization, formal modeling, systematic review methods, or empirical validation that would raise rigor to high. SampleNo primary data; draws on secondary literature from leadership theory, emotional intelligence research, and AI ethics (qualitative and theoretical sources), with illustrative examples rather than systematic empirical samples. Themeshuman_ai_collab skills_training labor_markets GeneralizabilityNormative, context-dependent claims not empirically validated across industries or countries, Assumes present-day AI capabilities; future AI improvements could change which capacities are distinctive, Cultural and institutional variation in valuing ‘human’ capacities (empathy, moral judgment) limits cross-country generalizability, Vague constructs (e.g., ‘adaptive purpose’) lack standardized measurement, complicating transferability, Time-horizon uncertainty: short-run vs long-run dynamics and transition costs are not quantified

Claims (8)

ClaimDirectionConfidenceOutcomeDetails
Artificial intelligence (AI) has redefined what it means to perform, achieve and succeed. Other mixed medium definition/criteria of 'success' (conceptual)
0.01
Algorithms now surpass human capability in processing speed, pattern recognition and data-driven decision-making. Other positive medium processing speed, pattern recognition capability, data-driven decision-making performance
0.01
As machines become increasingly intelligent, the question of what constitutes success in the human sense becomes increasingly important. Other mixed low perceived importance of 'human' criteria for success (conceptual)
0.01
Productivity-based definitions of success should be dismantled and reconstructed into a framework centered on adaptability and purpose. Other positive low formulation of success frameworks emphasizing adaptability and purpose (conceptual)
0.01
In an age of automation, being human is not a disadvantage; it is a defining strategic advantage. Other positive speculative strategic advantage conferred by human traits in automated contexts (conceptual)
0.0
The future of success will not depend on outpacing machines but on cultivating distinctly human capacities: empathy, discernment, imagination and moral reasoning. Other positive low future success (as determined by cultivation of specific human capacities)
0.01
This paper proposes the Human Excellence 2.0 model, positioning human consciousness and ethical awareness as the new frontier of achievement. Other positive speculative conceptual model components: human consciousness and ethical awareness as determinants of achievement
0.0
Drawing on leadership theory, emotional intelligence research and AI ethics informs the proposed framework. Other null_result high sources informing the framework (theoretical influences)
0.02

Notes