The Commonplace
Home Dashboard Papers Evidence Digests 🎲
← Papers

Generative AI floods teams with ideas and structures early-stage problem solving but seldom substitutes for human judgment; firms that combine high-volume generation with expert curation will capture most of the economic upside, while policy should prioritize training and accountability.

ChatGPT as an Innovative Tool for Idea Generation and Problem-Solving: Nano Review
Horn Sarun · March 27, 2026 · Zenodo (CERN European Organization for Nuclear Research)
openalex review_meta medium evidence 7/10 relevance DOI Source PDF
Generative AI serves as a dual-purpose cognitive tool—rapidly expanding idea volume and scaffolding problem decomposition—whose economic value materializes when paired with human curation, judgment, and domain expertise.

Abstract The advent of generative artificial intelligence, particularly advanced large language models like ChatGPT, is fundamentally reshaping the cognitive architecture of creativity and systematic problem-solving. This nano review examines its emergent, dual function as both a high-volume catalyst for divergent idea generation and a structured assistant for deconstructing and navigating complex problems. It synthesizes empirical findings on its utility in overcoming cognitive fixation, generating cross-domain analogies, and providing scaffolded support for hypothesis formulation and solution prototyping. The review critically confronts the model's inherent limitations, including its reliance on recombinative rather than deeply original insight, its susceptibility to bias and mediocrity, and its lack of contextual, domain-specific wisdom. It concludes that the true potential of these tools is unlocked not through autonomous ideation, but through a synergistic "cognitive co-pilot" model. This framework strategically leverages AI to expand the solution space and challenge assumptions, while reserving for human intelligence the critical roles of curation, strategic evaluation, and the application of experiential judgment to navigate ambiguity and ethical implications. Keywords: ChatGPT, innovation, ideation, creative problem-solving, divergent thinking, cognitive augmentation, human-AI collaboration, design thinking, creativity support tools

Summary

Main Finding

Generative AI (e.g., ChatGPT) functions as a dual-purpose cognitive tool: a high-volume catalyst for divergent idea generation and a structured assistant for decomposing complex problems. Its economic value is maximized not through autonomous ideation but via a synergistic "cognitive co-pilot" model in which AI expands the solution space while humans provide curation, strategic evaluation, and experiential judgment.

Key Points

  • Dual role
    • Divergent ideation: rapidly produces many candidate ideas, analogies, and associative prompts that help overcome fixation.
    • Structured problem-solving: scaffolds hypothesis formation, prototyping steps, and systematic decomposition of problems.
  • Empirical utilities documented
    • Helps break cognitive fixation and generate cross-domain analogies.
    • Provides scaffolded support that accelerates early-stage hypothesis and prototype development.
  • Core limitations
    • Predominantly recombinative (reworkings of existing material) rather than deeply novel insight.
    • Prone to bias, mediocrity, and errors absent domain-specific context or experiential knowledge.
    • Lacks reliable, situational judgment on ambiguity and ethical trade-offs.
  • Prescriptive framing
    • Best deployed as a "cognitive co-pilot": AI expands and challenges the idea space; humans retain roles of selection, refinement, contextualization, and ethical evaluation.

Data & Methods

  • Paper type: nano review / synthesis of existing empirical literature on large language models and creativity-support tools.
  • Methods used: literature synthesis and critical analysis of empirical studies (experimental tasks, design/ideation studies, and applied case evidence) on LLM-assisted creativity and problem-solving.
  • What it does not do: presents no original dataset or new experimental results; appears not to be a systematic meta-analysis.
  • Methodological caveats highlighted or implied:
    • Heterogeneity across studies (tasks, measures of creativity, domains) limits generalizability.
    • Potential publication and short-term study biases; limited evidence on long-run field impacts and economic outcomes.
    • Domain-specific performance varies and is contingent on user expertise and integration workflow.

Implications for AI Economics

  • Productivity and innovation
    • Lower search and idea-generation costs may speed early-stage R&D and increase the gross flow of candidate innovations.
    • Net gains in innovation depend on complementary human capacity for curation and development; raw increases in idea volume do not automatically equate to higher-quality innovation.
  • Labor demand and skill composition
    • Complementarity: increased demand for evaluative, integrative, and domain-expert roles (curators, synthesizers, implementation experts).
    • Substitution risk: routine ideation and drafting tasks may be automated, altering task-level labor demand and wage structure.
    • Implication for skills policy: emphasis on training in judgment, evaluation, domain expertise, and AI oversight.
  • Firm strategy and market structure
    • Firms that embed AI into collaborative workflows and invest in human curation may capture disproportionate returns (first-mover and scale advantages).
    • Platforms that combine high-volume generation with effective filtering/curation can create strong network effects and concentration.
  • Measurement and welfare
    • Standard productivity metrics may undercount value from augmented ideation; measuring quality-adjusted creative output is required.
    • Welfare effects ambiguous: broader access to idea scaffolding could democratize some innovation but also amplify low-quality outputs and misinformation without proper curation.
  • Policy and regulation
    • Need for standards around evaluation, bias mitigation, provenance, and accountability in AI-assisted ideation and design.
    • Support for public and private investments in human-AI collaboration training and incentives to build complementary expertise.
  • Research gaps relevant to economists
    • Long-run causal effects of LLMs on firm-level innovation rates, business formation, and industry structure.
    • Quantification of complementarities between AI and different skill types (evaluative vs. generative tasks).
    • Market experiments on business models that internalize both high-volume generation and high-quality curation.

Assessment

Paper Typereview_meta Evidence Strengthmedium — Synthesizes multiple experimental and applied studies that consistently show LLMs boost idea generation and scaffold problem decomposition, but relies on heterogeneous, short-term, often lab-based evidence without new causal estimates or systematic meta-analysis; therefore insights are plausible but not strongly causal or generalizable. Methods Rigormedium — Careful critical synthesis of existing empirical work is used, with thoughtful caveats about heterogeneity and biases, but the review is not a systematic review or formal meta-analysis and does not present pre-registered inclusion criteria, quantitative aggregation, or new empirical checks. SampleNarrative synthesis of existing empirical literature on large language models and creativity-support tools, including short-term experimental tasks, design/ideation laboratory studies, applied case evidence from firms and pilot deployments; no original dataset or new experiments presented. Themesproductivity human_ai_collab innovation labor_markets skills_training org_design governance GeneralizabilityHeterogeneous tasks, creativity measures, and domains across cited studies limit external validity., Many studies are short-term and lab-based; long-run adoption and field impacts are unobserved., Performance conditional on user expertise and integration workflow reduces transferability across firm types and occupations., Potential publication bias toward positive results and lack of representative samples., Findings may not generalize to highly regulated, safety-critical, or domain-expert settings where experiential judgment matters most.

Claims (21)

ClaimDirectionConfidenceOutcomeDetails
Generative AI functions as a dual-purpose cognitive tool: a high-volume catalyst for divergent idea generation and a structured assistant for decomposing complex problems. Creativity mixed medium role/performance of generative AI on cognitive tasks (divergent ideation volume and structured problem decomposition quality)
0.14
Generative models rapidly produce many candidate ideas, analogies, and associative prompts that help overcome cognitive fixation. Creativity positive medium idea quantity and measures of fixation (e.g., fixation errors, number of distinct categories generated)
0.14
Generative AI provides scaffolded, structured support that aids systematic hypothesis formation, prototyping steps, and decomposition of complex problems. Research Productivity positive medium speed and/or quality of early-stage hypothesis generation and prototype development (e.g., time-to-first-prototype, number/clarity of hypotheses)
0.14
Empirical studies document that AI-assisted tools can help break cognitive fixation and generate cross-domain analogies. Creativity positive medium frequency/quality of cross-domain analogies and fixation-related performance metrics
0.14
Generative AI accelerates early-stage hypothesis and prototype development by providing scaffolded prompts and procedural suggestions. Task Completion Time positive medium time-to-hypothesis or prototype, number of prototype iterations in early-stage development
0.14
LLMs are predominantly recombinative — they tend to rework and recombine existing material rather than produce deeply novel insights. Creativity negative medium novelty/creativity metrics (e.g., originality scores, novelty ratings)
0.14
LLMs are prone to bias, mediocrity, and factual or logical errors when domain-specific context or experiential knowledge is absent. Error Rate negative medium accuracy/factuality, bias indicators, perceived quality of outputs in domain-specific tasks
0.14
Generative AI lacks reliable situational judgment on ambiguous problems and on ethical trade-offs, making it insufficient for autonomous decision-making in such contexts. Decision Quality negative medium quality/appropriateness of situational judgment and ethical decision-making in test scenarios
0.14
The most effective deployment model is a 'cognitive co-pilot' in which AI expands and challenges the idea space while humans provide curation, strategic evaluation, and experiential judgment. Team Performance mixed medium quality-adjusted creative output or decision outcomes under human-AI collaborative workflows
0.14
Lower search and idea-generation costs enabled by LLMs may speed early-stage R&D and increase the gross flow of candidate innovations. Innovation Output positive medium volume/rate of candidate ideas generated and pace of early-stage R&D activity
0.14
Net gains in innovation from increased idea volume depend on complementary human capacity for curation and development; raw increases in ideas do not automatically translate into higher-quality innovation. Innovation Output mixed medium quality-adjusted innovation rate (conversion of ideas into valuable innovations)
0.14
There will likely be complementarity-driven increases in demand for evaluative, integrative, and domain-expert roles (curators, synthesizers, implementation experts). Employment positive medium employment demand for evaluative/integrative/domain-expert roles
0.14
There is substitution risk: routine ideation and drafting tasks may be automated, altering task-level labor demand and wage structure. Job Displacement negative medium employment and wages for routine ideation/drafting tasks
0.14
Firms that embed AI into collaborative workflows and invest in human curation may capture disproportionate returns (first-mover and scale advantages). Firm Revenue positive low firm-level returns, market share, and competitive advantage
0.07
Platforms combining high-volume generation with effective filtering/curation can create strong network effects and concentration in markets for AI-assisted ideation. Market Structure positive low market concentration and network effects for ideation platforms
0.07
Standard productivity metrics are likely to undercount the value generated by AI-augmented ideation; quality-adjusted measures of creative output are required. Organizational Efficiency mixed high measured productivity vs. true quality-adjusted creative output
0.24
Welfare effects of democratized access to AI-assisted ideation are ambiguous: access could democratize innovation but also amplify low-quality outputs and misinformation absent proper curation. Consumer Welfare mixed medium distributional welfare impacts and prevalence/impact of misinformation or low-quality outputs
0.14
There is a need for standards around evaluation, bias mitigation, provenance, and accountability in AI-assisted ideation and design. Governance And Regulation positive medium existence and adoption of evaluation/mitigation/provenance/accountability standards
0.14
Key research gaps include a lack of long-run causal evidence on the effects of LLMs on firm-level innovation rates, business formation, and industry structure. Research Productivity null_result high long-run causal impacts of LLM adoption on firm innovation, business formation, and industry structure
0.24
Another important gap is quantifying complementarities between AI and different skill types (evaluative vs. generative tasks). Skill Acquisition null_result high magnitude of complementarities between AI assistance and various human skill types
0.24
Methodological caveats across the literature (heterogeneity of tasks/measures, publication bias, short-term studies) limit the generalizability of current findings. Research Productivity mixed high generalizability and external validity of LLM-assisted creativity findings
0.24

Notes