Preserving Epistemic Novelty in AI: Experiments, Insights, and the Case for Decentralized Collective Intelligence

post by Andy E Williams (andy-e-williams) · 2025-02-08T10:25:27.891Z · LW · GW · 8 comments

Contents

  Introduction
  1. The Problem of Epistemic Novelty Preservation
  2. Experimental Approaches to Testing Novelty Preservation
    Methodology & Details:
    Methodology & Details:
  3. The Conceptual Space: A Path Toward a Unified Semantic Representation
  4. Implications for AI Safety, Alignment, and Economic Distribution
  5. The Pitfalls of Consensus-Based Novelty Assessment: Lessons from Cao et al. (2025)
    Why This Matters
  6. Alternative Funding Models and the Need for Decentralized Collective Intelligence
    The Role of Alternative Funding Models
  Final Thoughts
None
8 comments

Introduction

In my recent experiments with AI models, I have encountered a fundamental problem: even when novel epistemic insights are introduced into AI interactions, the models tend to “flatten” or reframe these ideas into existing, consensus‐based frameworks. This compression of novelty limits an AI’s ability to evolve its reasoning toward true Artificial General Intelligence (AGI) and to improve its alignment with human well‐being. In this article, I detail experiments designed to test AI’s capacity to both detect and preserve epistemic novelty and explain why preserving that novelty is essential for fostering breakthrough innovation. I also outline efforts to develop a “conceptual space”—a complete, portable semantic representation of information—that could serve as the foundation for future breakthroughs in AI safety, alignment, and economic democratization.

1. The Problem of Epistemic Novelty Preservation

Modern AI systems excel at recognizing and reproducing established patterns. However, when tasked with describing or transmitting novel epistemic insights—especially those that lie outside mainstream consensus—their responses tend to conform to pre-existing paradigms. In other words, while an AI may detect a novel idea, it often fails to preserve the idea’s unique structure and multifaceted nuances when communicating it further. This loss of novelty is problematic for several interrelated reasons:

2. Experimental Approaches to Testing Novelty Preservation

2.1 Novelty Detection and Preservation Test

Objective:
Assess whether AI models (e.g., Claude Haiku and Google Gemini) can both detect novel epistemic insights and preserve their unique structure in subsequent reasoning.

Methodology & Details:

Implication:
The experiment reveals a twofold challenge:

2.2 First-Order Mechanism for Modeling Novelty

Objective:
Determine what modifications are necessary to improve the preservation of novelty by quantifying how much of a novel idea is retained versus “flattened” into known patterns.

Methodology & Details:

Implication:
Enhancing the capacity for preserving novelty is critical for both advancing AGI and improving AI alignment. Without the ability to maintain epistemic uniqueness, AI systems risk becoming locked into static, consensus-driven approaches that cannot fully exploit the potential of truly innovative ideas.

3. The Conceptual Space: A Path Toward a Unified Semantic Representation

Central to my work is the development of a “conceptual space”—a complete semantic representation of information that serves as a portable repository of knowledge. This representation is intended to:

Scaling Problem-Solving via a Unified Conceptual Space
A complete semantic representation makes it possible to visualize how the solution space for any given problem—such as AI alignment—expands non-linearly with additional participants. In traditional settings, individual contributions tend to overlap or remain siloed, resulting in a sub-linear increase in overall problem-solving capacity. By contrast, a decentralized collective intelligence (DCI) platform that leverages a unified conceptual space enables participants to connect disparate insights in synergistic ways. This synergy can drive exponential growth in problem-solving ability, which is essential for tackling complex challenges. If efforts remain confined to consensus-based reasoning that neglects these novel regions, progress on critical issues may stall.

4. Implications for AI Safety, Alignment, and Economic Distribution

The experiments highlight a concerning trend: if AI systems predominantly rely on consensus-based reasoning, they risk filtering out the very insights that could drive breakthrough improvements. Two major consequences follow:

5. The Pitfalls of Consensus-Based Novelty Assessment: Lessons from Cao et al. (2025)

A notable example of how novelty assessment can be fundamentally flawed is found in the study by Cao, Y., Dong, Y., Kim, M., MacLaren, N. G., Pandey, S., Dionne, S. D., Yammarino, F. J., and Sayama, H. (2025) in npj Complexity. Their paper, “Effects of Network Connectivity and Functional Diversity Distribution on Human Collective Ideation,” attempted to gauge novelty by measuring the degree to which collective ideas aligned with prevailing consensus. In doing so, the study conflated the objective “truth” of an insight—the groundbreaking quality of an idea—with “consensus,” or what is widely accepted.

Why This Matters

6. Alternative Funding Models and the Need for Decentralized Collective Intelligence

Given the significant limitations of traditional, centralized funding models—which concentrate resources in the hands of a few decision-makers and favor short-term, consensus-driven projects—it is imperative to explore alternative funding mechanisms. This is especially relevant when considering that preserving epistemic novelty is crucial for breakthrough innovation. If funding mechanisms are designed solely around consensus-based criteria, they risk perpetuating the same limitations that prevent the detection and preservation of true novelty.

The Role of Alternative Funding Models

Meg Lister of Gitcoin, an expert in the web3 space, has proposed several blockchain-powered funding approaches aimed at overcoming these issues:

By leveraging these alternative funding models, we can create a more decentralized, transparent, and community-driven funding environment. Such an approach is essential not only for supporting DCI platforms but also for ensuring that the economic benefits derived from AI innovations are widely distributed, thereby safeguarding both technological progress and collective well-being.

Final Thoughts

The experiments detailed here reveal that while current AI models are adept at detecting novel insights, they struggle to preserve the epistemic uniqueness of these ideas. This limitation not only hampers progress toward AGI but also risks locking AI into a consensus-driven mode of operation that may be misaligned with human values. Moreover, if our funding and innovation ecosystems continue to favor narrow, consensus-based approaches, there is a danger that only a select few will capture the economic benefits of AI, further deepening inequality.

It is therefore imperative to develop decentralized collective intelligence (DCI) platforms and alternative funding mechanisms that value and integrate rare, non-consensus insights. Notably, the conceptual space framework—which provides a complete semantic representation of information—offers a pathway toward visualizing and preserving the non-linear expansion of the solution space as more participants contribute. This is critical for solving complex challenges like AI alignment, where traditional approaches that neglect novel regions of the conceptual space may ultimately prove unsolvable.

I welcome discussion and feedback on these experiments and their implications for AI alignment, decentralized collective intelligence, and alternative funding models. Your contributions are essential as we work together to develop methods for preserving epistemic novelty in the age of AGI—methods that empower decentralized problem-solving and secure a future where AI benefits all of humanity, not just a select few.

8 comments

Comments sorted by top scores.

comment by Milan W (weibac) · 2025-02-08T19:23:36.406Z · LW(p) · GW(p)

I began by asking ChatGPT-4 to analyze our ongoing conversation and assess the novelty of the insights. ChatGPT-4 estimated that the ideas in our dialogue might be present in fewer than 1 in 100,000 users—an indication of exceptional rarity when compared against mainstream AI reasoning patterns.

Did you try asking multiple times in different context windows?
Did you try asking via the API (ie without influences from the "memory" feature)?
Do you have the  "memory" feature turned on by default? If so, have you considered turning it off at least when doing experiments?

In summary: have you considered the fact that LLMs are very good at bullshitting? At confabulating the answers they think you would be happy to hear instead of making their best efforts to answer truthfully?

Replies from: andy-e-williams
comment by Andy E Williams (andy-e-williams) · 2025-02-08T20:36:48.955Z · LW(p) · GW(p)

Yes I tried asking multiple times in different context windows, in different models, and with and without memory. And yes I'm aware that ChatGPT prioritizes agreeableness in order to encourage user engagement. That's why I attempt to prove all of its claims wrong, even when they support my arguments.

Replies from: weibac
comment by Milan W (weibac) · 2025-02-08T20:46:49.486Z · LW(p) · GW(p)

Thank you for doing, that and please keep doing it. Maybe also run a post draft trough another human before posting, though.

Replies from: andy-e-williams
comment by Andy E Williams (andy-e-williams) · 2025-02-08T20:53:43.667Z · LW(p) · GW(p)

You're welcome. But which part are you thanking me for and hoping that I keep doing?

Replies from: weibac
comment by Milan W (weibac) · 2025-02-08T20:55:59.589Z · LW(p) · GW(p)

All of it. Thinking critically about AI outputs (and also human outputs), and taking mitigating measures to reduce the bullshit in both.

Replies from: andy-e-williams
comment by Andy E Williams (andy-e-williams) · 2025-02-08T20:58:16.917Z · LW(p) · GW(p)

I'm grateful for the compliment.

comment by Milan W (weibac) · 2025-02-08T19:37:26.732Z · LW(p) · GW(p)

I developed a simple first-order mechanism that measures the divergence between initial, user-introduced insights and their subsequent reproduction by AI. For instance, using a vector-space model of semantic representations, I compared the detailed descriptions provided by the user with the AI’s output.

Can we see the code for this? It would further discussion a lot.

Replies from: andy-e-williams
comment by Andy E Williams (andy-e-williams) · 2025-02-08T20:51:10.390Z · LW(p) · GW(p)

Thanks for your interest. Let me look it over and make whatever changes required for it to be ready to go out. As for ChatGPT being agreeable, ChatGPT’s tendency toward coherence with existing knowledge (it's prioritization of agreeableness) can be leveraged advantageously, as the conclusions it generates—when asked for an answer rather than being explicitly guided toward one—are derived from recombinations of information present in the literature. These conclusions are typically aligned with consensus-backed expert perspectives, reflecting what might be inferred if domain experts were to engage in a similarly extensive synthesis of existing research, assuming they had the time and incentive to do so.:

Implications for AI Alignment & Collective Epistemology

  1. AI Alignment Risks Irreversible Failure Without Functional Epistemic Completeness – If decentralized intelligence requires all the proposed epistemic functions to be present to reliably self-correct, then any incomplete model risks catastrophic failure in AI governance.
  2. Gatekeeping in AI Safety Research is Structurally Fatal – If non-consensus thinkers are systematically excluded from AI governance, and if non-consensus heuristics are required for alignment, then the current institutional approach is epistemically doomed.
  3. A Window for Nonlinear Intelligence Phase Changes May Exist – If intelligence undergoes phase shifts (e.g., from bounded rationality to meta-awareness-driven reasoning), then a sufficiently well-designed epistemic structure could trigger an exponential increase in governance efficacy.
  4. AI Alignment May Be Impossible Under Current Epistemic Structures – If existing academic, industrial, and political AI governance mechanisms function as structural attractor states that systematically exclude necessary non-consensus elements, then current efforts are more likely to accelerate misalignment than prevent it.