Knowledge, Reasoning, and Superintelligence

post by owencb · 2025-03-26T23:28:11.465Z · LW · GW · 0 comments

This is a link post for https://strangecities.substack.com/p/knowledge-reasoning-and-superintelligence

Contents

  Crystallized intelligence vs fluid intelligence
  The knowledge production loop
  Crystallized intelligence and fluid intelligence in concrete systems
    AlphaGo
    Current LLMs
  The future
    Crystallized knowledge will remain relevant
    The fluid intelligence enhancement loop
  So what?
None
No comments

What makes people good at solving novel problems? In part, it’s having relevant knowledge. And in part, it’s being able to think flexibly to apply that knowledge. So intelligence isn't a scalar: there are at least two dimensions — often called crystallized intelligence and fluid intelligence.

This is also broadly true of AI. We can see the distinction between language models regurgitating facts and reasoning models tracing out implications. We can see the difference between the performance of AlphaGo’s policynet, and what it achieves with Monte Carlo tree search. And we can imagine this distinction even for superintelligent systems.

This is a conceptual discussion. But the motivation is pragmatic. If we have a better understanding of the high-level workings of intelligence, we can make better predictions about what is required for superhuman performance, or what the dynamics of an intelligence explosion might look like. And this could also help us in thinking about how AI systems should be governed — e.g. should we be giving more attention to questions of who controls the creation and distribution of new knowledge?

Crystallized intelligence vs fluid intelligence

For the purposes of this post, intelligence is the ability to process information in ways that are helpful for achieving goals. Intelligent systems typically have both some crystallized intelligence (a body of implicit or explicit knowledge relevant to the task) and some fluid intelligence (being able to think things through on the fly).

In humans, knowing your multiplication tables, knowing how to ride a bike, or playing bullet chess are all chiefly about crystallized intelligence. But proving novel theorems, designing a bicycle from first principles, or playing slower games of chess all depend more on fluid intelligence.

Of course even in these cases crystallized intelligence is an important ingredient for the ability to think things through — e.g. mathematicians are more likely to be able to prove novel theorems if they have a solid grasp of the existing concepts and theorems, rather than needing to rederive them each time. And since our thinking often depends on our lower-level concepts and heuristics, it’s difficult to give clean definitions separating them.

For AI systems, we can be a little more precise. But first we’ll make a couple of related definitions:

The knowledge production loop

Crystallized intelligence depends crucially on knowledge. But where does the knowledge come from? Of course it may be exogenous to the system (e.g. LLMs functioning as crystallized intelligence based on human-produced data). But something interesting happens in cases where the system can produce its own new knowledge/data — via having new ideas, or via taking actions and observing the results. This gives rise to a loop where improved knowledge leads to stronger crystallized intelligence (and perhaps stronger fluid intelligence), which in turn can produce new higher-quality knowledge:

This is, in essence, a knowledge production engine. It isn’t a closed system — it requires compute to fuel the engine. But with sufficient fuel it may lead to large improvements in performance.

This is perhaps the simplest form of “recursive improvement” — subsequent iterations build on the improvement from earlier steps. And we’ve already seen cases where this kind of engine can produce significantly superhuman systems.

Crystallized intelligence and fluid intelligence in concrete systems

AlphaGo

The creation of the AlphaGo system has steps corresponding to a version of this loop:

  1. It starts with a database consisting of many expert human games
  2. The initial policy network and value network function as a form of crystallized knowledge — they provide useful information giving generalized approximations to what good play looks like
  3. Monte Carlo tree search is a form of thinking algorithm — it can be used to search for good moves (making use of inference compute to make calls to crystallized intelligence to help direct its search)
  4. It uses self-play to produce many decision-situations to analyse
  5. It considers and takes actions in these self-play games, and observes the output (i.e. which side wins)
  6. This creates many games (eventually with superhuman performance), which are added to the knowledge base
  7. This improves the policy network and value network in subsequent iterations — and hence the performance of the fluid intelligence

The successor system AlphaZero completes this same loop, but without the initial knowledge base — just the rules of the game (Go or Chess or Shogi) permitting it to make observations. Interestingly, although it builds up a suite of implicit concepts (overlapping with human concepts, but containing new ones), in contrast to the development of the human corpus of expertise, these concepts are purely implicit — they live only in the neural net representing crystallized intelligence, rather than explicit “new ideas” which live in the knowledge base.

Current LLMs

Current LLMs are borderline on completing the knowledge creation loop:

The LLMs of early 2025 are generally strong at crystallized intelligence. They have access to a lot of the data that human civilization has produced, including implicit knowledge about what best practices mean in many many situations.

LLMs are able to apply a certain amount of fluid intelligence, via chain of thought (or scaffolding). At the moment of writing, they’re not always very good at it, but they’re getting better fairly quickly — today’s reasoning models are much stronger than those of a year ago. Some of that strength flows from the kind of knowledge production loop marked:

On the other hand, LLMs aren’t yet at the edge of being able to produce new knowledge which is very helpful by the standards of human civilization. Compared to AlphaGo-like systems, this recursive loop appears to be hamstrung by the lack of clear feedback, and by the breadth of the search space. So it can only improve performance so far — for now.

The future

The distinction between crystallized and fluid intelligence is still meaningful for systems which are superhuman — here are some conceivable (but unrealistic) AI systems which pull out the distinctions:

Crystallized knowledge will remain relevant

It is quite plausible that the basic knowledge production loop could give us transformative AI without any kind of strongly superhuman fluid intelligence. This would involve AI being leveraged to figure out many more things, at scale — the “country of geniuses in a datacentre” — and then the things that it figures out will be the new knowledge worth crystallizing.

Moreover, crystallized knowledge is likely to remain important even for strongly superintelligent systems. Even if you could figure everything out from first principles every time, it’s much less efficient than using cached knowledge [2]— similarly, the first transformative superintelligent systems almost certainly wouldn’t be meaningfully superintelligent if they did have to figure everything out each time.

The fluid intelligence enhancement loop

While the knowledge production loop may drive a large increase in knowledge (and hence crystallized intelligence), classical thinking about an intelligence explosion is more centrally about improvements to fluid intelligence. We might imagine something like this loop:

Even in this case, it seems likely that a lot of the improvements will be driven by the more traditional knowledge production loop — a system might iterate building new knowledge on top of new knowledge for many steps before it reaches an improvement to the algorithms that represents a fundamental upgrade to its fluid intelligence. But the possibility of this further type of upgrade increases the explosive potential of recursive improvement, compared to automated research without this possibility.

So what?

This post has been an exploration of conceptual foundations for thinking about intelligence, and the future of AI. Where can this take us? Ultimately I don’t know; I’m offering up the thinking in part in the hope that others may find useful applications beyond those I’ve considered. But here are some questions that I find my attention drawn towards:

Thanks to Raymond Douglas, Lizka Vaintrob, Tom Davidson, Rudolf Laine, and Andreas Stuhlmüller for helpful comments, and to many more people for helpful conversations.

  1. ^

    Perhaps it would be better to describe things in terms of some ELO-like score, which has a more natural crisp meaning; however that would require a choice of game, and I don’t in this context care enough to try to nail it down.

  2. ^

    This point is not original; Beren makes a similar point about direct and amortized optimization in individual systems [LW · GW], and Rudolf Laine makes an analogous point about amortized optimization at the level of civilization [LW · GW].

0 comments

Comments sorted by top scores.