Posts

What is the best way to talk about probabilities you expect to change with evidence/experiments? 2024-04-19T15:35:47.779Z
What is the nature of humans general intelligence and it's implications for AGI? 2024-03-26T15:20:39.860Z
Will_Pearson's Shortform 2024-01-04T16:52:40.565Z
Should rationalists (be seen to) win? 2023-08-29T18:13:09.629Z
Summary of Improving Global Decision Making (around AI) 2023-08-05T18:46:44.268Z

Comments

Comment by Will_Pearson on Will_Pearson's Shortform · 2024-04-24T23:08:40.205Z · LW · GW

I was thinking of having evals that controlled deployment of LLMs could be something that needs multiple stakeholders to agree upon.

Butt really it is a general use pattern.

Comment by Will_Pearson on Will_Pearson's Shortform · 2024-04-24T16:39:35.636Z · LW · GW

Agreed code as coordination mechanism

Code nowadays can do lots of things, from buying items to controlling machines. This presents code as a possible coordination mechanism, if you can get multiple people to agree on what code should be run in particular scenarios and situations, that can take actions on behalf of those people that might need to be coordinated.

This would require moving away from the “one person committing code and another person reviewing” code model. 

This could start with many people reviewing the code, people could write their own test sets against the code or AI agents could be deputised to review the code (when that becomes feasible). Only when an agreed upon number of people thinking the code should it be merged into the main system.

Code would be automatically deployed, using gitops and the people administering the servers would be audited to make sure they didn’t interfere with running of the system without people noticing.

Code could replace regulation in fast moving scenarios, like AI. There might have to be legal contracts that you can’t deploy the agreed upon code or use the code by itself outside of the coordination mechanism.


 

Comment by Will_Pearson on Express interest in an "FHI of the West" · 2024-04-19T10:15:55.818Z · LW · GW

As well as thinking about the need for the place in terms of providing a space for research, it is probably worth thinking about the need for a place in terms of what it provides the world.  What subjects are currently under-represented in the world and need strong representation to guide us to a positive future? That will guide who you want to lead the organisation.

Comment by Will_Pearson on Deontic Explorations In "Paying To Talk To Slaves" · 2024-04-11T22:02:27.105Z · LW · GW

I admit that it is extreme circumstances that would make slavery consensual and justified. My thinking was if existential risk was involved, you might consent to slavery to avert it. It would have to be a larger entity than a single human doing the enslaving, because I think I agree that individuals shouldn't do consequentialism. Like being a slave to the will of the people, in general. Assuming you can get that in some way.

I don't follow the reasoning here

So let's say the person has given up autonomy to avert existential risk, they should perhaps get something in return. Maybe they get influence, but they can't use influence for their own benefit (as one of the deontological rules stipulates that is disallowed). So they are stuck trying to avert existential risk with no pay off. If you unenslave them you remove the will of the people's voice and maybe increase existential risk or s risks.

Hmm, sorry went off on a bit of tangent here. All very unlikely agreed.

Comment by Will_Pearson on Deontic Explorations In "Paying To Talk To Slaves" · 2024-04-11T19:00:18.825Z · LW · GW

Tangential but is there ever justified unconscious slavery. For example if you asked whether you consent to slavery and then your mind wiped, might you get into a situation where the slave doesn't know they consented to it, but the slave master is justified in treating them like a slave.

You would probably need a justification for the master slave relationship. Perhaps it is because it needs to be hidden for a good reason? Or to create a barrier against interacting with the ethical. In order to dissolve such slavery, understanding the justifications for why the slavery started would be important.

Comment by Will_Pearson on Will_Pearson's Shortform · 2024-04-10T09:55:30.891Z · LW · GW

Proposal for new social norm - explicit modelling

Something that I think would make rationalists more effective at convincing people is if we had explicit models of the things we care about.

Currently we are at the stage of physicists arguing that the atom bomb might ignite the atmosphere without concrete math and models of how that might happen.

If we do this for lots of issues and have a norm of making models composable this would have further benefits.
 

  • People would use the models to make real world decisions with more accuracy
  • We would create frameworks for modelling that would be easily composable, that other people would use

Both would raise the status and knowledge of the rationalist community.

Comment by Will_Pearson on Vanessa Kosoy's Shortform · 2024-04-08T16:00:16.791Z · LW · GW

Does it make sense to plan for one possible world or do you think that the other possible worlds are being adequately planned for and it is only the fast unilateral take off that is neglected currently?

Limiting AI to operating in space makes sense. You might want to pay off or compensate all space launch capability in some way as there would likely be less need.

Some recompense for the people who paused working on AI or were otherwise hurt in the build up to AI makes sense.

Also trying to communicate ahead of time what a utopic vision of AI and humans might look like, so the cognitive stress isn't too major is probably a good idea to commit to.

Committing to support multilateral acts if unilateral acts fail is probably a good idea too. Perhaps even partnering with a multilateral effort so that effort on shared goals can be spread around?

Comment by Will_Pearson on Will_Pearson's Shortform · 2024-04-08T10:46:14.038Z · LW · GW

Relatedly I am thinking about improving the wikipedia page on recursive self-improvement. Does anyone have any good papers I should include? Ideally with models.

Comment by Will_Pearson on Will_Pearson's Shortform · 2024-04-07T11:43:20.900Z · LW · GW

I'm starting a new blog here. It is on modelling self-modifying systems, starting with AI. Criticisms welcome 

Comment by Will_Pearson on What is the nature of humans general intelligence and it's implications for AGI? · 2024-03-26T20:57:42.009Z · LW · GW

I'm wary about that one, because that isn't a known "general" intelligence architecture, so we can expect AIs to make better learning algorithms for deep neural networks, but not necessarily themselves.

Comment by Will_Pearson on What is the nature of humans general intelligence and it's implications for AGI? · 2024-03-26T19:35:50.507Z · LW · GW

I'd like to see more discussion of this, I read some of the FOOM debate but I'm assuming that there has been more discussion of this important issue since?

I suppose the key question is for recursive self-improvement. We can give hardware improvement (improved hardware allows design of more complex and better hardware) because we are on the treadmill already.  But how likely is algorithmic self-improvement.  For an intelligence to be able to improve itself algorithmically the following seem to need to hold.

  1. The system needs to understand itself
  2. There has to be some capacity that can be improved without detriment to some other capacity (else you are doing some self-optimization and not necessarily improvement)
     

If it is the memeplex that gives us our generality (as is suggested by our flowering of discovery over the past 250 years compared to the past 300,000 years of homo sapiens), it might not be understandable. It would be in the weights or equivalents in whatever the AI uses. No human would understand it either.

Fiddling about with weights without knowledge would likely lead to trade offs and so you might not have the second consideration holding.

I'm not saying AI won't change history, but we need an accurate view of how it will change things.

Comment by Will_Pearson on Will_Pearson's Shortform · 2024-01-04T16:52:42.029Z · LW · GW

Found "The Future of Man" by Pierre Teilhard de Chardin in a bookshop. Tempted to wite a book review. It discusses some interesting topics, like the planetisation of Mankind. However it treats them as inevitable, rather as something contingent on us getting our act together. Anyone interested in a longer review?

Edit: I think his faith in the super natural plays a part in the assumption of inevitability.

Comment by Will_Pearson on Should rationalists (be seen to) win? · 2023-08-29T18:45:17.645Z · LW · GW

That's true. Communities that can encourage truth speaking and exploration will probably get more of it and be richer for it in the long term.

Comment by Will_Pearson on Conditional on living in a AI safety/alignment by default universe, what are the implications of this assumption being true? · 2023-07-17T22:27:42.853Z · LW · GW

Not really part of the lesswrong community at the moment, but I think evolutionary dynamics will be the next thing.

Not just of AI, but post humans, uploads etc. Someone will need to figure out what kind of selection pressures the should be so that things don't go to ruin in an explosion of variety.

Comment by Will_Pearson on Markets are Anti-Inductive · 2009-02-27T15:04:09.000Z · LW · GW

All competitive situations against ideal learning agents are anti inductive in this sense. Because they can note regularities in their actions and avoid them in the future as well as you can note regularities in their actions and exploit them. The usefulness of induction is based on the relative speeds of the induction of the learning agents.

As such anti induction appears in situations like bacterial resistance to antibiotics. We spot a chink in the bacterias armour, and we can predict that that chink will become less prevalent and our strategy less useful.

So I wouldn't mark markets as special, just the most extreme example.

Comment by Will_Pearson on Wise Pretensions v.0 · 2009-02-21T13:01:39.000Z · LW · GW

I find neither that convincing. Justice is not a terminal value for me, so I might sacrifice it for Winning. I prefered reading the first, but that is no indication of what a random person may prefer.

Comment by Will_Pearson on Pretending to be Wise · 2009-02-19T23:19:19.000Z · LW · GW

With international affairs, isn't stopping the aggression the main priority? That is stopping the death and suffering of humans on both sides? Sure it would be good to punish the aggressors rather than the retaliators but if that doesn't stop the fighting it just means more people are dying.

Also there is a difference between the adult and the child, the adult relies on the law of the land for retaliation the child takes it upon himself when he continues the fight. That is the child is a vigilante, and he may punish disproportionately e.g. breaking a leg for a dead leg.

Comment by Will_Pearson on Three Worlds Decide (5/8) · 2009-02-03T11:50:20.000Z · LW · GW

I don't really have a good enough grasp on the world to predict what is possible, it all seems to unreal.

One possibility is to jump one star away back towards earth and then blow up that star, if that is the only link to the new star.

Comment by Will_Pearson on The Baby-Eating Aliens (1/8) · 2009-01-31T01:11:28.000Z · LW · GW

Re: "MST3K Mantra"

Illustrative fiction is a tricky business, if this is to be part of your message to the world it should be as coherent as possible, so you aren't accidentally lying to make a better story.

If it is just a bit of fun, I'll relax.

Comment by Will_Pearson on The Baby-Eating Aliens (1/8) · 2009-01-30T19:24:13.000Z · LW · GW

I wonder why the babies don't eat each other. There must be a huge selective pressure to winnow down your fellows to the point where you don't need to be winnowed. This would in turn select for small brained, large and quick growing at the least. There might also be selective pressure to be partially distrusting of your fellows (assuming there was some cooperation), which might follow over into adulthood.

I also agree with the points Carl raised. It doesn't seem very evolutionarily plausible.

Comment by Will_Pearson on Value is Fragile · 2009-01-29T10:32:01.000Z · LW · GW

"Except to remark on how many different things must be known to constrain the final answer."

What would you estimate the probability of each thing being correct is?

Comment by Will_Pearson on Failed Utopia #4-2 · 2009-01-24T15:05:00.000Z · LW · GW

Reformulate to least regret after a certain time period, if you really want to worry about the resource usage of the genie.

Comment by Will_Pearson on Investing for the Long Slump · 2009-01-22T10:37:43.000Z · LW · GW

Personally I believe in the long slump. However I believe in human optimisim that will make people rally the market every so often. The very fact that most people believe the stock market will rise, will make it rise at least once or twice before people start to get the message that we are in the long slump.

Comment by Will_Pearson on Failed Utopia #4-2 · 2009-01-22T01:15:00.000Z · LW · GW

Eliezer, didn't you say that humans weren't designed as optimizers? That we satisfice. The reaction you got is probably a reflection of that. The scenario ticks most of the boxes humans have, existence, self-determination, happiness and meaningful goals. The paper clipper scenario ticks none. It makes complete sense for a satisficer to pick it instead of annihilation. I would expect that some people would even be satisfied by a singularity scenario that kept death as long as it removed the chance of existential risk.

Comment by Will_Pearson on Failed Utopia #4-2 · 2009-01-22T00:39:19.000Z · LW · GW

Dognab, your arguments apply equally well to any planner. Planners have to consider the possible futures and pick the best one (using a form of predicate), and if you give them infinite horizons they may have trouble. Consider a paper clip maximizer, every second it fails to use its full ability to paper clip things in its vicinity it is losing possible useful paper clipping energy to entropy (solar fusion etc). However if it sits and thinks for a bit it might discover a way to hop between galaxies with minimal energy. So what decision should it make? Obviously it would want to run some simulations, see if there gaps in its knowledge. How detailed simulations should it make, so it can be sure it has ruled out the galaxy hopping path?

I'll admit I was abusing the genie-trope some what. But then I am sceptical of FOOMing anyway, so when asked to think about genies/utopias, I tend to suspend all disbelief in what can be done.

Oh and belldandy is not annoying because she has broken down in tears (perfectly natural), but because she bases her happiness too much on what Stephen Grass thinks of her. A perfect mate for me would tell me straight what was going on and if I hated her for it (when not her fault at all), she'd find someone else because I'm not worth falling in love with. I'd want someone with standards for me to meet, not unconditional creepy fawning.

Comment by Will_Pearson on Failed Utopia #4-2 · 2009-01-21T16:04:39.000Z · LW · GW

Bogdan Butnaru:

What I meant was is that the AI would keep inside it a predicate Will_Pearson_would_regret_wish (based on what I would regret), and apply that to the universes it envisages while planning. A metaphor for what I mean is the AI telling a virtual copy of me all the stories of the future, from various view points, and the virtual me not regretting the wish. Of course I would expect it to be able to distill a non sentient version of the regret predicate.

So if it invented a scenario where it killed the real me, the predicate would still exist and say false. It would be able to predict this, and so not carry out this plan.

If you want to, generalize to humanity. This is not quite the same as CEV, as the AI is not trying to figure out what we want when we would be smarter, but what we don't want when we are dumb. Call it coherent no regret, if you wish.

CNR might be equivalent of CEV if humanity wishes not to feel regret in the future for the choice. That is if we would regret being in a future where people regret the decision, even though current people wouldn't.

Comment by Will_Pearson on Failed Utopia #4-2 · 2009-01-21T12:30:20.000Z · LW · GW

I don't believe in trying to make utopias but in the interest of rounding out your failed utopia series how about giving a scenario against this wish.

I wish that the future will turn out in such a way that I do not regret making this wish. Where I is the entity standing here right now, informed about the many different aspects of the future, in parallel if need be (i.e if I am not capable of groking it fully then many versions of me would be focused on different parts, in order to understand each sub part).

I'm reminded by this story that while we may share large parts of psychology, what makes a mate have an attractive personality is not something universal. I found the cat girl very annoying.

Comment by Will_Pearson on Building Weirdtopia · 2009-01-13T21:08:24.000Z · LW · GW

Personally I don't find the scientific weirdtopia strangely appealing. Finding knowledge for me is about sharing it later.

Utopia originally meant no-place, I have a hard time forgetting that meaning when people talk about them.

I'd personally prefer to work towards negated-dystopias. Which is not necessarily the same thing as working towards Utopia, depending on how broad your class of dystopia is. For example rather than try and maximise Fun, I would want to minimize the chance that humanity and all its work were lost to extinction. If there is time and energy to devote to Fun while humanity survives then people can figure it out for themselves.

Comment by Will_Pearson on Growing Up is Hard · 2009-01-04T15:55:35.000Z · LW · GW

Time scaling is not unproblematic. We don't have a single clock in the brain, clocks must be approximated by neurons and by neural firing. Speeding up the clocks may affect the ability to learn from the real world (if we have a certain time interval for associating stimuli).

We might be able to adapt, but I wouldn't expect it to be straight forward.

Comment by Will_Pearson on The Uses of Fun (Theory) · 2009-01-02T23:26:22.000Z · LW · GW

A random utility function will do fine, iff the agent has perfect knowledge.

Imagine, if you will a stabber, something that wants to turn the world into things that have been stabbed. If it knows that stabbing itself will kill itself, it will know to stab itself last. If it doesn't know know that stabbing itself will lead to it no longer being able to stab things, then it may not do well in actually achieving its stabbing goal by stabbing itself too early.

Comment by Will_Pearson on Can't Unbirth a Child · 2008-12-28T23:19:45.000Z · LW · GW

I'd agree with the sentiment in this post. I'm interested in building artificial brain stuff, more than building Artificial People. That is a computational substrate that allows the range of purpose-oriented adaptation shown in the brain, but with different modalities. Not neurally based, because simulating neural systems on a system where processing and memory is split defeats the majority of the point of them for me.

Comment by Will_Pearson on Nonperson Predicates · 2008-12-27T10:02:59.000Z · LW · GW

Don't you need a person predicate as well? If the RPOP is going to upload us all or something similar, doesn't ve need to be sure that the uploads will still be people.

Comment by Will_Pearson on Living By Your Own Strength · 2008-12-22T15:08:40.000Z · LW · GW

I suspect the knowledge you get from reading someones writings is very different than the knowledge you get from working with them or them teaching you. When you work or learn closely with someone they can see your reasoning processes and correct them when they go astray at the right point when they are still newly formed and not too ingrained. Otherwise it relies too much on luck. When in someone intellectual career should you read OB, too early it won't mean too much lacking the necessary background and too late you will be inured against it (assuming it is the right way to go!).

Autodidacts are going to be most intellectually useful when you need to break new ground and the methodologies of the past aren't the way to solve the problems needed to be solved.

Comment by Will_Pearson on What I Think, If Not Why · 2008-12-13T17:27:00.000Z · LW · GW

Are you saying "snakes are often deadly poisonous to humans" is an instrumental value?

I'd agree that dying is bad therefore avoid deadly poisonous things. But I still don't see that snakes have little xml tags saying keep away, might be harmful.... I don't see that as a value of any sort.

Comment by Will_Pearson on What I Think, If Not Why · 2008-12-13T15:29:00.000Z · LW · GW
Morality does not compress; it's not something you can learn just by looking at the (nonhuman) environment or by doing logic; if you want to get all the details correct, you have to look at human brains.

Why? Why can't you rewrite this as "complexity and morality"?

You may talk about the difference between mathematical and moral insights. Which is true, but then mathematical insights aren't sufficient for intelligence. Maths doesn't tell you whether a snake is poisonous and will kill you or not....

Comment by Will_Pearson on You Only Live Twice · 2008-12-12T22:28:34.000Z · LW · GW

The number of people living today because their ancestors invested their money in themselves/their status and their children, all of us:

The number of people living today because they or someone else invested their money in cryonics or other scheme to live forever, 0.

Not saying that things won't change in the future, but there is a tremendously strong bias to spend your resources on ambulatory people and new people, because that has been what has worked previously.

Women might have stronger instincts in this respect as they have been more strongly selected for the ability to care for their children (unlike men).

If you want to change this state of affairs, swiftly at least, you have to tap into our common psyche as successful replicators and have it pass the "useful for fitness test". This would be as easy as making it fashionable or a symbol of high status, get Obama to sign up publicly and I think you would see a lot more interest.

High status has been something sort after because it gets you better mates and more of them (perhaps illicitly).

Comment by Will_Pearson on What I Think, If Not Why · 2008-12-11T22:42:22.000Z · LW · GW
Will, your example, good or bad, is universal over singletons, nonsingletons, any way of doing things anywhere.

My point was not that non-singletons can see it coming. But if one non-singletons trys self-modification in a certain way and it doesn't work out then other non-singletons can learn from the mistake (or in worst the evolutionary case the descendents of people curious in a certain way would be out competed by those that instinctively didn't try the dangerous activity). Less so with the physics experiments, depending on dispersal of non-singletons, range of the physical destruction.

Comment by Will_Pearson on What I Think, If Not Why · 2008-12-11T22:08:13.000Z · LW · GW

There are some types of knowledge that seem hard to come by (especially for singletons). The type of knowledge is knowing what destroys you. As all knowledge is just an imperfect map, there are some things a priori that you need to know to avoid. The archetypal example is in-built fear of snakes in humans/primates. If we hadn't had this while it was important we would have experimented with snakes the same way we experiment with stones/twigs etc and generally gotten ourselves killed. In a social system you can see what destroys other things like you, but the knowledge of what can kill you is still hard won.

If you don't have this type of knowledge you may step into an unsafe region, and it doesn't matter how much processing power or how much you correctly use your previous data. Examples that might threaten singletons:

1) Physics experiments, the model says you should be okay but you don't trust your model under these circumstances, which is the reason to do the experiment. 2) Self-change, your model says that the change will be better but the model is wrong. It disables the system to a state it can't recover from, i.e. not an obvious error but something that renders it ineffectual. 3) Physical self-change. Large scale unexpected effects from feedback loops at a different levels of analysis, e.g. things like the swinging/vibrating bridge problem, but deadly.

Comment by Will_Pearson on Disjunctions, Antipredictions, Etc. · 2008-12-09T16:18:43.000Z · LW · GW
No diminishing returns on complexity in the region of the transition to human intelligence: "We're so similar to chimps in brain design, and yet so much more powerful; the upward slope must be really steep."

Or there is no curve and it is a random landscape with software being very important...

Scalability of hardware: "Humans have only four times the brain volume of chimps - now imagine an AI suddenly acquiring a thousand times as much power."

Bottle nosed dolphins have twice the brain volume as normal dolphins (and comparable to our brain volume), yet aren't massively more powerful compared to them. Asian elephants have 5 times the weight...

Comment by Will_Pearson on True Sources of Disagreement · 2008-12-09T01:00:49.000Z · LW · GW

I personally find the comparison between spike frequency and clockspeed unconvincing. It glosses over all sorts of questions of whether the system can keep all the working memory it needs in 2MB or whatever processor cache it has. Neurons have the advantage of having local memory, no need for the round trip off chip.

We also have no idea how neurons really work, there has been recent work on the role of methylation of dna in memory. Perhaps it would be better to view neural firing as communication between mini computers, rather than processing in itself.

I'm also unimpressed with large numbers, 10^15 operations is not enough to process the positions of 1 gram of hydrogen atoms, in fact it would take 20 million years for it to do so (assuming one op per atom). So this is what we have to worry about planning to atomically change our world to the optimal form. Sure it is far more than we can consciously do, and quite possibly a lot more than we can do unconsciously as well. But it is not mindboglingly huge compared to the real world.

Comment by Will_Pearson on Artificial Mysterious Intelligence · 2008-12-07T20:56:18.000Z · LW · GW

The universe doesn't have to be kind and make all problems amenable to insight....

There are only a certain number of short programs, and once a program gets above a certain length it is hard to compress (I can't remember the reference for this, so it may be wrong, can anyone help?). We can of course reorder things, but then we have to make things currently simple complex.

That said I do think insight will play some small part in the development of AI, but that there may well be a hell a lot of parameter tweaking that we don't understand or know why they are so.

Comment by Will_Pearson on Sustained Strong Recursion · 2008-12-06T11:00:49.000Z · LW · GW

You are right about the smaller is faster and local being more capable of reacting. But Eliezer's arguments are predicated on there being a type of AI that can change itself without deviation from a purpose. So an AI that splits itself into two may deviate in capability, but should share the same purpose.

Whether such an AI is possible or would be effective in the world is another matter.

Comment by Will_Pearson on Sustained Strong Recursion · 2008-12-05T23:54:56.000Z · LW · GW
We also suppose that the technology feeding Moore's Law has not yet hit physical limits. And that, as human brains are already highly parallel, we can speed them up even if Moore's Law is manifesting in increased parallelism instead of faster serial speeds - we suppose the uploads aren't yet being run on a fully parallelized machine, and so their actual serial speed goes up with Moore's Law. Etcetera.

Moore's Law says nothing about speed in the canonical form. You should probably define exactly what variant you are using.

Comment by Will_Pearson on Underconstrained Abstractions · 2008-12-05T07:02:49.000Z · LW · GW

Consider the following. Chimpanzees make tools. The first hominid tools were simple chipped stone from 2.5 million years ago. Nothing changed for a million years. Then homo erectus came along with Acheulian tech, nothing happened for a million years. Then two thousand years ago H. Sapiens appeared and tool use really diversified. The brains had been swelling from 3 million years ago.

If brains had been getting more generally intelligent at that time as they were increasing in size, it is not shown. They may have been getting better at wooing women and looking attractive to men.

This info has been cribbed from the Red Queen page 313 hardback edition.

I would say this shows a discontinous improvement in intelligence, where intelligence is defined as the ability to generally hit a small target in search space about the world. Rather than the ability to get into another hominids pants.

Comment by Will_Pearson on Recursive Self-Improvement · 2008-12-03T14:46:38.000Z · LW · GW
Because it shows that with constant optimization pressure from natural selection and no intelligent insight, there were no diminishing returns to a search for better brain designs up to at least the human level. There were probably accelerating returns (with a low acceleration factor). There are no visible speedbumps, so far as I know.

Were the brain designs better because they were more powerful or more intelligent?

That is how many of the improvements were adding more resources to the brain (because they paid off in this evolutionary case), rather than adding more efficient programs/systems.

Comment by Will_Pearson on Recursive Self-Improvement · 2008-12-02T20:15:03.000Z · LW · GW

Pearson: I think you have a tendency to overlook our lack of knowledge of how the brain works. You talk of constant brain circuitry, when people add new hippocampal cells through their life. We also expand the brain areas devoted to fingers if we are born blind and use braille.

Pearson, "constant brains" means "brains with constant adaptation-algorithms, such as an adaptation-algorithm for rewiring via reinforcement" not "brains with constant synaptic networks". I think a bit of interpretive charity would have been in order here. We don't know how deep the rabbit hole of adaptation goes. Are there constant adaptation-algortihms? Constant adaptation algorithms are not a prerequisite for an optimization process, evolution being the cannonical example. It gets by with a very changeable adaptation-algorithm embodied in the varieties of genetic transfer, reproduction rates etc. We have barely scratched the surface of adaption systems, assuming a constant adaptation-algorithm for intelligence is premature, as far as I am concerned.

Comment by Will_Pearson on Recursive Self-Improvement · 2008-12-01T23:58:32.000Z · LW · GW

I think you have a tendency to overlook our lack of knowledge of how the brain works. You talk of constant brain circuitry, when people add new hippocampal cells through their life. We also expand the brain areas devoted to fingers if we are born blind and use braille.

We don't know how else the brain rewires itself. In some sense all knowledge is wiring in the brain... I mean what else is it going to be. This is all invisible to us, and may throw a spanner in the works of any intelligence trying to improve a moving target.

Comment by Will_Pearson on Total Nano Domination · 2008-11-27T15:29:02.000Z · LW · GW

By nanocomputer I meant rod-logic or whatever the state of the art in hypothetical computing is. I want to see how it compares to standard computing.

I think the lure of nano computing is supposed to be low power consumption and easy 3d stackability that that entails as well. It it not sufficient to have small components if they are in a 2D design and you can't have too many together without overheating.

Some numbers would be nice though.

Comment by Will_Pearson on Total Nano Domination · 2008-11-27T11:21:26.000Z · LW · GW

How much of current R&D time is humans thinking, and how much is compiling projects, running computer simulations or doing physical experiments?

E.g. would having faster than human speed uploads, speed up getting results from the LHC by the ratio of their speed to us?

Also do you have some FLOPS per cubic centimeter estimations for nanocomputers? I looked at this briefly, and I couldn't find anything. It references a previous page that I can't find.

Comment by Will_Pearson on ...Recursion, Magic · 2008-11-25T10:01:21.000Z · LW · GW

Do you have any evidence that insight is applicable to understanding and creating intelligences? Because without that recursion isn't powerful and magic doesn't start to get off the ground.