What are important UI-shaped problems that Lightcone could tackle?
post by Raemon · 2025-04-27T00:02:36.311Z · LW · GW · 8 commentsContents
Some thoughts so far Cyborgism "Schleppy work" in narrow, technical domains What kinds of markets need to exist, that are difficult because of evaluation or reputation or (cognitive) transaction costs? ... None 8 comments
As I think about "what to do about AI x-risk?", some principles that seem useful to me:
- Short timelines seem plausible enough that, for the next year or so, I'd like to focus on plans that are relevant if takeoff begins in the next few years. In a year, if it looks more like there are some fundamental bottlenecks on true creative thinking, I may consider more projects that only payoff in longer-timeline stories.
- Given "short timelines", I feel most optimistic on plans that capitalize on skills that I'm already good at (but maybe multiclassing at things that I can learn quickly with LLM assistance).
- I think "UI design" is a skill that I (and Lightcone more broadly) am pretty good at. And, I believe the Interfaces as a Scarce Resource [LW · GW] hypothesis – the world is often bottlenecked on ability to process and make-use-of information in complicated, messy domains.
(In addition to LessWrong, the Lightcone team has worked on the S-Process, which makes it much easier for grantmakers to argue and negotiate pretty complex positions about how much they value various orgs).
If I've got a UI-shaped hammer, what are some nails that seem like they need doing? (In particular, that are somehow relevant to x-risk)
Some thoughts so far
Cyborgism
Last fall, I was oriented around building good UI for LLM-assisted-thinking.
In addition to it generally seeming like a fruitful green field, I had a specific hypothesis that, once LLMs Get Real Gud, there will be an important distinction between "being able to get useful work out of them given a minute's work, and, being able to get useful work out of them of them in <5 seconds." The latter is something that can become a true "part of your exobrain." The former is still more like a tool you're using.
I'm less bullish on that now because, while I think most people aren't quite tackling this with the particular taste I'd apply, it does sure seem like everyone is working on "do stuff with LLMs" and it's not where the underpicked fruit is.
"Schleppy work" in narrow, technical domains
It seems like there may be narrow, technical domains specific narrow domains, where there's some kinds of tasks that rarely get done because they're too hard to think about –you need tons of context, the context is littered around various places. Or, maybe it's all technically in one place but you have to sift through a lot of extraneous details.
A past example of this would be "hoverovers in coding IDEs for types and docs", linters, etc. The ability to right-click on a function call to go to the original implementation of the function.
For a less technical example: spellchecking and grammarchecking in word processor docs.
A possible (current, important) example might be "something something Mech Interp a la Chris Olah's earlier work on distill.pub". I don't actually know how Mech Interp works these days, I vaguely believe there are visualizers/heat maps for neurons, but I'm not sure how useful those actually are for the kinds of analysis that are most important.
An example John Wentworth has mentioned a couple times is "automatically generating examples of your abstractions and making sure they type-check." (This involves both UI, and some deep technical work to actually verify the type checking)
What kinds of markets need to exist, that are difficult because of evaluation or reputation or (cognitive) transaction costs?
It's sort of surprising that Amazon, Uber or Lugg work. Why aren't people receiving bobcats when they order a stapler all the time? A large part of the answer are rating systems, and an ecosystem where it's not trivial to build up a reputation.
What are some places where you can't easily buy or find a thing because of adversarial optimization?
...
With those illustrative examples: do you work on x-risk or x-risk adjacent things? What are some places in your work where it's confusing or annoying to find things, or figure things out?
8 comments
Comments sorted by top scores.
comment by Alice Blair (Diatom) · 2025-04-27T01:57:10.484Z · LW(p) · GW(p)
I work mostly as a distiller (of xrisk-relevant topics). I try to understand some big complex thing, package it up all nice, and distribute it. The "distribute it" step is something society has already found a lot of good tech for. The other two steps, not so much.
Loom [LW · GW] is lovely in the times I've used it. I would love to see more work done on things like this, things that enhance my intelligence while keeping me very tightly in the loop. Other things in this vein include:
- memory augmentation beyond just taking notes (+Anki where appropriate). I'm both interested in working memory and long-term memory, but moreso in the former.
- For digesting complex pieces, I'd like something better than a chat window for interactively getting up to my desired level of familiarity with some existing work. When I'm digesting the paper, do I want the one-sentence summary, the abstract, or to have a better understanding than the researchers behind it? NotebookLM is sort of doing this but I've found it lacking for this task for UI reasons (I also wish that it would automatically bring in other relevant sources).
comment by romeostevensit · 2025-04-27T01:10:56.814Z · LW(p) · GW(p)
I think there's a possibility for ui people to make progress on the reputation tracking problem by virtue of tight feedback loops relative to people thinking more abstractly about it. The most rapid period of learning in this regard that I know of is early days at PayPal eBay where they were burning millions a day in fraud at certain points.
Secondly: the chat interface for llm is just bad for power users. Ai Labs is slightly better but still bad.
Edit: meant aistudio
Replies from: Raemon, Diatom↑ comment by Raemon · 2025-04-27T02:07:05.753Z · LW(p) · GW(p)
I think there's a possibility for ui people to make progress on the reputation tracking problem by virtue of tight feedback loops relative to people thinking more abstractly about it.
Are there particular reputation-tracking-problems you're thinking of? (I'm sure there are some somewhere, but I'm looking to get more specific)
I'm working on a poweruser LLM interface but honestly it's not going to be that much better than Harpa AI or Sider.
↑ comment by Alice Blair (Diatom) · 2025-04-27T01:32:19.678Z · LW(p) · GW(p)
Ai Labs is slightly better but still bad.
Could you give a link to this or a more searchable name? "Ai Labs" is very generic and turns up every possible result. Even if it's bad, I'd be interested in investigating something "slightly better" and hearing a bit about why.
Replies from: romeostevensit↑ comment by romeostevensit · 2025-04-27T01:37:58.272Z · LW(p) · GW(p)
Oops meant aistudio.google.com
comment by Thane Ruthenis · 2025-04-27T09:47:39.176Z · LW(p) · GW(p)
while I think most people aren't quite tackling this with the particular taste I'd apply, it does sure seem like everyone is working on "do stuff with LLMs" and it's not where the underpicked fruit is
I disagree, I think pretty much nobody is attempting anything useful with LLM-based interfaces. Almost all projects I've seen in the wild are terrible and there are tons of unpicked low-hanging fruits.
I'd been thinking, on and off, about ways to speed up agent-foundations research using LLMs. An LLM-powered exploratory medium for mathematics is one possibility.
A big part of highly theoretical research is flipping between different representations of the problem: viewing it in terms of information theory, in terms of Bayesian probability, in terms of linear algebra; jumping from algebraic expressions to the visualizations of functions or to the nodes-and-edges graphs of the interactions between variables; et cetera.
The key reason behind it is that research heuristics bind to representations. E. g., suppose you're staring at some graph-theory problem. Certain problems of this type are isomorphic to linear-algebra problems, and they may be trivial in linear-algebra terms. But unless you actually project the problem into the linear-algebra ontology, you're not necessarily going to see the trivial solution when staring at the graph-theory representation. (Perhaps the obvious solution is to find the eigenvectors of the adjacency matrix of the graph – but when you're staring at a bunch of nodes connected by edges, that idea isn't obvious in that representation at all.)
This is a bit of a simplified example – the graph theory/linear algebra connection is well-known, so experienced mathematicians may be able to translate between those representations instinctively – but I hope it's illustrative.[1]
As a different concrete example, consider John Wentworth's Bayes Net Algebra [LW · GW]. This is essentially an interface for working with factorizations of joint probability distributions. The nodes-and-edges representation is more intuitive and easy to tinker with than the "formulas" representation, which means that having concrete rules for tinkering with graph representations without committing errors would significantly speed up how quickly you can reason through related math problems. Imagine if the derivation of such frameworks was automated: if you could set up a joint PD in terms of formulas, automatically project the setup into graph terms, start tinkering with it by dragging nodes and edges around, and get errors if and only if back-projecting the changed "graph" representation into the "formulas" representations results in a setup that's non-isomorphic to the initial one.
(See also this video, and the article linked above.)
A related challenge are refactors. E. g., suppose you're staring at some complicated algebraic expression with an infinite sum. It may be the case that a certain no-loss-of-generality change of variables would easily collapse that expression into a Fourier series, or make some Obscure Theorem #418152/Weird Trick #3475 trivially applicable. But unless you happen to be looking at the problem through those lens, you're not going to be able to spot it. (Especially if you don't know the Obscure Theorem #418152/Weird Trick #3475.)
It's plausible that the above two tasks is what 90% of math research consists of (the "normal-science" part of it), in terms of time expenditure. Flipping between representations in search of a representation-chain where every step is trivial.
Those problems would be ameliorated by (1) reducing the friction costs of flipping between representations, and (2) being able to set up an automated searches for simplifying refactors of the problem.
Can LLMs help with (1)? Maybe. They can write code and they can, more or less, reason mathematically, as long as you're not asking them for anything creative. One issue is that they're also really sloppy and deceptive when writing proofs... But that problem can potentially be ameliorated by fine-tuning e. g. r1 to justify all its conclusions using rigorous Lean code, which could be passed to automated proof-checkers before being shown to you.[2]
Can LLMs help with (2)? Maybe. I'm thinking something like the Pantheon interface [LW · GW], where you're working through the problem on your own, and in a side window LLMs offer random ideas regarding how to simplify the problem.
LLMs have bad research taste, which would extend to figuring out what refactorings they should try. But they also have a superhuman breadth of knowledge regarding theorems/math results. A depths-first search might thus be productive here. Most of LLM suggestions would be trash, but as long as complete nonsense is screened off by proof-checkers, and the ideas are represented in a quickly-checkable manner (e. g., equipped with one-sentence summaries), and we're giving LLMs an open-ended task, some results may be useful.
I expect I'd pay $200-$500/month for a working, competently executed tool of this form; even more the more flexible it is. I expect plenty of research mathematicians (not only agent-foundations folks) would, as well. There's a lucrative startup opportunity there.
@johnswentworth [LW · GW], any thoughts?
- ^
A more realistic example would concern ansatzes, i. e., various "weird tricks" for working through problems. They likewise bind to representations, such that the idea of using one would only occur to you if you're staring at a specific representation of the problem, and would fail to occur if you're staring at an isomorphic-but-shallowly-different representation.
- ^
Or using o3 with a system prompt where you yell at it a lot to produce rigorous Lean code, with a proof-checker that returns errors if it ever uses a placeholder always-passes "sorry" expression. But I don't know whether you can yell at it loudly enough using just the system prompt, and this latest generation of LLMs seems really into Goodharting, so it might straight-up try to exploit bugs in your proof-checker.
comment by Tenoke · 2025-04-27T09:39:32.272Z · LW(p) · GW(p)
I still want something even closer to Givewell but for AI Safety (though it is easier to find where to donate now than before). Hell, I wouldn't mind if LW itself had recommended charities in a prominent place (though I guess LW now mostly asks for Lightcone donations instead).
comment by Knight Lee (Max Lee) · 2025-04-27T06:34:21.283Z · LW(p) · GW(p)
One big UI shaped problem, is that when I visit the website of an extremely corrupt and awful company with a lot of scandals, they often trick me into thinking they are totally good, because I'm too lazy to search up their Wikipedia page.
What if we create a new tiny wiki as a browser extension, commenting on every website?
The wiki should only say one or two sentences about every website, since we don't want to use up too much of the user's screen space while she is navigating the website.
The user should only see the wiki when scrolled to the top of the webpage. If the user clicks "hide for this site," the wiki collapses into a tiny icon (which is red or green depending on the organization's overall score). If the wiki for one website has already been shown for 5 minutes, it automatically hides (but it expands again the next week).
Details
Whenever people make an alternative to Wikipedia, they always start off by simply copying Wikipedia.
This is okay! Wikipedia as a platform does not own the work of its editors, its editors are not loyal to the platform but loyal to the idea of sharing their knowledge, and don't mind if you copy their work to your own platform. There's no copyright.
The current Wikipedia is longer than one or two sentences, so you might need to use summarize it with AI (sadly). But as soon as a user edits it, her edit replaces the AI slop.
Where do we display the one or two sentences about the website? The simplest way is to create a thin horizontal panel on the top or bottom of the website. A more adaptive way, is to locate some whitespace in the website and add it there.
It might only display in certain webpages within a website. E.g. for a gaming website, it might not display while the user is gaming, since even one or two sentences uses up too much screen space. It might only display in the homepage of the gaming website.
Font size is preferably small.
If the user mouse-hovers the summary, it opens up the full Wikipedia page (in a temporary popup). If the website has no Wikipedia page (due to Wikipedia's philosophy of "deletionism"), your wiki users can write their own. Even if it has a Wikipedia page, your wiki users can add annotations to the existing Wikipedia page (e.g. if they disagree with Wikipedia's praise of a bad company).
In addition to the full Wikipedia page, there might be a comments section (Wikipedia frustratingly disallows comments), and possibly a web search.
Worthwhile gamble
80%, trying to create it will fail. But 20%, it works, at least a little.
But the cost is a mere bit of UI work, and the benefit is huge.
It can greatly help the world on judging bad companies! This feels "unrelated to AI risk," but helps a lot if you think about it.
If it works, then whichever organization implements it first will win a lot of donations, and act as the final judge in savage fights over website reputations.