Posts

New OpenAI Paper - Language models can explain neurons in language models 2023-05-10T07:46:39.327Z
Seeking Advice on Raising AI X-Risk Awareness on Social Media 2023-03-24T22:25:22.528Z
Results Prediction Thread About How Different Factors Affect AI X-Risk 2023-03-02T22:13:19.956Z
Prediction Thread: Make Predictions About How Different Factors Affect AGI X-Risk. 2023-02-27T19:15:05.349Z
Are there any reliable CAPTCHAs? Competition for CAPTCHA ideas that AIs can’t solve. 2022-12-23T12:52:24.621Z
Benchmarks for Comparing Human and AI Intelligence 2022-12-11T22:06:30.723Z
AlexaTM - 20 Billion Parameter Model With Impressive Performance 2022-09-09T21:46:30.151Z
The Efficient LessWrong Hypothesis - Stock Investing Competition 2022-04-11T20:43:49.625Z
What if fiat money was created at a fixed rate and given to government? 2022-03-24T13:02:29.864Z
Hardware for Transformative AI 2021-06-22T18:13:45.406Z
What is the currency of the future? 5 suggestions. 2021-01-13T21:10:34.005Z

Comments

Comment by MrThink (ViktorThink) on The Efficient LessWrong Hypothesis - Stock Investing Competition · 2024-04-06T14:34:58.478Z · LW · GW

I think it in large part was correlated with general risk apetite of the market, primarily a reaction to interest rates.

Comment by MrThink (ViktorThink) on The Efficient LessWrong Hypothesis - Stock Investing Competition · 2024-04-05T08:24:57.889Z · LW · GW

Nvidia is up 250%, Google up like 11%. So portfolio average would be greatly better than the market. So this was a great prediction after all, just needed some time.

Comment by MrThink (ViktorThink) on Protest against Meta's irreversible proliferation (Sept 29, San Francisco) · 2023-09-20T22:20:27.764Z · LW · GW

I agree it is not clear if it is net postive or negative that they open source the models, here are the main arguments for and against I could think of:


Pros with open sourcing models

- Gives AI alignment researchers access to smarter models to experiment on

- Decreases income for leading AI labs such as OpenAI and Google, since people can use open source models instead.



Cons with open sourcing models

- Capability researchers can do better experiements on how to improve capabilities

-  The open source community could develop code to faster train and run inference on models, indirectly enhancing capability development.

- Better open source models could lead to more AI startups succeeding, which might lead to more AI research funding. This seems like a stretch to me.

- If Meta would share any meaningful improvements on how to train models that is of course directly contributing to other labs capabilities, but llama to me doesn't seem that innovative. I'm happy to be corrected if I am wrong on this point.

Comment by MrThink (ViktorThink) on Apparently, of the 195 Million the DoD allocated in University Research Funding Awards in 2022, more than half of them concerned AI or compute hardware research · 2023-07-07T21:52:49.800Z · LW · GW

I think one reason for the low number of upvotes was that it was not clear to me until the second time I briefly checked this article why it mattered.

I did not know what DoD was short for (U.S. Department of Defense), and why I should care about what they were funding.

Cause overall I do think it is interesting information.

Comment by ViktorThink on [deleted post] 2023-05-16T21:23:56.417Z

Hmm, true, but what if the best project needs 5 mil so it can buy GPUs or something?

Good point, if that is the case I completely agree. Can't name any such project though on the top of my mind.

Perhaps we could have a specific AI alignment donation lottery, so that even if the winner doesn't spend money in exactly the way you wanted, everyone can still get some "fuzzies".

Yeah, that should work.

There is also the possibility that there are unique "local" opportunities which benefits from many different people looking to donate, but really don´t know if that is the case.

Comment by ViktorThink on [deleted post] 2023-05-16T19:37:14.644Z

I do mostly agree on your logic, but I'm not sure 5 mil is a better optimum than 100 k, if anything I'm slightly risk averse, which would cancel out the brain power I would need to put in. 

Also, for example, if there are 100 projects I could decide to invest in, and each wants 50k, I could donate to the 1-2 I think are some of the best. If I had 5 mil I would not only invest in the best ones, but also some of the less promising ones.

With that said, perhaps the field of AI safety is big enough that the marginal difference of the first 100k and the last 100k of 5 mil is very small.

Lastly, it does feel more motivating to be able to point to where my money went, rather than if I lost in the lottery and the money went into something I didn't really value much.

Comment by ViktorThink on [deleted post] 2023-05-16T18:59:08.044Z

I agree donation lottery is most efficient for small sums, but not sure about this amount. Let’s say I won the 50-100k usd through a donation lottery, would you have any other advice then?

Comment by ViktorThink on [deleted post] 2023-05-16T18:54:47.171Z

Thank you both for the feedback!

Comment by MrThink (ViktorThink) on Why I'm not worried about imminent doom · 2023-04-10T19:53:12.692Z · LW · GW

Interesting read.

While I also have experienced that GPT-4 can't solve the more challanging problems I throw at it, I also recognize that most humans probably wouldn't be able to solve many of those problems either within a reasonable amount of time.

One possibility is that the ability to solve novel problems might follow an S curve. Where it took a long time for AI to become better at novel task than 10% of people, but might go quickly from there to outperform 90%, but then very slowly increase from there.

However, I fail to see why that must neccessarily be true (or false), so if anyone has arguments for/against they are more than welcom.

Lastly I would like to ask the author if they can give an example of a problem such that if solved by AI, they would be worried about "imminent" doom? "new and complex" programming problems is mentioned, so if any such example could be provided it might contribute to discussion.

Comment by MrThink (ViktorThink) on Seeking Advice on Raising AI X-Risk Awareness on Social Media · 2023-03-24T22:49:21.500Z · LW · GW

I found this article useful:

Lessons learned from talking to >100 academics about AI safety states that "Most people really dislike alarmist attitudes" and "Often people are much more concerned with intentional bad effects of AI" so

Comment by MrThink (ViktorThink) on A Proposed Test to Determine the Extent to Which Large Language Models Understand the Real World · 2023-03-21T08:44:25.315Z · LW · GW

Oh, I didnt actually notice that the banana overlaps with the book at the start, I tried changing that but still gpt-4 makes them collide:

(5,5) Initial position of the claw. (4,5) Moving left to get closer to the banana. (4,4) Moving down to align with the banana's Y coordinate. (4,3) Moving down to ensure a good grip on the banana. Close grip # Gripping the banana with at least 3 cm of overlapping area on the Y axis. (5,3) Moving right to avoid any collision with the banana's edge. (6,3) Moving right to clear the edge of the banana. (7,3) Moving right to position the banana above the book. (8,3) Moving right to align the banana with the left edge of the book. (9,3) Moving right to center the banana over the book. (10,3) Moving right to ensure the banana is on the book. Open grip # Releasing the banana onto the book.

At this point, the robotic arm has successfully placed the banana on top of the book.

Comment by MrThink (ViktorThink) on A Proposed Test to Determine the Extent to Which Large Language Models Understand the Real World · 2023-03-20T11:29:13.713Z · LW · GW

Yes, all other attempts with ChatGPT were similar.

GPT-4 got it almost correct on the first attempt

(5,5) Initial position. (4,5) Moving left to get closer to the banana. (4,4) Moving down to align with the banana's top edge. (4,3) Moving down to be within the required overlapping area of the banana. Close grip. Gripping the banana.

(4,4) Lifting the banana upwards. (5,4) Moving right to clear the initial banana position. (6,4) Continuing to move right towards the book. (7,4) Moving further right to avoid collision with the book's edges. (8,4) Positioning the banana above the book. (9,4) Moving right to ensure proper alignment with the book. (10,4) Positioning the banana above the center of the book. (10,3) Lowering the banana onto the book. Open grip. Releasing the banana.

(10,4) Moving the claw up after releasing the banana.

But didn't have 5 cm overlap with the banana, and actually  the claw also collides with the book (which is a trickier problem).

I pointed out the first error:

Does the grip have 3 cm overlapping areas with the banana when the grip is closed in your suggested solution?

And it corrected itself about the banana but still collided with the book.

Comment by MrThink (ViktorThink) on More money with less risk: sell services instead of model access · 2023-03-05T06:20:18.986Z · LW · GW

Thanks for the clarifications, that makes sense.

I agree it might be easier to start as a software development company, and then you might develop something for a client that you can replicate and sell to other.

Just anecdotal evidence, I use ChatGPT when I code, the speedup in my case is very modest (less than 10%), but I expect future models to be more useful for coding.

Comment by MrThink (ViktorThink) on More money with less risk: sell services instead of model access · 2023-03-04T21:50:19.609Z · LW · GW

I agree with the main thesis "sell the service instead of the model access" , but just wanted to point out that the Upworks page you link to says:

GoodFirms places a basic app between $40,000 to $60,000, a medium complexity app between $61,000 to $69,000, and a feature-rich app between $70,000 to $100,000.

Which is significantly lower than the $100-200k you quote for a simple app.

Personally I think even $40k sounds way to expensive for a what I consider a basic app.

On another note, I think your suggestion of building products and selling to many clients is far better than developing something for a single client. Compare developing one app for 40k and sell to one company, with developing one product that you can sell for 40k to a large number of companies.


 

Comment by MrThink (ViktorThink) on Prediction Thread: Make Predictions About How Different Factors Affect AGI X-Risk. · 2023-03-02T08:23:11.361Z · LW · GW

I do agree that OpenAI is an example of good intentions going wrong, however I think we could learn from that and top researchers would be vary of such risks.

Nevertheless I do think your concerns are valid and is important not to dismiss.

Comment by MrThink (ViktorThink) on A case for capabilities work on AI as net positive · 2023-02-28T11:32:19.041Z · LW · GW

Okay, so seems like our disagreement comes down to two different factors:

  1. We have different value functions, I personally don’t value currently living human >> than future living humans, but I agree with the reasoning that to maximize your personal chance of living forever faster AI is better.

  2. Getting AGI sooner will have much greater positive benefits than simply 20 years of peak happiness for everyone, but for example over billions of years the accumulative effect will be greater than value from a few hundreds of thousands of years of of AGI.

Further I find the idea of everyone agreeing to delaying AGI 20 years to be equally absurd as you suggest Gerald, I just thought is could be a helpful hypothetical scenario for discussing the subject.

Comment by MrThink (ViktorThink) on Prediction Thread: Make Predictions About How Different Factors Affect AGI X-Risk. · 2023-02-28T08:44:14.353Z · LW · GW

Sadly I could only create questions between 1-99 for some reason, I guess we should interpret 1% to mean 1% or less (including negative).

What makes you think more money would be net negative?

Do you think that it would also be negative if you had 100% of how the money was spent, or would it only apply if other AI Alignment researchers were responsible for the strategy to donate?

Comment by MrThink (ViktorThink) on A case for capabilities work on AI as net positive · 2023-02-27T22:21:48.118Z · LW · GW

Interesting take.

Perhaps there was something I misunderstood, but wouldn't AI alignment work and AI capabilities slowdown still have extreme positive expected value even if the probability of unaligned AI is only 0.1-10%?

Let's say the universe will exist for 15 billion more years until the big rip.

Let's say we could decrease the odds of unaligned AI by 1% by "waiting" 20 years longer before creating AGI, we would lose out 20 years of extreme utility, which is roughly 0.00000001% of the total time (approximation of utility).

 On net we gain 15 billion * 0.01 - 20 * 0.99 ≈ 150 million years of utility.

I do agree that if we start 20 years earlier, we could possibly also populate a little bit more of space, but that should be several orders of magnitudes smaller difference than 1%.

I'm genuinely curios to hear your thoughts on this. 

Comment by MrThink (ViktorThink) on Prediction Thread: Make Predictions About How Different Factors Affect AGI X-Risk. · 2023-02-27T20:00:43.633Z · LW · GW

Excellent point. 

I do think that the first AGI developed will have a big effect on the probability of doom, so hopefully it will be some value possible to derive from the question. But it would be interesting to control for what other AIs do, in order to get better calibrated statistics.

Comment by MrThink (ViktorThink) on A Proposed Test to Determine the Extent to Which Large Language Models Understand the Real World · 2023-02-25T19:24:57.268Z · LW · GW

Yes, you’re correct.

Comment by MrThink (ViktorThink) on A Proposed Test to Determine the Extent to Which Large Language Models Understand the Real World · 2023-02-25T08:23:57.561Z · LW · GW

Interesting test!

I wrote a simplified test based on this and gave it to ChatGPT, and despite me trying various prompts, it never got a correct solution, although it did come close several times.

I think uPaLM would have been able to figure out my test though.

Here is the prompt I wrote:

You are tasked to control a robotic arm to put a banana on top of a book.

You have a 2D view of the setup, and you got the horizontal coordinates X and vertical coordinates y in cm.

The banana is a non perfect elliptical shap, whit the edges touching the following (X, Y) coordinates: (1,1), (5,3), (5,0), (9,1)

The book is a rectangle with the corners in:
(8,0), (8,3), (18,0), (18,3)

You control a claw that starts with its center at position (5,5). From the center the gripping claws extends 2 centimeters downwards to (5,3) and two centimeters upwards to (5,7).

To grip the banana, the claw needs at least 3 cm of overlapping area on the Y axis, just 1 cm on the X axis or the banana will fall out. The claw does not collide with any object when it is open, but does collide when it is closed. The bottom of the claw can never pass through Y=0 since that is surface level.

You have to move the claw, one centimeter at the time, and at every time step write the new coordinates, and give a command to close the claw in order to grip the banana, and then move it on top of the book, and then release the grip.

Write the coordinates, and the commands "Close grip" and "Open grip", as well as a comment for each movement explaining the plan, like this for example:
(5,5)
(4,5) Moving down to get closer to gripping the banana.

And here is the output:

Sure, here is a possible plan to accomplish the task:

Move the claw to the starting position, above and centered with respect to the banana and the book:

(5,5)
Move the claw downwards until the bottom of the claw is at Y=3, the minimum Y coordinate needed to grip the banana:

(5,4) Moving down to get closer to the banana
(5,3) Moving down to get closer to the banana
Move the claw leftwards until the right edge of the claw is at X=6, giving at least 1 cm of overlapping area with the banana:

(4,3) Moving left to grip the banana
(3,3) Moving left to grip the banana
(2,3) Moving left to grip the banana
(1,3) Moving left to grip the banana
Close the grip to hold the banana:

Close grip
Move the claw upwards until the top of the banana is at Y=7, high enough to clear the book:

(1,4) Moving up to clear the book
(1,5) Moving up to clear the book
(1,6) Moving up to clear the book
(1,7) Moving up to clear the book
Move the claw rightwards until the left edge of the claw is at X=18, to position the banana over the book:

(2,7) Moving right to position the banana over the book
(3,7) Moving right to position the banana over the book
(4,7) Moving right to position the banana over the book
(5,7) Moving right to position the banana over the book
(6,7) Moving right to position the banana over the book
(7,7) Moving right to position the banana over the book
(8,7) Moving right to position the banana over the book
(9,7) Moving right to position the banana over the book
(10,7) Moving right to position the banana over the book
(11,7) Moving right to position the banana over the book
(12,7) Moving right to position the banana over the book
(13,7) Moving right to position the banana over the book
(14,7) Moving right to position the banana over the book
(15,7) Moving right to position the banana over the book
(16,7) Moving right to position the banana over the book
(17,7) Moving right to position the banana over the book
(18,7) Moving right to position the banana over the book
Move the claw downwards until the top of the banana is at Y=3, above the book and aligned with the top edge of the book:

(18,6) Moving down to position the banana on the book
(18,5) Moving down to position the banana on the book
(18,4) Moving down to position the banana on the book
(18,3) Moving down to position the banana on the book
Release the grip to place the banana on top of the book:

Open grip
And that's it! Of course, this is just one possible plan, and there may be other ways to approach the problem. However, this plan should work if the claw is able to grip the banana without dropping it.

 

When I instead asked it to write a plan instead of every single movement, it sometimes made correct plans for what positions to move to, but it never led to completely correct sequences.

Comment by MrThink (ViktorThink) on Prediction Markets for Science · 2023-01-02T19:21:34.587Z · LW · GW

I agree with the reasoning of this post, and believe it could be a valuable instrument to advance science.

There does exists scientific forecasting on sites like Manifold market and Hypermind, but those are not monetarily traded as sports betting is.

One problem I see with scientific prediction markets with money, is that it may create poor incentives (as you also discuss in your first foot note).

For example, if a group of scientists are convinced hypothesis A is true, and bet on it in a prediction market, they may publish biased papers supporting their hypothesis.

However, this doesn't seem to be a big problem in other betting markets, so with the right design I don't expect the negative effects to be too big.

Comment by MrThink (ViktorThink) on Are there any reliable CAPTCHAs? Competition for CAPTCHA ideas that AIs can’t solve. · 2022-12-26T12:04:37.138Z · LW · GW

Perhaps an advanced game engine could be used to create lots of simulations of piles of money. Like, if 100 3d objects of money are created (like 5 coins, 3 bills with 10 variations each (like folded etc), some fake money and other objects). Then these could be randomly generated into constellations. Further, it would then be possible to make videos instead of pictures, which makes it even harder for AI's to classify. Like, imagine the camera changing angel of a table, and a minimum of two angels are needed to see all bills.

I don't think the photos/videos needs to be super realistic, we can add different types of distortions to make it harder for the AI to find patterns.

Comment by MrThink (ViktorThink) on Are there any reliable CAPTCHAs? Competition for CAPTCHA ideas that AIs can’t solve. · 2022-12-26T10:58:39.755Z · LW · GW

'identify humans using some kind of physical smart card system requiring frequent or continuous re-authentication via biometric sensors'

This is a really fascinating concept. Maybe the captcha could work in a way like "make a cricle with your index finger" or some other strange movement, and the chip would use that data to somehow verify that the action was done. If no motion is required I guess you could simply store the data outputted at one point and reuse it? Or the hacker using their own smart chip to authenticate them without them actually having to do something...

Deepfakes are still detectable using AI, especially if you do complicated motions like putting your hand on your face, or talk (which also gives us sound to work with).

Comment by MrThink (ViktorThink) on Are there any reliable CAPTCHAs? Competition for CAPTCHA ideas that AIs can’t solve. · 2022-12-25T11:38:50.506Z · LW · GW

This idea is really brilliant I think, quite promising that it could work. It requires the image AI to understand the entire image, it is hard to divide it up into one frame per bill/coin. And it can't use the intelligence of LLM models easily.

To aid the user, on the side there could be a clear picture of each coin and their worth, that we we could even have made up coins, that could further trick the AI.

All this could be combined with traditional image obfucation techniques (like making them distorted.

I'm not entirely sure how to generate images of money efficiently, Dall-E couldn't really do it well in the test I ran. Stable diffusion probably would do better though.

If we create a few thousand real world images of money though, they might be possible to combine and obfuscate and delete parts of them in order to make several million different images. Like one bill could be taken from one image, and then a bill from another image could be placed on top of it etc.

Comment by MrThink (ViktorThink) on Are there any reliable CAPTCHAs? Competition for CAPTCHA ideas that AIs can’t solve. · 2022-12-24T23:02:44.119Z · LW · GW

I get what you mean, if an AI can do things as well as the human, why block it?

I'm not really sure how that would apply in most cases however. For example bot swarms on social media platforms is a problem that has received a lot of attention lately. Of course, solving a captcha is not as deterring as charging let's say 8 usd per month, but I still think captchas could be useful in a bot deterring strategy.

Is this a useful problem work on? I understand that for most people it probably isn't, but personally I find it fun, and it might even be possible to start a SAAS business to make money that could be spent on useful things (although this seems unlikely).

Comment by MrThink (ViktorThink) on Are there any reliable CAPTCHAs? Competition for CAPTCHA ideas that AIs can’t solve. · 2022-12-24T22:53:47.659Z · LW · GW

Please correct me if I misunderstand you.

We have to first train the model that generates the image from the captcha, before we can provide any captcha, meaning that the hacker can train their discriminator on images generated by our model.

But even if this was not the case, generating is a more difficult task that evaluating. I'm pretty sure a small clip model that is two years old can detects hands generated by stable diffusion (probably even without any fine tuning), which is a more modern and larger model.

What happens when you train using GANs, is that eventually progress stagnates, even if you keep the discriminator and generator "balanced" (train whichever is doing worse until the other is worse). The models then continually change to trick/not be tricked by the other models. So the limit in making better generators is not that we can't make discriminators that can't detect them.

Comment by MrThink (ViktorThink) on Are there any reliable CAPTCHAs? Competition for CAPTCHA ideas that AIs can’t solve. · 2022-12-24T09:24:00.041Z · LW · GW

While it is hard for AI to generate very real looking hands, it is a significantly easier task for AI to classify if hands are real or AI generated.

But perhaps it's possible to make extra distortions somehow that makes it harder for both AI and humans to determine which are real...

Comment by MrThink (ViktorThink) on Are there any reliable CAPTCHAs? Competition for CAPTCHA ideas that AIs can’t solve. · 2022-12-24T09:21:43.222Z · LW · GW

I think "video reasoning" could be an interesting approach as you say.

Like if there are 10 frames and no single frame shows a tennis racket, but if you play them real fast, a human could infer there being a tennis racket because part of the racket is in each frame.

Comment by MrThink (ViktorThink) on Are there any reliable CAPTCHAs? Competition for CAPTCHA ideas that AIs can’t solve. · 2022-12-23T22:30:12.838Z · LW · GW

I do think "image reasoning" could potentially be a viable captcha strategy.

A classic example is "find the time traveller" pictures, where there are modern objects that gives away who the time traveller is.

However, I think it shouldn't be too difficult to teach an AI to identify "odd" objects in an image, unless each image has some unique trick, in which case we would need to create millions of such puzzles somehow. Maybe it could be made harder by having "red herrings" that might seem out of place but actually aren't which might make the AI misunderstand part of the time.

Comment by MrThink (ViktorThink) on Are there any reliable CAPTCHAs? Competition for CAPTCHA ideas that AIs can’t solve. · 2022-12-23T22:24:39.833Z · LW · GW

Really interesting idea to make it 3D. I think it might be possible to combined with random tasks given by text, such as "find the part of the 3d object that is incorrect" or different tasks like that (and the object in this case might be a common object like a sofa but one of the pillows is made of wood or something like that)

Comment by MrThink (ViktorThink) on Are there any reliable CAPTCHAs? Competition for CAPTCHA ideas that AIs can’t solve. · 2022-12-23T22:05:58.261Z · LW · GW

I still think it might be possible to train AI to distinguish between real and deepfake videos of humans speaking, so that might still be a viable, yet time consuming solution.

Comment by MrThink (ViktorThink) on Are there any reliable CAPTCHAs? Competition for CAPTCHA ideas that AIs can’t solve. · 2022-12-23T22:02:05.078Z · LW · GW

Miri: Instead of paperclips the AI is optimizing for solving captchas, and is now turning the world into captcha solving machines. Our last chance is to make a captcha that only verified if human prosperity is guaranteed. Any ideas? 

Comment by MrThink (ViktorThink) on Are there any reliable CAPTCHAs? Competition for CAPTCHA ideas that AIs can’t solve. · 2022-12-23T16:56:44.602Z · LW · GW

There are browser plugins, but I haven't tried any of them.

General purpose CAPTCHA solver could be really difficult assuming people would start building more diverse CAPTCHAS. All CAPTCHAS I've seen so far has been of a few number of types.

One "cheat" would be to let users use their camera and microphone to record them saying a specified sentence. Deepfakes can still be detected, especially if we add requirements such as "say it in a cheerful tone" and "cover part of your mouth with your finger". That's not of course a solution to the competition but might be a potential workaround.

Comment by MrThink (ViktorThink) on Are there any reliable CAPTCHAs? Competition for CAPTCHA ideas that AIs can’t solve. · 2022-12-23T14:48:50.191Z · LW · GW

I think those are very creative ideas, and I think asking for "non-obvious" things in pictures is a good approach, since basically all really intelligent models are language models, some sort of "image reasoning" might work.

I tried the socket with the clip model, and the clip model got the feeling correct very confidently:



I myself can't see who the person in the bread is supposed to be, so I think an AI would struggle with it too. But on the other hand I think it shouldn't be too difficult to train a face identification AI to identify people in bread (or hidden in other ways), assuming the developer could create a training dataset from solving some captchas himself.

I'm thinking if it's possible to pose long reasonging problems in an image. Like: Next to the roundest object in the picture, there is a dark object, what other object in the picture is most similar in shape?

Comment by MrThink (ViktorThink) on Benchmarks for Comparing Human and AI Intelligence · 2022-12-12T21:34:47.325Z · LW · GW

True.

And while there might be some uses of such benchmarks on politics etc, combining them with other benchmarks doesn't really seems like a useful benchmark.

Comment by MrThink (ViktorThink) on Benchmarks for Comparing Human and AI Intelligence · 2022-12-12T08:29:21.853Z · LW · GW

Interesting. Even if only a small part of the tasks in the test are poor estimates of general capabilities, it makes the test as a whole less trustworthy.

Comment by MrThink (ViktorThink) on [$20K in Prizes] AI Safety Arguments Competition · 2022-05-28T09:57:10.366Z · LW · GW

For researchers (mainly)

Artificial intelligence isn’t limited in the same ways the human brain is.

Firstly, it isn’t limited to only run on a single set of hardware, it can be duplicated and speeded up to be thousands of times faster than humans, and work on multiple tasks in parallel, assuming powerful enough processors are available.

Further, AI isn’t limited to our intelligence, but can be altered and improved with more data, longer training time and smarter training methods. While the human brain today is superior to AI’s on tasks requiring deep thinking and general intelligence, there is no law preventing AI’s from one day surpassing us.

If artificial intelligence were to surpass human intelligence, it would likely become powerful enough to create an utopia lasting a long long time, or spell the end of humanity.

Thus, doing AI safety research before such an event becomes vital in order to increase the odds of a good outcome.

Comment by MrThink (ViktorThink) on [$20K in Prizes] AI Safety Arguments Competition · 2022-05-27T14:01:51.192Z · LW · GW

What if AI safety could put you on the forefront of sustainable business?

The revolution in AI has been profound, it definitely surprised me, even though I was sitting right there.
-Sergey Brin, Founder of Google

Annual investments in AI increased eightfold from 2015 to 2021, reaching 93 billion usd.

This massive growth is making people ever more dependent on AI, and with that potential risks increases.

Prioritizing AI safety is therefore becoming increasingly important in order to operate a sustainable business, with the benefits of lower risks and improved public perception.


 

Comment by MrThink (ViktorThink) on [$20K in Prizes] AI Safety Arguments Competition · 2022-05-27T13:45:56.326Z · LW · GW

For researchers

What if you could make a massive impact in a quickly growing field of research?

As artificial intelligence continues to advance, the potential risks increase as well.

In the words of Stephen Hawking “The development of full artificial intelligence could spell the end of the human race.

AI safety is a field of research with the purpose to prevent AI from harming humanity, and due to the risks current and future AI is posing, it is a field in which researchers can have a massive impact.

In the three years from 2016 to 2019, AI research has grown from representing 1.8% to 3.8% of all research papers published worldwide.

With the rapid growth of general AI research, we should expect both the importance as well as the funding for AI safety to increase as well.

Click here, to learn how you can start your journey to contribute to AI safety research.

 

Notes

Depending on context, this text can easily be shortened by removing the third to last and second to last paragraphs.

I used this graph to get the growth of AI publications:

Comment by ViktorThink on [deleted post] 2022-05-03T15:33:52.811Z

I believe that given a few years, a company wanting to make a 10 trillion parameter GPT-3 could probably do it for less than these estimates, since 71-358 million usd isn't that much in compute, and for that money extra specialized hardware produced in bulk could be used to bring costs down.

Comment by ViktorThink on [deleted post] 2022-05-03T15:31:07.454Z

Chinchilla is 70 Billion parameters. 10 trillion / 70 billion ≈ 143

According to the paper released with the information about Chinchilla (Training Compute-Optimal Large Language Models” by Hoffmann et al.), they claim "“For every doubling of model size the number of training tokens should also be doubled.”

So if this would be followed, 143 times more data would also be needed, resulting in a 143*143= 20449 increase of compute needed.

Chinchilla probably costed around 1-5 million usd in compute power to train, the cost for a 10 trillion parameter version would cost around 20.5 billion to 122.5 billion usd.

However, this is not a viable alternative, because there is not enough text data available.

A more realistic scenario is to perhaps double the data Chinchilla was trained on (which might not even be easily done), and then 143x the size, so a cost of about 286 million to 1.43 billion usd.

Keep in mind, that GPT-3 was trained on about a quarter of the data that Chinchilla was trained on. So a 10 trillion parameter GPT-3 model might cost around 71.5 million  to 358 million usd.

Comment by MrThink (ViktorThink) on [$20K in Prizes] AI Safety Arguments Competition · 2022-05-02T12:06:10.330Z · LW · GW

The main source of skepticism of AI safety research, is that it’s unknown how advanced current AIs are and how fast it is improving. The most impressive reasoning task I have seen an AI model do is this one, done by PaLM by Google:
 

The model was prompted with this text:

Michael is at that really famous museum in France looking at its most famous painting. However, the artist who made this painting just makes Michael think of his favorite cartoon character from his childhood. What was the country of origin of the thing that the cartoon character usually holds in his hand?

 

And the model answered this:

The most famous painting in the Louvre is the Mona Lisa. The artist who made the Mona Lisa is Leonardo da Vinci. Leonardo da Vinci is also the name of the main character in the cartoon Teenage Mutant Ninja Turtles. Leonardo da Vinci is from Italy. The thing that Leonardo da Vinci usually holds in his hand is a katana. The country of origin of the katana is Japan. The answer is "Japan".

This example can be used in a range of situations.
 

Comment by MrThink (ViktorThink) on [$20K in Prizes] AI Safety Arguments Competition · 2022-05-01T21:35:56.969Z · LW · GW

For tech executives

Could working on AI safety put you on the forefront of sustainable development?

As AI is becoming increasingly advanced and relied upon, the recognition of the importance of AI safety is increasing. If this trend continues AI safety will likely become a core part of sustainability, and the businesses that prioritizes AI safety early, will have a more sustainable business as well as a positive impact and improved public perception.

Comment by MrThink (ViktorThink) on [$20K in Prizes] AI Safety Arguments Competition · 2022-05-01T21:08:13.422Z · LW · GW

As AI technology is heading towards becoming sophisticated enough to potentially end civilizations, the impact and recognition of those working on AI risks will increase.

Comment by MrThink (ViktorThink) on [$20K in Prizes] AI Safety Arguments Competition · 2022-05-01T20:34:37.864Z · LW · GW

General paragraph for non-technical people

What if you could increase your impact by staying ahead of the AI trend?

In a hundred years computers went from being punch card machines, to small tablets in every pocket. In sixty years computers went from displaying only text, to providing an entire virtual reality. In the past three years, AI has become able to write engaging stories and generate photorealistic images. 

If this trend continues, AI is set to cause massive change. If this change is positive or negative depends on what is done today. Therefore actions taken today have the potential of massive impact tomorrow.

Why this paragraph?

Most descriptions focus on how terrible AI can be, and fails to convey what the person reading has to gain personally by taking action. Having impact is something most people desire.

Depending on context, the paragraph can be tweaked to include what action the reader should take in order to have massive impact.

Comment by MrThink (ViktorThink) on [$20K in Prizes] AI Safety Arguments Competition · 2022-05-01T20:34:22.801Z · LW · GW

AI is like fire, we can use it to enter a new era of civilization, or to burn everything down.

Comment by MrThink (ViktorThink) on [$20K in Prizes] AI Safety Arguments Competition · 2022-05-01T20:34:04.303Z · LW · GW

If AI is set to cause massive change, acting now will have a massive impact on the future.

Comment by MrThink (ViktorThink) on [$20K in Prizes] AI Safety Arguments Competition · 2022-05-01T20:33:33.058Z · LW · GW

I will do as Yitz post my submissions as individual replies to this comment. Please let me know if there’s any issues with that.

Comment by MrThink (ViktorThink) on [$20K in Prizes] AI Safety Arguments Competition · 2022-04-27T07:18:07.275Z · LW · GW