Comment by alexei on Crypto quant trading: Intro · 2019-04-18T20:00:15.137Z · score: 2 (1 votes) · LW · GW

Well, I was learning together with the rest of our team. So there were a lot of conversations and an inflow of information from all sorts of places: books, blogs, Quora, advisors, internet research, etc... Honestly, it's all a little helpful to get you thinking along the right lines. So I can't recommend any specific resource, but I can recommend learning broadly.

Okay, one particular resource I'll recommend is: https://www.youtube.com/channel/UC-dLWl8etTtPSGdbbcYffGw

If you have things that you found helpful, please share them here too.

Comment by alexei on Crypto quant trading: Intro · 2019-04-18T19:49:20.916Z · score: 2 (1 votes) · LW · GW

Depending on how complex your strategies are, may be something like this could work: https://www.youtube.com/watch?v=P_YmXyUsb9A

Comment by alexei on Crypto quant trading: Intro · 2019-04-18T19:33:45.956Z · score: 2 (1 votes) · LW · GW

Hmm, that's a good question. I haven't looked into it too much, but I'd definitely try to leverage a 3rd party platform for hosting or at least use a 3rd party library to do the trading. https://enigma.co/catalyst/ was kind of decent when I used it for a while a year ago. By now, if they're still improving it, it's probably pretty good. One advantage there is that they have the data too. https://github.com/ccxt/ccxt is a wonderful API that generalizes how you interact with exchanges.

Hosting wise, AWS (what we use) or Google Cloud would both work. I personally like the latter a lot more.

Comment by alexei on Crypto quant trading: Intro · 2019-04-17T23:41:41.344Z · score: 22 (7 votes) · LW · GW

This comment was written in response to you feeling confused about what strategies to explore. I might write a fuller post about it, but for now here're the thoughts off the top of my head:

Calling marking anti-inductive is correct, but it's not helpful when trying to find strategies (as you've just noticed). I'd break down the strategy research process steps into:

1) Can you find a strategy (algorithm + data) that historically has performed well?

2) Can you find this strategy in such a way so as not to find a ton of other strategies that worked by random chance?

3) What % of the market has figured out this strategy?

From Eliezer's post:

Let's say you see me flipping a coin.  It is not necessarily a fair coin.  It's a biased coin, and you don't know the bias.  I flip the coin nine times, and the coin comes up "heads" each time.  I flip the coin a tenth time.  What is the probability that it comes up heads?
If you answered "ten-elevenths, by Laplace's Rule of Succession", you are a fine scientist in ordinary environments, but you will lose money in finance.
In finance the correct reply is, "Well... if everyone else also saw the coin coming up heads... then by now the odds are probably back to fifty-fifty."

Right. But if it's slightly more complicated than just looking at the coin, then suddenly we can have an edge:

1) May be not everyone can write the code to compute which way the coin is facing (algorithm). May be not everyone can see the coin (data).

2) May be other people are looking at the weather and the weather has been sunny nine days in a row.

3) May be not everyone can run their algorithm fast enough to make the trading decision in time. May be others figured out this strategy, but they're not confident in it enough to deploy a lot of money.

So once you find your strategy, you might be in a pretty small group of people who have discovered it. So you'll be fine in proportion to how much money is allocated to this strategy vs how much capacity it has.

And the lesson is: aim for a strategy complexity that's simple enough to pass 2), but complicated enough that most people haven't found it. And the bar for that is actually not that high (at least in crypto).

Crypto quant trading: Intro

2019-04-17T20:52:53.279Z · score: 58 (18 votes)
Comment by alexei on Highlights from "Integral Spirituality" · 2019-04-15T04:35:35.106Z · score: 2 (1 votes) · LW · GW

Awesome! I’ve really enjoyed reading this. Quite a lot of ideas resonated in a surprising new way. I might actually read this book. Thanks for writing up the summary!

Comment by alexei on Towards a Quieter Life · 2019-04-07T23:11:32.332Z · score: 3 (2 votes) · LW · GW

Reminds me of this: https://youtu.be/vl-44jDYDJQ

Comment by alexei on How do people become ambitious? · 2019-04-05T18:21:59.748Z · score: 10 (5 votes) · LW · GW

Ruby recently wrote this, which seems relevant: https://www.lesswrong.com/posts/k2KwBEq4ToXnvZSbz/on-the-nature-of-agency

Comment by alexei on User GPT2 is Banned · 2019-04-02T20:29:46.398Z · score: 29 (16 votes) · LW · GW

I think overall I just appreciate that you guys did something for April 1st. It made the website / community feel a bit more alive.

Comment by alexei on Subagents, akrasia, and coherence in humans · 2019-03-30T08:07:27.262Z · score: 6 (3 votes) · LW · GW

Having studied and achieved stream entry (in the Vipassana tradition), I very much doubt many people have stumbled into it. Although for clarity, what % of the population are we taking about? Quick fermi: from what I’ve seen in the spiritual community about 1/1000 have achieved stream entry spontaneously / easily. Out of my bubble, I’d say 1/100 is spiritually inclined. Then I’d add another factor of at least 1/100 to control for my bubble being the Bay Area.

The reason why I doubt it is because most people will tell you (and have written) that it has take them (and people they know) many years and intense practice to get it.

I do think a lot of people have gotten A&P though.

Comment by alexei on The Unexpected Philosophical Depths of the Clicker Game Universal Paperclips · 2019-03-29T02:03:34.221Z · score: 5 (5 votes) · LW · GW

Not long after Lantz made Bostrom’s ideas playable, a young filmmaker named Alberto Roldán followed a link to the game on Facebook and was entranced. He contacted Lantz about making Universal Paperclips into a feature-length film. Roldán is close to finishing a draft of the script, and he has already curried the interest of several producers. His version will be much more explicitly peopled than the game itself is. “It has to involve characters who feel like it’s their story,” he told me. “As you play, you keep getting these increasingly large multimillion dollar gifts to sort of placate your ostensible keepers. We know who these characters are in our modern world: hyper-rich tech titans who are not super interested in questioning the thing that has enriched them—with the justification that they are healing the world at large.”

Comment by alexei on "Other people are wrong" vs "I am right" · 2019-03-18T17:21:16.477Z · score: 7 (4 votes) · LW · GW

One way I’ve experienced this was playing poker in a casino. I was pretty decent and could see a lot of the the mistakes other people were making. I thought that meant I should just be able to come in and win their money, but I was sorely mistaken. It was only after a full week of training and practice that I was able to get to that level. (A week sounds like not a lot of time, but it took me a few years of playing on and off to get to that level + realization,)

Comment by alexei on Active Curiosity vs Open Curiosity · 2019-03-15T01:54:48.154Z · score: 13 (8 votes) · LW · GW

Yes! For me one area where it’s super clear is relationships. On first dates I’m 100% active curiosity. I’ll ask lots of questions, trying to get to the heart of who the person across the dinner table is. But after a while, once I’ve “accepted” them, I switch more to open curiosity. It feels more playful. I’m less interested in their version and more interested in exploring together. There’s trust.

Comment by alexei on Formalising continuous info cascades? [Info-cascade series] · 2019-03-14T05:16:15.467Z · score: 1 (3 votes) · LW · GW

Note that the post you linked by Johnicholas contains a mistake that the author admits invalidates his point.

[Link] OpenAI LP

2019-03-12T23:22:59.861Z · score: 15 (5 votes)
Comment by alexei on What Vibing Feels Like · 2019-03-11T23:58:17.448Z · score: 8 (6 votes) · LW · GW

Actually I take that last part back. I could see people who love to vibe about rationality. I think I actually do too, just at a certain level of depth / casually,

Comment by alexei on What Vibing Feels Like · 2019-03-11T23:55:54.914Z · score: 3 (2 votes) · LW · GW

Yes, but also once you get good at vibing you want to go vibe with people on your favorite frequency / topic. Also I think some topics are more “vibable“ than others, and rationality is just not one of them.

Comment by alexei on Have a camel! · 2019-03-01T05:50:41.786Z · score: 2 (1 votes) · LW · GW

Also I think this post gets at part of what you’re pointing at: https://www.lesswrong.com/posts/4QemtxDFaGXyGSrGD/other-people-are-wrong-vs-i-am-right

Comment by alexei on Have a camel! · 2019-03-01T05:49:20.152Z · score: 3 (2 votes) · LW · GW

Camel: you have a piece of data, then someone asks if you’ve seen a camel, then you realize that fits with your data. Elephant: you have a piece of data and so you imagine what the “thing” might be. But nobody walked in and asked if someone has seen an elephant.

Comment by alexei on A Strange Situation · 2019-02-18T21:51:54.052Z · score: 4 (3 votes) · LW · GW

Why do you think you should be reading / learning more vs going and doing / making something?

Comment by alexei on Limiting an AGI's Context Temporally · 2019-02-17T21:51:32.203Z · score: 13 (5 votes) · LW · GW

Seems like it would throw a brick at you, because it wanted to throw a brick, not caring that in 2 seconds it’ll hit your face. (You can probably come up with a better example with a slightly longer timeframe.)

Comment by alexei on The RAIN Framework for Informational Effectiveness · 2019-02-13T23:48:56.217Z · score: 2 (1 votes) · LW · GW

RAIN is easiest and most memorable.

Comment by alexei on When should we expect the education bubble to pop? How can we short it? · 2019-02-12T20:12:50.495Z · score: 7 (4 votes) · LW · GW

I learned recently that some states used to offer an equivalent of "forever stamps" for education. Meaning you pay $X at any time and you've guaranteed your payment for a state university in the future. Obviously, they discontinued it, since the costs rose and they lost money. But if you wanted to short *education cost*, you'd basically want to sell these guarantees yourself.

Comment by alexei on Probability space has 2 metrics · 2019-02-10T07:47:23.006Z · score: 4 (4 votes) · LW · GW

I don’t think I’ve read this view before, or if I have, I’ve forgotten it. Thanks for writing this up!

Comment by alexei on The Case for a Bigger Audience · 2019-02-10T07:39:59.366Z · score: 13 (7 votes) · LW · GW

I’ve been reading basically every post for the past few months. I don’t usually leave comments though, unless it’s to support and thank the author. (Thanks for writing this! Funny enough I also noticed recently how few comments there are, and it seemed worth bringing up.) I guess I feel like I just don’t have much to add to most posts.

Comment by alexei on The Question Of Perception · 2019-01-30T00:48:13.535Z · score: 2 (1 votes) · LW · GW

“view them as a set of open-ended concepts that have yet to reach a satisfactory conclusion“ is a good description of how I feel about this post.

Comment by alexei on Río Grande: judgment calls · 2019-01-27T04:24:43.674Z · score: 5 (3 votes) · LW · GW

I call it “making an executive decision.” And I used that term before getting into startups.

Link: That Time a Guy Tried to Build a Utopia for Mice and it all Went to Hell

2019-01-23T06:27:05.219Z · score: 15 (6 votes)
Comment by alexei on Some Thoughts on My Psychiatry Practice · 2019-01-17T06:00:35.763Z · score: 15 (9 votes) · LW · GW

Thanks for writing your thoughts out loud. Curious to follow your progress.

Comment by alexei on Optimizing for Stories (vs Optimizing Reality) · 2019-01-07T22:23:01.879Z · score: 4 (2 votes) · LW · GW
Of course, truly effective tea plus a well-conveyed story about its great properties will generate more sales than effective tea or a good story alone.

Sometimes not, I think. It's almost like a measure of the efficiency / effectiveness of the given market. If the market is really good at recognizing reality, then you don't need to tell a story. (Basic software libraries are like that: do they give the compute the right thing? If yes, then it's good.) If the market is not at recognizing reality, then creating stories is often way cheaper than then doing the real thing. (And also transfers better across domains.)

Comment by alexei on The E-Coli Test for AI Alignment · 2018-12-17T15:45:54.273Z · score: 2 (1 votes) · LW · GW

Is there anything in the world that we know of that does alignment for something else? Can we say that humans are doing "coherent extrapolated volition" for evolution? Keeping in mind that under this view, evolution itself would evolve and change into something more complex and may be better.

Comment by alexei on The E-Coli Test for AI Alignment · 2018-12-17T15:41:48.376Z · score: 2 (1 votes) · LW · GW

I think there's a bias when we consider optimizing for X's values to consider only X without its environment. But the environment gave rise to X and actually much of X doesn't make sense without it. So I think to some extent we also would need to find ways to do alignment on the environment itself. And that means to some extent helping evolution.

Comment by alexei on The E-Coli Test for AI Alignment · 2018-12-17T15:36:45.546Z · score: 4 (3 votes) · LW · GW

First off, great thought experiment! I like it, and it was a nice way to view the problem.

The most obvious answer is: “Wow, we sure don’t know how to help. Let’s design a smarter intelligence that’ll know how to help better.”

At that point I think we’re running the risk of passing the buck forever. (Unless we can prove that process terminates.) So we should probably do at least something. Instead of trying to optimize, I’d focus on doing things that are most obvious. Like helping it not to die. And making sure it has food.

Comment by alexei on Argue Politics* With Your Best Friends · 2018-12-16T01:30:21.457Z · score: 9 (5 votes) · LW · GW

Just wanted to say I’ve been reading all your recent posts. And I really like all of them, the ideas you lay out, and how you talk about them. You really make LW worth coming to almost every day. Thank you! FWIW, if you continue down this path, I could see you having your own SSC-sized community.

Comment by alexei on Review: Slay the Spire · 2018-12-10T06:03:59.540Z · score: 4 (2 votes) · LW · GW

I love that game too. Thanks for the write up, it’s been fun to read your take on it. I haven’t played it recently, so didn’t know they added the ending. I’ll come back and play it again.

Comment by alexei on What's up with Arbital? · 2017-08-10T19:19:56.537Z · score: 3 (1 votes) · LW · GW

This will take a long time to load, but it's comprehensive: https://arbital.com/explore/math/

Comment by alexei on What's up with Arbital? · 2017-04-13T20:58:03.116Z · score: 3 (1 votes) · LW · GW

I'm not sure when you tried. It works right now.

Comment by alexei on What's up with Arbital? · 2017-04-13T20:56:48.530Z · score: 3 (1 votes) · LW · GW

Hmm, I'm skeptical a barter system would work. I don't think I've seen a successful implementation of it anywhere, though I do hear about people trying.

Yes, we've considered paying people, but that's not scalable. (A good 3-5 page explanation might take 10 hours to write.)

Comment by alexei on What's up with Arbital? · 2017-04-10T23:09:43.374Z · score: 3 (1 votes) · LW · GW

It's not open sourced.

The pages might take a while to load (up to 30 seconds).

Comment by alexei on What's up with Arbital? · 2017-04-03T20:20:28.772Z · score: 3 (1 votes) · LW · GW

I can imagine a neural-activation-like effect coming out of that, where frequently co-active posts naturally rise to the top of each other's links and become threads or topics.

Not sure what you mean by this.

Comment by alexei on What's up with Arbital? · 2017-04-01T17:06:32.277Z · score: 5 (4 votes) · LW · GW

My guess for Wikipedia's success is that they were one of the first; and there was more of a sense of an online community back then. Also it's easier to create Wikipedia content than, say, a good explanation. StackOverflow succeeded because asking and answering questions is pretty easy, you get instant feedback, and they got community management right. (They solved exactly one problem well!) The founders were also really well known so it was easy for them to seed the platform.

I can't open-source the platform as long as I'm doing the for-profit venture, since the platforms are too similar. However, if at some point I have to stop, then I'll be happy to open source everything at that point.

Comment by alexei on What's up with Arbital? · 2017-04-01T17:01:40.983Z · score: 3 (1 votes) · LW · GW

I'm not into persuading people. :) If you want to write, go for it. I still think Arbital is a really good platform for writing up math explanations.

Comment by alexei on What's up with Arbital? · 2017-03-30T16:48:15.776Z · score: 3 (1 votes) · LW · GW

Not someone with sufficient authority, just the blog owner. That seems fair though. You can create you own blog and then you would be in charge of which comments to approve.

Comment by alexei on What's up with Arbital? · 2017-03-30T16:20:00.003Z · score: 3 (1 votes) · LW · GW

Yes, but that's not "invite-only".

Comment by alexei on What's up with Arbital? · 2017-03-30T15:29:46.778Z · score: 4 (2 votes) · LW · GW

How is it invite only? Are you talking about the comment section?

Originally the plan was to do exactly that if we couldn't figure how to build a "joyful maze": just throw open the doors and see what people do with it. Unfortunately there is still a significant amount of work left to do that well, and right now I'm more optimistic about the new platform than I am about scavenging the current version.

Comment by alexei on What's up with Arbital? · 2017-03-30T15:20:44.189Z · score: 1 (5 votes) · LW · GW

Here are like 5 pages explaining all the visions: https://arbital.com/p/more_about_arbital/

Basically what we tried is: "let's figure out how people are supposed to have truth-seeking conversations, build a platform that facilities that, and then grow it." Step 1 is very hard. Step 3 is made harder because your platform only attracts truth-seeking people.

New approach: "build a platform that facilities communication, grow it, then shape the ongoing discussion to be more truth-seeking." Step 1 is still hard, but not made harder. Step 3 sounds a lot more doable.

Comment by alexei on What's up with Arbital? · 2017-03-30T15:15:55.467Z · score: 6 (6 votes) · LW · GW

Don't use the outside view. Use your brain. If Arbital was confusing, but you didn't look closer at it, then you didn't use your brain. If MIRI seems confusing and you don't look closer at it, then you aren't using your brain. The whole concept of "smart people" whom you can trust with anything is just wrong. There are only niche experts.

My two cents: Arbital had very little to do with MIRI, aside from Arbital being Eliezer's idea. But this was definitely out of his realm of expertise. MIRI/AI stuff is not.

Comment by alexei on What's up with Arbital? · 2017-03-30T15:09:05.739Z · score: -2 (2 votes) · LW · GW

It's a complicated answer and also somewhat outside of the topic I want to discuss here (which is Arbital 1.0). For part of the answer see: http://lesswrong.com/lw/otq/whats_up_with_arbital/dqbc

Comment by alexei on What's up with Arbital? · 2017-03-30T15:07:04.730Z · score: 3 (1 votes) · LW · GW
  1. Currently it's not clear to anyone what Arbital is, what it can do, who it's for, etc.. It needs to solve a real problem and present itself as solving that clear problem.
  2. The tech we used is now somewhat obsolete. The codebase has accumulated a lot of unnecessary features. Also Google Material UI turned out to be too heavyweight and not as pleasant to design with as I thought initially. (These are all arguments for remaking the platform.)
  3. The blogging platform will be "as open and as inviting of contribution as possible."
Comment by alexei on What's up with Arbital? · 2017-03-30T14:55:42.162Z · score: 4 (2 votes) · LW · GW

Yes. There is already a pretty large demand for blogging platforms, and Arbital 2.0 will have features which will make it a much better option for some users. I'll also be personally reaching out to a lot of bloggers to interview them about their experience / wishes. I'll also be testing the key value propositions with a graphic that's being created right now.

But also sometimes you just have to go ahead and build the thing to really test it.

Comment by alexei on What's up with Arbital? · 2017-03-30T14:49:36.188Z · score: 7 (6 votes) · LW · GW

Yes, many students would benefit from a math explanation platform. But it was hard for us to find writers, and we weren't getting as much traction with them as we wanted. We reached out to some forums and to many individuals. That version of Arbital was also promoted by Eliezer on FB. When we switched away from math, it wasn't because we thought it was hopeless. We had a lot of ideas left to try out. But when it's not going well, you have to call it quits at some point, and so we did. There was also the consideration that if we built a platform for (math) explanations, it would be hard to eventually transition to a platform that solved debates (which always seemed like the more important part).

I think if someone wanted to give it a shot with another explanation platform and had a good strategy for getting writers, I'd feel pretty optimistic about their chance of success.

Comment by alexei on What's up with Arbital? · 2017-03-30T03:42:43.274Z · score: 4 (2 votes) · LW · GW

See this comment: http://lesswrong.com/lw/otq/whats_up_with_arbital/dq9h

I think we likely made a mistake with respect to openness, but it's not obvious when/how. Probably the biggest problem is that we couldn't settle on what we wanted the users to do once they were on the platform.

Comment by alexei on What's up with Arbital? · 2017-03-30T03:35:58.904Z · score: -2 (2 votes) · LW · GW

Noted, but I disagree.

What's up with Arbital?

2017-03-29T17:22:21.751Z · score: 24 (27 votes)

Toy problem: increase production or use production?

2014-07-05T20:58:48.962Z · score: 4 (5 votes)

Quantum Decisions

2014-05-12T21:49:11.133Z · score: 1 (6 votes)

Personal examples of semantic stopsigns

2013-12-06T02:12:01.708Z · score: 44 (49 votes)

Maximizing Your Donations via a Job

2013-05-05T23:19:05.116Z · score: 115 (117 votes)

Low hanging fruit: analyzing your nutrition

2012-05-05T05:20:14.372Z · score: 7 (8 votes)

Robot Programmed To Love Goes Too Far (link)

2012-04-28T01:21:45.465Z · score: -5 (12 votes)

I'm starting a game company and looking for a co-founder.

2012-03-18T00:07:01.670Z · score: 16 (23 votes)

Water Fluoridation

2012-02-17T04:33:00.064Z · score: 1 (9 votes)

What happens when your beliefs fully propagate

2012-02-14T07:53:25.005Z · score: 22 (50 votes)

Rationality and Video Games

2011-09-18T19:26:01.716Z · score: 6 (11 votes)

Credit card that donates to SIAI.

2011-07-22T18:30:35.207Z · score: 5 (8 votes)

Futurama does an episode on nano-technology.

2011-06-27T02:44:14.496Z · score: 3 (6 votes)

Considering all scenarios when using Bayes' theorem.

2011-06-20T18:11:34.810Z · score: 9 (10 votes)

Discussion for Eliezer Yudkowsky's paper: Timeless Decision Theory

2011-01-06T00:28:29.202Z · score: 10 (11 votes)

Life-tracking application for android

2010-12-11T01:48:11.676Z · score: 20 (21 votes)