Posts

Delta's of Change 2024-03-19T21:03:28.204Z
Navigating the Attackspace 2023-12-12T13:59:14.738Z
WHO Biological Risk warning 2023-04-25T15:10:14.985Z
We might need to rethink the Hard Reset , aka the AI Pause. 2023-03-30T21:38:44.564Z
Negative Expertise 2023-01-14T00:51:30.280Z
Distributed blind review site for papers 2022-04-21T22:51:45.625Z
What are the causality effects of an agents presence in a reinforcement learning environment 2022-03-01T21:57:15.928Z
What are some ways to do a PhD without an educational institution 2022-03-01T21:57:15.896Z

Comments

Comment by Jonas Kgomo (jonas-kgomo) on SAE-VIS: Announcement Post · 2024-03-31T18:25:35.792Z · LW · GW

is this something that can work for a hosted web version (npm i or api) 

Comment by jonas-kgomo on [deleted post] 2023-06-21T21:07:05.121Z
Comment by jonas-kgomo on [deleted post] 2023-06-19T20:09:04.865Z
Comment by jonas-kgomo on [deleted post] 2023-06-19T18:34:59.099Z

For context, i mean posting a link without much comment, but people can engage on the post in the comment section. 

Comment by jonas-kgomo on [deleted post] 2023-06-18T22:47:25.028Z

Why doesn't LW have ShowHN style posts, no comments , just a link to some page? 

Comment by Jonas Kgomo (jonas-kgomo) on MIRI's technical research agenda · 2023-06-09T19:16:55.169Z · LW · GW
Comment by jonas-kgomo on [deleted post] 2023-04-25T14:23:48.060Z
Comment by Jonas Kgomo (jonas-kgomo) on gwern's Shortform · 2022-07-12T21:06:05.464Z · LW · GW

Gwern,  i wonder what you think about this question i asked a while ago on causality in relation to the article you posted on reddit. Do we need more general causal agents for addressing issues in RL environments? 

Apologies for posting here, i didn't know how to mention/tag someone on a post in LW. 

https://www.lesswrong.com/posts/BDf7zjeqr5cjeu5qi/what-are-the-causality-effects-of-an-agents-presence-in-a?commentId=xfMj3iFHmcxjnBuqY

Comment by jonas-kgomo on [deleted post] 2022-05-05T11:02:01.696Z

Is bias simply human in the loop problem(is it something that can be solved by data refinement and having diverse programmers), or is it also related to explainability of AI, the fact that we can not explain why AI decided to make some decisions.  A simple example would be if an AGI was supposed to identify extreme ideology in a persons posts on social media: one  AI (honest) tells us an extreme person A is extreme, while the other AI (dishonest) tells us an extreme person B is not extreme (even thou it knows the person is extreme). In the above scenario, having a human trying to understand if there is bias would be futile, since the untruthful AI would basically perpetuate bias by lying about there being no bias. Does this mean algorithmic bias is beyond human in the loop, but also an architectural bias (if we had more causal models and logic in neural networks then we could have less of such bias and side effects).

Comment by jonas-kgomo on [deleted post] 2022-04-21T22:50:49.820Z