+John Wentworth—
+> I chose the "train a shoulder advisor" framing specifically to keep my/Eliezer's models separate from the participants' own models.
+https://www.greaterwrong.com/posts/Afdohjyt6gESu4ANf/most-people-start-with-the-same-few-bad-ideas#comment-zL728sQssPtXM3QD9
+
+https://twitter.com/ESYudkowsky/status/1355712437006204932
+> A "Physics-ist" is trying to engage in a more special human activity, hopefully productively, where they *think* about light in order to use it better.
+
+Wentworth on my confusion about going with the sqaured-error criterion in "Unnatural Categories"—
+> I think you were on the right track with mutual information. They key insight here is not an insight about what metric to use, it's an insight about the structure of the world and our information about the world. [...] If we care more about the rough wall-height than about brick-parity, that’s because the rough wall-height is more relevant to the other things which we care about in the world. And that, in turn, is because the rough wall-height is more relevant to more things in general. Information about brick-parity just doesn’t propagate very far in the causal graph of the world; it's quickly wiped out by noise in other variables. Rough wall-height propagates further.
+
+not interested in litigating "lying" vs. "rationalizing" vs. "misleading-by-implicature"; you can be _culpable_ for causing people to be misled in a way that isn't that sensitive to what exactly was going on in your head
+
+-----
+
+https://www.facebook.com/yudkowsky/posts/pfbid02ZoAPjap94KgiDg4CNi1GhhhZeQs3TeTc312SMvoCrNep4smg41S3G874saF2ZRSQl?comment_id=10159410429909228&reply_comment_id=10159410748194228
+
+> Zack, and many others, I think you have a vulnerability where you care way too much about the reasons that bullies give for bullying you, and the bullies detect that and exploit it.
+
+
+
+> Everyone. (Including organizers of science fiction conventions.) Has a problem of "We need to figure out how to exclude evil bullies." We also have an inevitable Kolmogorov Option issue but that should not be confused with the inevitable Evil Bullies issue, even if bullies attack through Kolmogorov Option issues.
+
+----
+
+Someone else's Dumb Story that you can read about on someone else's blog
+
+all he does these days is sneer about Earth people, but he _is_ from Earth—carrying on the memetic legacy of Richard Feynmann and Douglas Hofstadter and Greg Egan
+
+"Robust Cooperation in the Prisoner's Dilemma" https://arxiv.org/abs/1401.5577
+
+https://www.lesswrong.com/posts/anCubLdggTWjnEvBS/your-rationality-is-my-business
+> One of those interests is the human pursuit of truth, which has strengthened slowly over the generations (for there was not always Science). I wish to strengthen that pursuit further, in _this_ generation. That is a wish of mine, for the Future. For we are all of us players upon that vast gameboard, whether we accept the responsibility or not.
+
+https://www.washingtonexaminer.com/weekly-standard/be-afraid-9802
+
+https://www.lesswrong.com/posts/TQSb4wd6v5C3p6HX2/the-pascal-s-wager-fallacy-fallacy#pART2rjzcmqATAZio
+> egoists, but not altruists, should immediately commit suicide in case someone is finishing their AI project in a basement, right now.
+(I remembered this as suggesting some plausibility in sudden Singularity even then, but in context it's more clearly in thought-experimental mode)
+
+from "Go Forth and Create the Art"—
+> To the best of my knowledge there is _no_ true science that draws its strength from only one person. To the best of my knowledge that is _strictly_ an idiom of cults. A true science may have its heroes, it may even have its lonely defiant heroes, but _it will have more than one_.
+
+contrast the sneering at Earth people with the attitude in "Whining-Based Communities"
+
+from "Why Quantum?"—
+> But would you believe that I had such strong support, if I had not shown it to you in full detail? Ponder this well. For I may have other strong opinions. And it may seem to you that _you_ do't see any good reason to form such strong beliefs. Except this is _not_ what you will see; you will see simply that there _is_ no good reason for the strong belief, that there _is_ no strong support one way or the other. For our first-order beliefs are how the world seems to _be_. And you may think, "Oh, Eliezer is just opinionated—forming strong beliefs in the absence of lopsided support." And I will not have time to do another couple of months worth of blog posts.
+>
+> I am _very_ far from infallible, but I do not hold strong opinions at random.