✓ comments on "Self-Consciousness wants to make"
✓ tussle with Ruby on "Causal vs. Social Reality"
✓ The End (of Sequences)
+_ retrieve own-blog links for "futurist-themed delusions"
+_ URL for "Why Quantum?"
_ double-check that "Changing Emotions" was in January
_ No such thing as a tree
_ Yudkowsky on AlphaGo
https://www.lesswrong.com/posts/wustx45CPL5rZenuo/no-safe-defense-not-even-science
> I'm not sure that human beings realistically _can_ trust and think at the same time.
+"Why Quantum" has another reference to "doesn't seem possible to think and trust"
+
https://www.lesswrong.com/posts/t6Fe2PsEwb3HhcBEr/the-litany-against-gurus
https://www.lesswrong.com/posts/TQSb4wd6v5C3p6HX2/the-pascal-s-wager-fallacy-fallacy#pART2rjzcmqATAZio
> egoists, but not altruists, should immediately commit suicide in case someone is finishing their AI project in a basement, right now.
-(I remembered this as suggesting some plausibility in sudden Singularity even then, but in context it's more clearly in thought-experimental mode)
\ No newline at end of file
+(I remembered this as suggesting some plausibility in sudden Singularity even then, but in context it's more clearly in thought-experimental mode)
+
+from "Go Forth and Create the Art"—
+> To the best of my knowledge there is _no_ true science that draws its strength from only one person. To the best of my knowledge that is _strictly_ an idiom of cults. A true science may have its heroes, it may even have its lonely defiant heroes, but _it will have more than one_.
+
+contrast the sneering at Earth people with the attitude in "Whining-Based Communities"
+
+from "Why Quantum?"—
+> But would you believe that I had such strong support, if I had not shown it to you in full detail? Ponder this well. For I may have other strong opinions. And it may seem to you that _you_ do't see any good reason to form such strong beliefs. Except this is _not_ what you will see; you will see simply that there _is_ no good reason for the strong belief, that there _is_ no strong support one way or the other. For our first-order beliefs are how the world seems to _be_. And you may think, "Oh, Eliezer is just opinionated—forming strong beliefs in the absence of lopsided support." And I will not have time to do another couple of months worth of blog posts.
+>
+> I am _very_ far from infallible, but I do not hold strong opinions at random.