X-Git-Url: http://unremediatedgender.space/source?a=blobdiff_plain;f=notes%2Fa-hill-of-validity-sections.md;h=71a7eb2ef443e710442e03c36a8e93458f98c7c9;hb=30c3567400705b877008cedbec46348642d36d7c;hp=0fd8c19e69cb1be13b1f01a061a1485421e8e74d;hpb=0d7e6c46bb0268b6aaad8221a151f626c90edb50;p=Ultimately_Untrue_Thought.git diff --git a/notes/a-hill-of-validity-sections.md b/notes/a-hill-of-validity-sections.md index 0fd8c19..71a7eb2 100644 --- a/notes/a-hill-of-validity-sections.md +++ b/notes/a-hill-of-validity-sections.md @@ -1,15 +1,18 @@ with internet available— +_ "What the Tortise Said" link +_ 13th century word meanings _ the comment from old-OB where someone brought up the Dolphin objection and Yudkowsky linked "Mutual Information" in reply _ the other Twitter conversation incl. Tail where Yudkowsky says he's not taking a stand _ the exchange on old OB where I said to acknowledge that not all men _want_ to be masculine _ screenshot Rob's Facebook comment which I link _ compile Categories references from the Dolphin War +_ dates of subsequent philosophy-of-language posts far editing tier— +_ clarify why Michael thought Scott was "gaslighting" me, include "beeseech bowels of Christ" _ address the "maybe it's good to be called names" point from "Hill" thread _ 2019 Discord discourse with Alicorner _ edit discussion of "anti-trans" side given that I later emphasize that "sides" shouldn't be a thing -_ clarify why Michael thought Scott was "gaslighting" me, include "beeseech bowels of Christ" _ the right way to explain how I'm respecting Yudkowsky's privacy _ explain the adversarial pressure on privacy norms _ first EY contact was asking for public clarification or "I am being silenced" (so Glomarizing over "unsatisfying response" or no response isn't leaking anything Yudkowksy cares about) @@ -30,6 +33,8 @@ _ January 2019 meeting with Ziz and Gwen _ better summary of Littman + + people to consult before publishing, for feedback or right of objection— _ Iceman _ Ben/Jessica @@ -984,3 +989,15 @@ Thoughts on your proposed cruxes: 1 (harmful inferences) is an unworkable AI des https://twitter.com/ESYudkowsky/status/1436025983522381827 > Well, Zack hopefully shouldn't see this, but I do happen to endorse everything you just said, for your own personal information. + + +[SECTION about monastaries (with Ben and Anna in April 2019) +I complained to Anna: "Getting the right answer in public on topic _X_ would be too expensive, so we won't do it" is _less damaging_ when the set of such Xes is _small_. It looked to me like we added a new forbidden topic in the last ten years, without rolling back any of the old ones. + +"Reasoning in public is too expensive; reasoning in private is good enough" is _less damaging_ when there's some sort of _recruiting pipeline_ from the public into the monasteries: lure young smart people in with entertaining writing and shiny math, _then_ gradually undo their political brainwashing once they've already joined your cult. (It had [worked on me](/2021/May/sexual-dimorphism-in-the-sequences-in-relation-to-my-gender-problems/)!) + +I would be sympathetic to "rationalist" leaders like Anna or Yudkowsky playing that strategy if there were some sort of indication that they had _thought_, at all, about the pipeline problem—or even an indication that there _was_ an intact monastery somewhere. +] + +> Admitting something when being pushed a little, but never thinking it spontaneously and hence having those thoughts absent from your own thought processes, remains not sane. +https://twitter.com/ESYudkowsky/status/1501218503990431745