+To his credit, he _will_ admit that he's only willing to address a selected subset of arguments—but while doing so, he claims an absurd "confidence in [his] own ability to independently invent everything important that would be on the other side of the filter and check it [himself] before speaking" while _simultaneously_ blatantly mischaracterizing his opponents' beliefs! ("Gendered Pronouns For Everyone and Asking To Leave The System Is Lying" doesn't pass anyone's [ideological Turing test](https://www.econlib.org/archives/2011/06/the_ideological.html).)
+
+Counterarguments aren't completely causally _inert_: if you can make an extremely strong case that Biological Sex Is Sometimes More Relevant Than Self-Declared Gender Identity, Yudkowsky will put some effort into coming up with some ingenious excuse for why he _technically_ never said otherwise, in ways that exhibit generally rationalist principles. But at the end of the day, Yudkowsky is going to say what he needs to say in order to protect his reputation, as is personally prudent.
+
+Even if one were to agree with this description of Yudkowsky's behavior, it doesn't immediately follow that Yudkowsky is making the wrong decision. Again, "bad faith" is meant as a literal description that makes predictions about behavior, not a contentless attack—maybe there are some circumstances in which engaging some amount of bad faith is the right thing to do, given the constraints one faces! For example, when talking to people on Twitter with a very different ideological background from me, I sometimes anticipate that if my interlocutor knew what I was actually thinking, they wouldn't want to talk to me, so I engage in a bit of what is sometimes called ["concern trolling"](https://geekfeminism.fandom.com/wiki/Concern_troll): I take care to word my replies in a way that makes it look like I'm more ideologically aligned with them than I actually am. (For example, I [never say "assigned female/male at birth" in my own voice on my own platform](/2019/Sep/terminology-proposal-developmental-sex/), but I'll do it in an effort to speak my interlocutor's language.) I think of this as the _minimal_ amount of strategic bad faith needed to keep the conversation going, to get my interlocutor to evaluate my argument on its own merits, rather than rejecting it for coming from an ideological enemy. In cases such as these, I'm willing to defend my behavior as acceptable—there _is_ a sense in which I'm being deceptive by optimizing my language choice to make my interlocutor make bad guesses about my ideological alignment, but I'm comfortable with that amount and scope of deception in the service of correcting the distortion where I don't think my interlocutor _should_ be paying attention to my personal alignment.
+
+That is, my bad faith concern-trolling gambit of deceiving people about my ideological alignment in the hopes of improving the discussion seems like something that makes our collective beliefs about the topic-being-argued-about _more_ accurate. (And the topic-being-argued-about is presumably of greater collective interest than which "side" I personally happen to be on.)
+
+In contrast, the "it is sometimes personally prudent [...] to post your agreement with Stalin" gambit is the exact reverse: it's _introducing_ a distortion into the discussion in the hopes of correcting people's beliefs about the speaker's ideological alignment. (Yudkowsky is not a right-wing Bad Guy, but people would tar him as a right-wing Bad Guy if he ever said anything negative about trans people.) This doesn't improve our collective beliefs about the topic-being-argued about; it's a _pure_ ass-covering move.
+
+Yudkowsky names the alleged fact that "people do _know_ they're living in a half-Stalinist environment" as a mitigating factor. But the _reason_ censorship is such an effective tool in the hands of dictators like Stalin is because it ensures that many people _don't_ know—and that those who know (or suspect) don't have [game-theoretic common knowledge](https://www.lesswrong.com/posts/9QxnfMYccz9QRgZ5z/the-costly-coordination-mechanism-of-common-knowledge#Dictators_and_freedom_of_speech) that others do too.
+
+Zvi Mowshowitz has [written about how the false assertion that "everybody knows" something](https://thezvi.wordpress.com/2019/07/02/everybody-knows/) is typically used justify deception: if "everybody knows" that we can't talk about biological sex (the reasoning goes), then no one is being deceived when our allegedly truthseeking discussion carefully steers clear of any reference to the reality of biological sex when it would otherwise be extremely relevant.
+
+But if it were _actually_ the case that everybody knew (and everybody knew that everybody knew), then what would be the point of the censorship? It's not coherent to claim that no one is being harmed by censorship because everyone knows about it, because the entire appeal and purpose of censorship is precisely that _not_ everybody knows and that someone with power wants to _keep_ it that way.
+
+For the savvy people in the know, it would certainly be _convenient_ if everyone secretly knew: then the savvy people wouldn't have to face the tough choice between acceding to Power's demands (at the cost of deceiving their readers) and informing their readers (at the cost of incurring Power's wrath).
+
+Policy debates should not appear one-sided. Faced with this kind of dilemma, I can't say that defying Power is necessarily the right choice: if there really _were_ no other options between deceiving your readers with a bad faith performance, and incurring Power's wrath, and Power's wrath would be too terrible to bear, then maybe deceiving your readers with a bad faith performance is the right thing to do.
+
+But if you actually _cared_ about not deceiving your readers, you would want to be _really sure_ that those _really were_ the only two options. You'd [spend five minutes by the clock looking for third alternatives](https://www.lesswrong.com/posts/erGipespbbzdG5zYb/the-third-alternative)—including, possibly, not issuing proclamations on your honor as leader of the so-called "rationalist" community on topics where you _explicitly intend to ignore counteraguments on grounds of their being politically unfavorable_. Yudkowsky rejects this alternative on the grounds that it allegedly implies "utter silence about everything Stalin has expressed an opinion on including '2 + 2 = 4' because if that logically counterfactually were wrong you would not be able to express an opposing opinion", but this seems like yet another instance of Yudkowsky motivatedly playing dumb: if he _wanted_ to, I'm sure Eliezer Yudkowsky could think of _some relevant differences_ between "2 + 2 = 4" (a trivial fact of arithmetic) and "the simplest and best protocol is, "'He' refers to the set of people who have asked us to use 'he'" (a complex policy proposal whose numerous flaws I have analyzed in detail).
+
+"[P]eople do _know_ they're living in a half-Stalinist environment," Yudkowsky says. "I think people are better off at the end of that," he says. But who are "people", specifically? I'm asking because I have a long sob story about how _I_ didn't know, and _I'm_ not better off.
+
+------
+
+https://twitter.com/ESYudkowsky/status/1404697716689489921
+> I have never in my own life tried to persuade anyone to go trans (or not go trans)—I don't imagine myself to understand others that much.
+
+If you think it "sometimes personally prudent and not community-harmful" to strategically say positive things about Republican candidates, and make sure to never, ever say negative things about Democratic candidates (because you "don't see what the alternative is besides getting shot"), you can see why people might regard you as a _Republican shill_—even if all the things you said were true, and even if you never told any specific individual, "You should vote Republican."
+
+https://www.facebook.com/yudkowsky/posts/10154110278349228
+> Just checked my filtered messages on Facebook and saw, "Your post last night was kind of the final thing I needed to realize that I'm a girl."
+> ==DOES ALL OF THE HAPPY DANCE FOREVER==
+
+https://twitter.com/ESYudkowsky/status/1404821285276774403
+> It is not trans-specific. When people tell me I helped them, I mostly believe them and am happy.
+
+[The claim that the "Hill of Validity" thread wasn't partisan, but was just neutrally trying to teach the difference between facts and policy decisions is not credible. Imagine if someone was complaining about being required to say "Peace Be Upon Him" before referencing the prophet Muhammad. "It's a speech act, there's nothing factually false about saying 'peace be unto him'"]
+unremediatedgender.space/2020/Aug/yarvin-on-less-wrong/
+
+
+-----
+
+comments to "Wilhelm", January 2019—
+
+I got concessions on all the important parts (categories should make predictions, trans women differ from cis women in a masc direction), and these people just don't fucking CARE ... like, if I'm trying to be agreeable, I could agree that trans women resemble women if you restrict your vision to the subspace spanned by the "preferred pronouns" and "self-identified gender identity" dimensions ... but, but, WTF, be serious, guys
+
+Scott or Eliezer know better and could put an end to this bullshit (or at least put a dent in it), and I begged and I pleaded, and they just don't CARE
+
+even Ozy knows better
+
+I said: I probably do put too much rhetorical emphasis on passing; like, I agree that that's not the only criterion that one can use. I like drawing attention to that particular distinction because it at least has the benefit of not requiring people to override their perceptual system they way that self-identity does?
+
+and Ozy (correctly!) chimed in: "in fact it is the only criterion that doesn't involve people overriding their perceptual system!"
+
+as if she's objectively pro-gaslighting
+
+more charitably: people care a lot about this very thin layer of socual constructions (if you mindfuck ppl into believing that AGPs are women, that really does make it easier to transition) and are being instrumentally rational about that, whereas I'm an aspiring epistemic rationalist and want to study the deep structure if which social constructions are feasible, how we can be more reflective about them, &c.
+
+Ppl with less power than Scott or Eliezer can afford to be more honest with me that they see but don't care enough to pay the cost of fighting
+
+the rationalist lore here is that status makes you stupid; maybe the NRx twist is that status plus rivals/insecurity makes you stupid
+
+You _can't_ optimize your group's culture for not-talking-about-atheism without also optimizing against understanding Occam's razor; you _can't_ optimize for not questioning gender self-identity without also optimizing against understanding "A Human's Guide to Words."
+
+"yeah in public we say 'cis women' but tran to tran we just say 'women,' we‘re not insane lol"
+no transsexual is like 'from a young age all i ever wanted was to be.. cis'
+https://twitter.com/theorygurl/status/1062451652836446208
+
+Keltham and Carissa's attitude towards Pharima mirrors my attitude towards Yudkowsky (I'm grateful for him having created me, but he can't be allowed to get away with this shit)
+https://www.glowfic.com/replies/1834769#reply-1834769
+
+sometimes I get a cite, too—
+https://putanumonit.com/2022/05/02/genders-discrimination/
+https://axrp.net/episode/2022/05/23/episode-15-natural-abstractions-john-wentworth.html
+
+People learn a lot from Godel Escher Bach, too, but they don't form an identity around Douglas Hofstadter being the most important person in the world
+
+and Keltham tells Carissa (null action pg 39) to keep the Light alive as long as possible, not do throw away your deontology too quickly.
+
+> It, it—the fe—it, flame—flames. Flames—on the side of my face. Breathing—breathl—heaving breaths, heaving—
+
+like a crazy ex-girlfriend (I have no underlying issues to address; I'm certifiably cute, and adorably obsessed)
+
+But he is willing to go to bat for killing babies, but not for "Biological Sex is Actually Real Even If That Hurts Your Feelings" https://mobile.twitter.com/AuronMacintyre/status/1547562974927134732
+
+https://extropians.weidai.com/extropians.3Q97/4361.html
+> Half the time I regard myself as "an AI wannabee trapped in a male human body"
+
+Nate's "missing the hard part" post is all strawmen—I'm not looking down on it because it's a blog post and not an "official" arXiv paper; I'm looking down because it's visibly low-effort
+
+"What do you say to the Republican?" !!!
+
+subject: "nothing left to lose; or, the end of my rope"
+
+4 November 2018 email to Marcus—
+> Concrete anecdote about how my incredibly-filtered Berkeley social circle is nuts: at a small gathering this weekend I counted seven MtTs. (I think seven; I guess it's possible that physically-very-passable Cassandra is actually female, but given the context and her personality, I doubt it.) Plus me (a man wearing a dress and makeup), and three ordinary men, one ordinary woman, and my FtM friend. Looking up the MtTs' birthdays on Facebook was instructive in determining exactly how many years I was born too early. (Lots of 1992-3 births, so about five years.)
+
+Scathing rhetoric to Scott—
+> (I've been told that I'm not using the word "gaslighting" correctly. Somehow no one seems to think I have the right to define that category boundary however I want.)
+
+> If our vaunted rationality techniques result in me having to spend dozens of hours patiently explaining why I don't think that I'm a woman and that the person in this photograph isn't a woman, either (where "isn't a woman" is a convenient rhetorical shorthand for a much longer statement about naïve Bayes models and high-dimensional configuration spaces and defensible Schelling points for social norms), then our techniques are worse than useless.
+
+> But at the same time, if Galileo ever muttered "And yet it moves", there's a very similar long and important discussion to be had about the consequences of using the word "moves" in Galileo's preferred sense or some other sense that happens to result in the theory needing more epicycles. It may not have been obvious in 2014, but in retrospect, maybe it was a bad idea to build a memetic superweapon that says the number of epicycles doesn't matter.
+
+> It's gotten the point where Nature (!!!) has an editorial proclaiming ["US proposal for defining gender has no basis in science"](https://www.nature.com/articles/d41586-018-07238-8). And if otherkin started becoming sufficiently popular, I have no doubt that they could write an equally sober editorial denouncing species classification as having no basis in science. (Have you heard of ring species? Checkmate, cisspeciesists!)
+
+> I don't care about the social right to call Sam "he", because I wasn't going to do that anyway. What I do care about is the right to say "That's not what I meant by 'woman' in this context, and you fucking know it" when I think that my interloctuor does, in fact, fucking know it.
+
+Anna thinks trust and integrity is an important resource
+https://www.lesswrong.com/posts/mmHctwkKjpvaQdC3c/what-should-you-change-in-response-to-an-emergency-and-ai
+
+The HEXACO personality model considers "honesty" and "humility" a single factor
+
+I'm not usually—at least, not always—so much of a scrub as to play chess with a pigeon (which shits on the board and then struts around like it's won), or wrestle with a pig (which gets you both dirty, and the pig likes it), or dispute what the Tortise said to Achilles
+
+(You might group things together _on the grounds_ of their similarly positive consequences—that's what words like _good_ do—but that's distinct from choosing _the categorization itself_ because of its consequences.)
+
+—and would be unforgivable if it weren't so _inexplicable_.
+
+... not _actually_ inexplicable. There was, in fact, an obvious explanation: that Yudkowsky was trying to bolster his reputation amongst progressives by positioning himself on the right side of history, and was tailoring a fake rationality lesson to suit that goal. But _Eliezer Yudkowsky wouldn't do that_. I had to assume this was a honest mistake.
+
+At least, a _pedagogy_ mistake. If Yudkowsky _just_ wanted to make a politically neutral technical point about the difference between fact-claims and policy claims _without_ "picking a side" in the broader cultural war dispute, these Tweets did a very poor job of it. I of course agree that pronoun usage conventions, and conventions about who uses what bathroom, are not, themselves, factual assertions about sex chromosomes in particular. I'm not saying that Yudkowsky made a false statement there. Rather, I'm saying that it's
+
+
+Rather, previously sexspace had two main clusters (normal females and males) plus an assortment of tiny clusters corresponding to various [disorders of sex development](https://en.wikipedia.org/wiki/Disorders_of_sex_development), and now it has two additional tiny clusters: females-on-masculinizing-HRT and males-on-feminizing-HRT. Certainly, there are situations where you would want to use "gender" categories that use the grouping {females, males-on-feminizing-HRT} and {males, females-on-masculinizing-HRT}.
+
+[TODO: relevance of multivariate—
+
+(And in this case, the empirical facts are _so_ lopsided, that if we must find humor in the matter, it really goes the other way. Lia Thomas trounces the entire field by _4.2 standard deviations_ (!!), and Eliezer Yudkowsky feels obligated to _pretend not to see the problem?_ You've got to admit, that's a _little_ bit funny.)
+
+https://www.lesswrong.com/posts/cu7YY7WdgJBs3DpmJ/the-univariate-fallacy
+https://www.lesswrong.com/posts/vhp2sW6iBhNJwqcwP/blood-is-thicker-than-water
+
+]
+
+[TODO: sentences about studies showing that HRT doesn't erase male advantage
+https://twitter.com/FondOfBeetles/status/1368176581965930501
+]
+
+[TODO sentences about Lia Thomas and Cece Tefler] https://twitter.com/FondOfBeetles/status/1466044767561830405 (Thomas and Tefler's —cite South Park)
+https://www.dailymail.co.uk/news/article-10445679/Lia-Thomas-UPenn-teammate-says-trans-swimmer-doesnt-cover-genitals-locker-room.html
+https://twitter.com/sharrond62/status/1495802345380356103 Lia Thomas event coverage
+https://www.realityslaststand.com/p/weekly-recap-lia-thomas-birth-certificates Zippy inv. cluster graph!
+
+]
+
+Writing out this criticism now, the situation doesn't feel _confusing_, anymore. Yudkowsky was very obviously being intellectually dishonest in response to very obvious political incentives. That's a thing that public intellectuals do. And, again, I agree that the distinction between facts and policy decisions _is_ a valid one, even if I thought it was being selectively invoked here as an [isolated demand for rigor](http://slatestarcodex.com/2014/08/14/beware-isolated-demands-for-rigor/) because of the political context. Coming from _anyone else in the world_, I would have considered the thread fine—a solidly above-average performance, really. I wouldn't have felt confused or betrayed at all. Coming from Eliezer Yudkowsky, it was—confusing.
+
+Because of my hero worship, "he's being intellectually dishonest in response to very obvious political incentives" wasn't in my hypothesis space; I _had_ to assume the thread was an "honest mistake" in his rationality lessons, rather than (what it actually was, what it _obviously_ actually was) hostile political action.
+
+(I _want_ to confidently predict that everything I've just said is completely obvious to you, because I learned it all specifically from you! A 130 IQ _nobody_ like me shouldn't have to say _any_ of this to the _author_ of "A Human's Guide to Words"! But then I don't know how to reconcile that with your recent public statement about [not seeing "how there's scientific truth at stake"](https://twitter.com/ESYudkowsky/status/1067482047126495232). Hence this desperate and [_confused_](https://www.lesswrong.com/posts/5JDkW4MYXit2CquLs/your-strength-as-a-rationalist) email plea.)
+
+And I'm still really confused, because I still feel like everything I'm saying here is a trivial application of Sequences-lore. If I'm getting something wrong, I should be overjoyed to be harshly corrected by the Great Teacher! A simple person like me is as but a mere worm in the presence of the great Eliezer Yudkowsky! But if it looks like the Great Teacher is getting something wrong (wrong with respect to the balanced flow of arguments and evidence in which every "step is precise and has no room in it for your whims", although not wrong in the sense of making a factually incorrect statement) and the Great Teacher neither corrects me nor says "OK, you're right and I was wrong, well done, my student", what am I supposed to conclude? Is this a prank—a test? Am I like Brennan in "Initiation Ceremony", being evaluated to see if I have the guts to stand by my vision of the Way in the face of social pressure? (If so, I'm not doing a very good job, because I definitely wouldn't be writing this if I hadn't gotten social proof from Michael, Ben, and Sarah.) Did I pass??
+
+In a functioning rationalist community, there should never be any occasion in which "appeal to Eliezer Yudkowsky's personal authority" seems like a good strategy: the way this is supposed to work is that I should just make my arguments with the understanding that good arguments will be accepted and bad arguments will be rejected. But I've been trying that, and it's mostly not working. On any other topic, I probably would have just given up and accepted the social consensus by now: "Sure, OK, whatever, trans women are women by definition; who am I to think I've seen into the Bayes-structure?" I still think this from time to time, and feel really guilty about arguing for the Bad Guys (because in my native Blue Tribe culture, only Bad people want to talk about sexual dimorphism). But then I can't stop seeing the Bayes-structure that says that biological sex continues to be a predictively-useful concept even when it's ideologically unfashionable—and I've got Something to Protect. What am I supposed to do?
+
+I agree that this is the only reason you should care.
+
+> People probably change their mind more often than they explicitly concede arguments, which is fine because intellectual progress is more important than people who were wrong performing submission.
+> If your interlocutor is making progress arguing against your claim X, just say, "Oh, X is a strawman, no one actually believes X; therefore I'm not wrong and you haven't won" (and then don't argue for X in the future now that you know you can't get away with it).
+https://twitter.com/zackmdavis/status/1088459797962215429
+
+
+
+My 28 November 2018 text to Michael—
+> I just sent an email to Eliezer ccing you and Anna; if you think it might help inject sanity in into the world, maybe your endorsement would help insofar as Eliezer Aumman-updates [_sic_] with you?
+> hope all is well
+> just a thread reply to Eliezer that says "I trust Zack's rationality and think you should pay attention to what he has to say" (if and only if you actually believe that to be true, obviously)?
+
+
+(don't know how to summarize the part with Ian—)
+I remember going downstairs to impulsively confide in a senior engineer, an older bald guy who exuded masculinity, who you could tell by his entire manner and being was not infected by the Berkeley mind-virus, no matter how loyally he voted Democrat—not just about the immediate impetus of this Twitter thread, but this whole _thing_ of the past couple years where my entire social circle just suddenly decided that guys like me could be women by means of saying so. He was sympathetic.
+
+
+[TODO: paraphrase remaining interaction with Scott, or not worth the space?
+
+> I don't have a simple, mistake-theoretic characterization of the language and social conventions that everyone should use such that anyone who defected from the compromise would be wrong. The best I can do is try to objectively predict the consequences of different possible conventions—and of conflicts over possible conventions.
+
+helping Norton live in the real world
+
+Scott says, "It seems to me pretty obvious that the mental health benefits to trans people are enough to tip the object-level first-pass uilitarian calculus."; I don't think _anything_ about "mental health benefits to trans people" is obvious
+]
+
+[TODO: connecting with Aurora 8 December, maybe not important]
+
+What do think submitting to social pressure looks like, if it's not exactly this thing (carefully choosing your public statements to make sure no one confuses you with the Designated Ideological Bad Guy)?!? The credible threat of being labeled an Ideological Bad Guy is _the mechanism_ the "Good" Guys use to retard potentially-ideologically-inconvenient areas of inquiry.
+
+Kerry Vaughan on deferral
+https://twitter.com/KerryLVaughan/status/1552308109535858689
\ No newline at end of file