+[SECTION: treachery, faith, and the great river
+
+I concluded that further email prosecution was not useful at this time. My revised Category War to-do list was:
+
+ * Send a _brief_ wrapping-up/end-of-conversation email to Scott (with the anecdote from Discord and commentary on his orc story).
+ * Mentally write-off Scott, Eliezer, and the so-called "rationalist" community as a loss so that I wouldn't be in horrible emotional pain from cognitive dissonance all the time.
+ * Write up the long, engaging, depoliticized mathy version of the categories argument for _Less Wrong_ (which I thought might take a few months—I had a dayjob, and write slowly, and might need to learn some new math, which I'm also slow at).
+ * _Then_ email the link to Scott and Eliezer asking for a signal-boost and/or court ruling.
+
+Ben didn't think the mathematically precise categories argument was the most important thing for _Less Wrong_ readers to know about: a similarly careful explanation of why I've written off Scott, Eliezer, and the "rationalists" would be way more valuable.
+
+I could see the value he was pointing at, but something in me balked at the idea of _attacking my friends in public_ (Subject: "treachery, faith, and the great river (was: Re: DRAFTS: 'wrapping up; or, Orc-ham's razor' and 'on the power and efficacy of categories')").
+
+Ben had previously written (in the context of the effective altruism movement) about how [holding criticism to a higher standard than praise distorts our collective map](http://benjaminrosshoffman.com/honesty-and-perjury/#A_tax_on_criticism).
+
+He was obviously correct that this was a distortionary force relative to what ideal Bayesian agents would do, but I was worried that when we're talking about criticism of _people_ rather than ideas, the removal of the distortionary force would just result in an ugly war (and not more truth). Criticism of institutions and social systems _should_ be filed under "ideas" rather than "people", but the smaller-scale you get, the harder this distinction is to maintain: criticizing, say, "the Center for Effective Altruism", somehow feels more like criticizing Will MacAskill personally than criticizing "the United States" does, even though neither CEA nor the U.S. is a person.
+
+This is was I felt like I couldn't give up faith that [honest discourse _eventually_ wins](https://slatestarcodex.com/2017/03/24/guided-by-the-beauty-of-our-weapons/). Under my current strategy and consensus social norms, I could criticize Scott or Kelsey or Ozy's _ideas_ without my social life dissolving into a war of all against all, whereas if I were to give in to the temptation to flip a table and say, "Okay, now I _know_ you guys are just fucking with me," then I didn't see how that led anywhere good, even if they really _are_ just fucking with me.
+
+Jessica explained what she saw as the problem with this. What Ben was proposing was _creating clarity about behavioral patterns_. I was saying that I was afraid that creating such clarity is an attack on someone. But if so, then my blog was an attack on trans people. What was going on here?
+
+Socially, creating clarity about behavioral patterns _is_ construed as an attack and _can_ make things worse for someone: for example, if your livelihood is based on telling a story about you and your flunkies being the only sane truthseeking people in the world, then me demonstrating that you don't care about the truth when it's politically inconvenient for you is a threat to your marketing story and therefore a threat to your livelihood. As a result, it's easier to create clarity down power gradients than up power gradients: it was easy for me to blow the whistle on trans people's narcissistic delusions, but hard to blow the whistle on Eliezer Yudkowsky's narcissistic delusions.
+
+But _selectively_ creating clarity down but not up power gradients just reinforces existing power relations—just like how selectively criticizing arguments with politically unfavorable conclusions only reinforces your current political beliefs. I shouldn't be able to get away with claiming that [calling non-exclusively-androphilic trans women delusional perverts](/2017/Mar/smart/) is okay on the grounds that that which can be destroyed by the truth should be, but that calling out Alexander and Yudkowsky would be unjustified on the grounds of starting a war or whatever. If I was being cowardly or otherwise unprincipled, I should own that instead of generating spurious justifications. Jessica was on board with a project to tear down narcissistic fantasies in general, but not on board with a project that starts by tearing down trans people's narcissistic fantasies, but then emits spurious excuses for not following that effort where it leads.
+
+Somewhat apologetically, I replied that the distinction between truthfully, publicly criticizing group identities and _named individuals_ still seemed very significant to me? I would be way more comfortable writing [a scathing blog post about the behavior of "rationalists"](/2017/Jan/im-sick-of-being-lied-to/), than about a specific person not adhering to good discourse norms in an email conversation that they had good reason to expect to be private. I thought I was consistent about this: contrast my writing to the way that some anti-trans writers name-and-shame particular individuals. (The closest I had come was [mentioning Danielle Muscato as someone who doesn't pass](/2018/Dec/untitled-metablogging-26-december-2018/#photo-of-danielle-muscato)—and even there, I admitted it was "unclassy" and done in desperation of other ways to make the point having failed.) I had to acknowledge that criticism of non-exclusively-androphilic trans women in general _implied_ criticism of Jessica, and criticism of "rationalists" in general _implied_ criticism of Yudkowsky and Alexander and me, but the extra inferential step and "fog of probability" seemed useful for making the speech act less of an attack? Was I wrong?
+
+Michael said this was importantly backwards: less precise targeting is more violent. If someone said, "Michael Vassar is a terrible person", he would try to be curious, but if they don't have an argument, he would tend to worry more "for" them and less "about" them, whereas if someone said, "The Jews are terrible people", he saw that more serious threat to his safety. (And rationalists and trans women are exact sort of people that get targeted by the same people to target Jews.)
+
+]
+
+
+[SECTION about monastaries (with Ben and Anna in April 2019)
+
+I complained to Anna: "Getting the right answer in public on topic _X_ would be too expensive, so we won't do it" is _less damaging_ when the set of such <em>X</em>es is _small_. It looked to me like we added a new forbidden topic in the last ten years, without rolling back any of the old ones.
+
+"Reasoning in public is too expensive; reasoning in private is good enough" is _less damaging_ when there's some sort of _recruiting pipeline_ from the public into the monasteries: lure young smart people in with entertaining writing and shiny math, _then_ gradually undo their political brainwashing once they've already joined your cult. (It had [worked on me](/2021/May/sexual-dimorphism-in-the-sequences-in-relation-to-my-gender-problems/)!)
+
+I would be sympathetic to "rationalist" leaders like Anna or Yudkowsky playing that strategy if there were some sort of indication that they had _thought_, at all, about the pipeline problem—or even an indication that there _was_ an intact monastery somewhere.
+
+]
+
+[TODO: Jessica on corruption—
+
+> I am reminded of someone who I talked with about Zack writing to you and Scott to request that you clarify the category boundary thing. This person had an emotional reaction described as a sense that "Zack should have known that wouldn't work" (because of the politics involved, not because Zack wasn't right). Those who are savvy in high-corruption equilibria maintain the delusion that high corruption is common knowledge, to justify expropriating those who naively don't play along, by narratizing them as already knowing and therefore intentionally attacking people, rather than being lied to and confused.
+]
+
+
+[TODO small section: concern about bad faith nitpicking—
+
+One reason someone might be reluctant to correct mistakes when pointed out, is the fear that such a policy could be abused by motivated nitpickers. It would be pretty annoying to be obligated to churn out an endless stream of trivial corrections by someone motivated to comb through your entire portfolio and point out every little thing you did imperfectly, ever.
+
+I wondered if maybe, in Scott or Eliezer's mental universe, I was a blameworthy (or pitiably mentally ill) nitpicker for flipping out over a blog post from 2014 (!) and some Tweets (!!) from November. Like, really? I, too, had probably said things that were wrong _five years ago_.
+
+But, well, I thought I had made a pretty convincing that a lot of people are making a correctable and important rationality mistake, such that the cost of a correction (about the philosophy of language specifically, not any possible implications for gender politics) would actually be justified here. If someone had put _this much_ effort into pointing out an error I had made four months or five years ago and making careful arguments for why it was important to get the right answer, I think I _would_ put some serious thought into it rather than brushing them off.
+
+]
+
+
+[TODO: after some bouncing off the posse, what was originally an email draft became a public _Less Wrong_ post, "Where to Draw the Boundaries?" (note, plural)
+ * Wasn't the math overkill?
+ * math is important for appeal to principle—and as intimidation https://slatestarcodex.com/2014/08/10/getting-eulered/
+ * four simulacra levels got kicked off here
+ * I could see that I'm including subtext and expecting people to only engage with the text, but if we're not going to get into full-on gender-politics on Less Wrong, but gender politics is motivating an epistemology error, I'm not sure what else I'm supposed to do! I'm pretty constrained here!
+ * I had already poisoned the well with "Blegg Mode" the other month, bad decision
+ * We lost?! How could we lose??!!?!?
+]
+
+
+[TODO: my reluctance to write a memoir, displacement behavior
+Ben thought it was imporant on 30 Apr, 12 Aug , I confess to being stuck on 9 Nov
+I write more about the philosophy of language instead
+"Univariate fallacy" also a concession
+]
+
+curation hopes ... 22 Jun: I'm expressing a little bit of bitterness that a mole rats post got curated https://www.lesswrong.com/posts/fDKZZtTMTcGqvHnXd/naked-mole-rats-a-case-study-in-biological-weirdness
+
+[TODO scuffle on "Yes Requires the Possibility of No"]
+
+[TODO scuffle on LessWrong FAQ https://www.lesswrong.com/posts/MqrzczdGhQCRePgqN/feedback-requested-draft-of-a-new-about-welcome-page-for#iqEEme6M2JmZEXYAk ]
+
+[TODO section on factional conflict:
+Michael on Anna as an enemy
+24 Aug: I had told Anna about Michael's "enemy combatants" metaphor, and how I originally misunderstood
+me being regarded as Michael's pawn
+assortment of agendas
+mutualist pattern where Michael by himself isn't very useful for scholarship (he just says a lot of crazy-sounding things and refuses to explain them), but people like Sarah and me can write intelligible things that secretly benefited from much less legible conversations with Michael.
+]
+
+[TODO: Yudkowsky throwing NRx under the bus; tragedy of recursive silencing
+15 Sep Glen Weyl apology
+]
+
+[TODO: Ziz incident; more upset about gender validation than the felony charges, which were equally ridiculous and more obviously linked to physical violence
+complicity with injustice "Ziz isn't going to be a problem for you anymore"]
+
+[TODO: a culture that has gone off the rails; my warning points to Vaniver]
+
+[TODO: plan to reach out to Rick]
+
+[TODO: Christmas party
+people reading funny GPT-2 quotes
+Tsvi said it would be sad if I had to leave the Bay Area
+motivation deflates after Christmas victory
+5 Jan memoir as nuke
+]
+
+(There's another very important part of the story that would fit around here chronologically, but unfortunately, it's not my story to tell.)
+
+[TODO: "out of patience" email]
+[TODO: Sep 2020 categories clarification from EY—victory?!]
+
+[TODO: That should have been the end of the story, but then—he revisited the pronouns issue!!!]
+
+[TODO: based on the timing, the Feb. 2021 pronouns post was likely causally downstream of me being temporarily more salient to EY because of my highly-Liked response to his "anyone at this point that anybody who openly hates on this community generally or me personally is probably also a bad person inside" from 17 February; it wasn't gratuitously out of the blue]
+
+If Yudkowsky is obviously playing dumb (consciously or not) and his comments can't be taken seriously, what's _actually_ going on here? When smart people act dumb, [it's usually wisest to assume that their behavior represents _optimized_ stupidity](https://www.lesswrong.com/posts/sXHQ9R5tahiaXEZhR/algorithmic-intent-a-hansonian-generalized-anti-zombie)—apparent "stupidity" that achieves a goal through some other channel than their words straightforwardly reflecting the truth. Someone who was _actually_ stupid wouldn't be able to generate text with a specific balance of insight and selective stupidity fine-tuned to reach a gender-politically convenient conclusion without explicitly invoking any controversial gender-political reasoning.
+
+Fortunately, Yudkowsky graciously grants us a clue in the form of [a disclaimer comment](https://www.facebook.com/yudkowsky/posts/10159421750419228?comment_id=10159421833274228):
+
+> It unfortunately occurs to me that I must, in cases like these, disclaim that—to the extent there existed sensible opposing arguments against what I have just said—people might be reluctant to speak them in public, in the present social atmosphere. That is, in the logical counterfactual universe where I knew of very strong arguments against freedom of pronouns, I would have probably stayed silent on the issue, as would many other high-profile community members, and only Zack M. Davis would have said anything where you could hear it.
+>
+> This is a filter affecting your evidence; it has not to my own knowledge filtered out a giant valid counterargument that invalidates this whole post. I would have kept silent in that case, for to speak then would have been dishonest.
+>
+> Personally, I'm used to operating without the cognitive support of a civilization in controversial domains, and have some confidence in my own ability to independently invent everything important that would be on the other side of the filter and check it myself before speaking. So you know, from having read this, that I checked all the speakable and unspeakable arguments I had thought of, and concluded that this speakable argument would be good on net to publish, as would not be the case if I knew of a stronger but unspeakable counterargument in favor of Gendered Pronouns For Everyone and Asking To Leave The System Is Lying.
+>
+> But the existence of a wide social filter like that should be kept in mind; to whatever quantitative extent you don't trust your ability plus my ability to think of valid counterarguments that might exist, as a Bayesian you should proportionally update in the direction of the unknown arguments you speculate might have been filtered out.
+
+So, the explanation of [the problem of political censorship filtering evidence](https://www.lesswrong.com/posts/DoPo4PDjgSySquHX8/heads-i-win-tails-never-heard-of-her-or-selective-reporting) here is great, but the part where Yudkowsky claims "confidence in [his] own ability to independently invent everything important that would be on the other side of the filter" is just _laughable_. My point that _she_ and _he_ have existing meanings that you can't just ignore by fiat given that the existing meanings are _exactly_ what motivate people to ask for new pronouns in the first place is _really obvious_.
+
+Really, it would be _less_ embarassing for Yudkowsky if he were outright lying about having tried to think of counterarguments. The original post isn't _that_ bad if you assume that Yudkowsky was writing off the cuff, that he clearly just _didn't put any effort whatsoever_ into thinking about why someone might disagree. If he _did_ put in the effort—enough that he felt comfortable bragging about his ability to see the other side of the argument—and _still_ ended up proclaiming his "simplest and best protocol" without even so much as _mentioning_ any of its incredibly obvious costs ... that's just _pathetic_. If Yudkowsky's ability to explore the space of arguments is _that_ bad, why would you trust his opinion about _anything_?
+
+[TODO: as Yudkowsky put it
+> Admitting something when being pushed a little, but never thinking it spontaneously and hence having those thoughts absent from your own thought processes, remains not sane.
+https://twitter.com/ESYudkowsky/status/1501218503990431745
+]
+
+But perhaps it's premature to judge Yudkowsky without appreciating what tight constraints he labors under. The disclaimer comment mentions "speakable and unspeakable arguments"—but what, exactly, is the boundary of the "speakable"? In response to a commenter mentioning the cost of having to remember pronouns as a potential counterargument, Yudkowsky [offers us another clue](https://www.facebook.com/yudkowsky/posts/10159421750419228?comment_id=10159421833274228&reply_comment_id=10159421871809228):
+
+> People might be able to speak that. A clearer example of a forbidden counterargument would be something like e.g. imagine if there was a pair of experimental studies somehow proving that (a) everybody claiming to experience gender dysphoria was lying, and that (b) they then got more favorable treatment from the rest of society. We wouldn't be able to talk about that. No such study exists to the best of my own knowledge, and in this case we might well hear about it from the other side to whom this is the exact opposite of unspeakable; but that would be an example.
+
+(As an aside, the wording of "we might well hear about it from _the other side_" (emphasis mine) is _very_ interesting, suggesting that the so-called "rationalist" community, is, effectively, a partisan institution, despite its claims to be about advancing the generically human art of systematically correct reasoning.)
+
+I think (a) and (b) _as stated_ are clearly false, so "we" (who?) fortunately aren't losing much by allegedly not being able to speak them. But what about some _similar_ hypotheses, that might be similarly unspeakable for similar reasons?
+
+Instead of (a), consider the claim that (a′) self-reports about gender dysphoria are substantially distorted by [socially-desirable responding tendencies](https://en.wikipedia.org/wiki/Social-desirability_bias)—as a notable and common example, heterosexual males with [sexual fantasies about being female](http://www.annelawrence.com/autogynephilia_&_MtF_typology.html) [often falsely deny or minimize the erotic dimension of their desire to change sex](/papers/blanchard-clemmensen-steiner-social_desirability_response_set_and_systematic_distortion.pdf) (The idea that self-reports can be motivatedly inaccurate without the subject consciously "lying" should not be novel to someone who co-blogged with [Robin Hanson](https://en.wikipedia.org/wiki/The_Elephant_in_the_Brain) for years!)
+
+And instead of (b), consider the claim that (b′) transitioning is socially rewarded within particular _subcultures_ (although not Society as a whole), such that many of the same people wouldn't think of themselves as trans or even gender-dysphoric if they lived in a different subculture.
+
+I claim that (a′) and (b′) are _overwhelmingly likely to be true_. Can "we" talk about _that_? Are (a′) and (b′) "speakable", or not?
+
+We're unlikely to get clarification from Yudkowsky, but based on my experiences with the so-called "rationalist" community over the past coming-up-on-six years—I'm going to _guess_ that the answer is broadly No: no, "we" can't talk about that.
+
+But if I'm right that (a′) and (b′) should be live hypotheses and that Yudkowsky would consider them "unspeakable", that means "we" can't talk about what's _actually going on_ with gender dysphoria and transsexuality, which puts the whole discussion in a different light. In another comment, Yudkowsky lists some gender-transition interventions he named in [a November 2018 Twitter thread](https://twitter.com/ESYudkowsky/status/1067183500216811521) that was the precursor to the present discussion—using a different bathroom, changing one's name, asking for new pronouns, and getting sex reassignment surgery—and notes that none of these are calling oneself a "woman". [He continues](https://www.facebook.com/yudkowsky/posts/10159421750419228?comment_id=10159421986539228&reply_comment_id=10159424960909228):
+
+> [Calling someone a "woman"] _is_ closer to the right sort of thing _ontologically_ to be true or false. More relevant to the current thread, now that we have a truth-bearing sentence, we can admit of the possibility of using our human superpower of language to _debate_ whether this sentence is indeed true or false, and have people express their nuanced opinions by uttering this sentence, or perhaps a more complicated sentence using a bunch of caveats, or maybe using the original sentence uncaveated to express their belief that this is a bad place for caveats. Policies about who uses what bathroom also have consequences and we can debate the goodness or badness (not truth or falsity) of those policies, and utter sentences to declare our nuanced or non-nuanced position before or after that debate.
+>
+> Trying to pack all of that into the pronouns you'd have to use in step 1 is the wrong place to pack it.
+
+Sure, _if we were in the position of designing a constructed language from scratch_ under current social conditions in which a person's "gender" is a contested social construct, rather than their sex an objective and undisputed fact, then yeah: in that situation _which we are not in_, you definitely wouldn't want to pack sex or gender into pronouns. But it's a disingenuous derailing tactic to grandstand about how people need to alter the semantics of their _already existing_ native language so that we can discuss the real issues under an allegedly superior pronoun convention when, _by your own admission_, you have _no intention whatsoever of discussing the real issues!_
+
+(Lest the "by your own admission" clause seem too accusatory, I should note that given constant behavior, admitting it is _much_ better than not-admitting it; so, huge thanks to Yudkowsky for the transparency on this point!)
+
+Again, as discussed in "Challenges to Yudkowsky's Pronoun Reform Proposal", a comparison to [the _tú_/_usted_ distinction](https://en.wikipedia.org/wiki/Spanish_personal_pronouns#T%C3%BA/vos_and_usted) is instructive. It's one thing to advocate for collapsing the distinction and just settling on one second-person singular pronoun for the Spanish language. That's principled.
+
+It's quite another thing altogether to _simultaneously_ try to prevent a speaker from using _tú_ to indicate disrespect towards a social superior (on the stated rationale that the _tú_/_usted_ distinction is dumb and shouldn't exist), while _also_ refusing to entertain or address the speaker's arguments explaining _why_ they think their interlocutor is unworthy of the deference that would be implied by _usted_ (because such arguments are "unspeakable" for political reasons). That's just psychologically abusive.
+
+If Yudkowsky _actually_ possessed (and felt motivated to use) the "ability to independently invent everything important that would be on the other side of the filter and check it [himself] before speaking", it would be _obvious_ to him that "Gendered Pronouns For Everyone and Asking To Leave The System Is Lying" isn't the hill anyone would care about dying on if it weren't a Schelling point. A lot of TERF-adjacent folk would be _overjoyed_ to concede the (boring, insubstantial) matter of pronouns as a trivial courtesy if it meant getting to _actually_ address their real concerns of "Biological Sex Actually Exists", and ["Biological Sex Cannot Be Changed With Existing or Foreseeable Technology"](https://www.lesswrong.com/posts/QZs4vkC7cbyjL9XA9/changing-emotions) and "Biological Sex Is Sometimes More Relevant Than Self-Declared Gender Identity." The reason so many of them are inclined to stand their ground and not even offer the trivial courtesy is because they suspect, correctly, that the matter of pronouns is being used as a rhetorical wedge to try to prevent people from talking or thinking about sex.
+
+Having analyzed the _ways_ in which Yudkowsky is playing dumb here, what's still not entirely clear is _why_. Presumably he cares about maintaining his credibility as an insightful and fair-minded thinker. Why tarnish that by putting on this haughty performance?
+
+Of course, presumably he _doesn't_ think he's tarnishing it—but why not? [He graciously explains in the Facebook comments](https://www.facebook.com/yudkowsky/posts/10159421750419228?comment_id=10159421833274228&reply_comment_id=10159421901809228):
+
+> it is sometimes personally prudent and not community-harmful to post your agreement with Stalin about things you actually agree with Stalin about, in ways that exhibit generally rationalist principles, especially because people do _know_ they're living in a half-Stalinist environment [...] I think people are better off at the end of that.
+
+Ah, _prudence_! He continues:
+
+> I don't see what the alternative is besides getting shot, or utter silence about everything Stalin has expressed an opinion on including "2 + 2 = 4" because if that logically counterfactually were wrong you would not be able to express an opposing opinion.
+
+The problem with trying to "exhibit generally rationalist principles" in an line of argument that you're constructing in order to be prudent and not community-harmful, is that you're thereby necessarily _not_ exhibiting the central rationalist principle that what matters is the process that _determines_ your conclusion, not the reasoning you present to _reach_ your presented conclusion, after the fact.
+
+The best explanation of this I know was authored by Yudkowsky himself in 2007, in a post titled ["A Rational Argument"](https://www.lesswrong.com/posts/9f5EXt8KNNxTAihtZ/a-rational-argument). It's worth quoting at length. The Yudkowsky of 2007 invites us to consider the plight of a political campaign manager:
+
+> As a campaign manager reading a book on rationality, one question lies foremost on your mind: "How can I construct an impeccable rational argument that Mortimer Q. Snodgrass is the best candidate for Mayor of Hadleyburg?"
+>
+> Sorry. It can't be done.
+>
+> "What?" you cry. "But what if I use only valid support to construct my structure of reason? What if every fact I cite is true to the best of my knowledge, and relevant evidence under Bayes's Rule?"
+>
+> Sorry. It still can't be done. You defeated yourself the instant you specified your argument's conclusion in advance.
+
+The campaign manager is in possession of a survey of mayoral candidates on which Snodgrass compares favorably to other candidates, except for one question. The post continues (bolding mine):
+
+> So you are tempted to publish the questionnaire as part of your own campaign literature ... with the 11th question omitted, of course.
+>
+> **Which crosses the line between _rationality_ and _rationalization_.** It is no longer possible for the voters to condition on the facts alone; they must condition on the additional fact of their presentation, and infer the existence of hidden evidence.
+>
+> Indeed, **you crossed the line at the point where you considered whether the questionnaire was favorable or unfavorable to your candidate, before deciding whether to publish it.** "What!" you cry. "A campaign should publish facts unfavorable to their candidate?" But put yourself in the shoes of a voter, still trying to select a candidate—why would you censor useful information? You wouldn't, if you were genuinely curious. If you were flowing _forward_ from the evidence to an unknown choice of candidate, rather than flowing _backward_ from a fixed candidate to determine the arguments.
+
+The post then briefly discusses the idea of a "logical" argument, one whose conclusions follow from its premises. "All rectangles are quadrilaterals; all squares are quadrilaterals; therefore, all squares are rectangles" is given as an example of _illogical_ argument, even though the both premises are true (all rectangles and squares are in fact quadrilaterals) _and_ the conclusion is true (all squares are in fact rectangles). The problem is that the conclusion doesn't _follow_ from the premises; the _reason_ all squares are rectangles isn't _because_ they're both quadrilaterals. If we accepted arguments of the general _form_ "all A are C; all B are C; therefore all A are B", we would end up believing nonsense.
+
+Yudkowsky's conception of a "rational" argument—at least, Yudkowsky's conception in 2007, which the Yudkowsky of the current year seems to disagree with—has a similar flavor: the stated reasons should be the actual reasons. The post concludes:
+
+> If you really want to present an honest, rational argument _for your candidate_, in a political campaign, there is only one way to do it:
+>
+> * _Before anyone hires you_, gather up all the evidence you can about the different candidates.
+> * Make a checklist which you, yourself, will use to decide which candidate seems best.
+> * Process the checklist.
+> * Go to the winning candidate.
+> * Offer to become their campaign manager.
+> * When they ask for campaign literature, print out your checklist.
+>
+> Only in this way can you offer a _rational_ chain of argument, one whose bottom line was written flowing _forward_ from the lines above it. Whatever _actually_ decides your bottom line is the only thing you can _honestly_ write on the lines above.
+
+I remember this being pretty shocking to read back in 'aught-seven. What an alien mindset! But it's _correct_. You can't rationally argue "for" a chosen conclusion, because only the process you use to _decide what to argue for_ can be your real reason.