-[At this point, the nature of the game is very clear. Yudkowsky wants to mood-affiliate with being on the right side of history, subject to the constraint of not saying anything false. I want to actually make sense of what's actually going on in the world, because _I need the correct answer to decided whether or not to cut my dick off_. On "his turn", he comes up with some pompous proclamation that's optimized to make the "pro-trans" faction look smart and good and the "anti-trans" faction look dumb and bad, "in ways that exhibit generally rationalist principles." On my turn, I put in an absurd amount of effort explaining in exhaustive, _exhaustive_ detail why Yudkowsky's pompous proclamation was substantively misleading as constrated to what you would say if you were actually trying to make sense of the world.]
+At this point, the nature of the game is very clear. Yudkowsky wants to mood-affiliate with being on the right side of history (as ascertained by the current year's progressive _Zeitgeist_), subject to the constraint of not saying anything he knows to be false. Meanwhile, I want to actually make sense of what's actually going on in the world as regards sex and gender, because _I need the correct answer to decide whether or not to cut my dick off_.
+
+On "his turn", he comes up with some pompous proclamation that's very obviously optimized to make the "pro-trans" faction look smart and good and make the "anti-trans" faction look dumb and bad, "in ways that exhibit generally rationalist principles."
+
+On "my turn", I put in an _absurd_ amount of effort explaining in exhaustive, _exhaustive_ detail why Yudkowsky's pompous proclamation, while [not technically saying anything definitively "false"](https://www.lesswrong.com/posts/MN4NRkMw7ggt9587K/firming-up-not-lying-around-its-edge-cases-is-less-broadly), was _substantively misleading_ as constrated to what any serious person would say if they were actually trying to make sense of the world without worrying what progressive activists would think of them.
+
+In the context of AI alignment theory, Yudkowsky has written about a "nearest unblocked strategy" phenomenon: if you directly prevent an agent from accomplishing a goal via some plan that you find undesirable, the agent will search for ways to route around that restriction, and probably find some plan that you find similarly undesirable for similar reasons.