Title: Sexual Dimorphism in Yudkowsky's Sequences, in Relation to My Gender Problems Date: 2021-01-01 Category: commentary Tags: autogynephilia, Eliezer Yudkowsky, epistemic horror, my robot cult, personal, sex differences Status: draft So, as I sometimes allude to, I've spent basically my entire adult life in this insular intellectual subculture that was founded in the late 'aughts to promulgate an ideal of _systematically correct reasoning_—general methods of thought that result in true beliefs and successful plans—and, incidentally, to use these methods of systematically correct reasoning to prevent superintelligent machines from [destroying all value in the universe](https://www.lesswrong.com/posts/GNnHHmm8EzePmKzPk/value-is-fragile). Lately I've been calling it my "robot cult" (a phrase [due to Dale Carrico](https://amormundi.blogspot.com/2011/08/ten-reasons-to-take-seriously.html))—the pejorative is partially [ironically affectionate](https://www.lesswrong.com/posts/gBma88LH3CLQsqyfS/cultish-countercultishness), and partially an expression of resentment and white-hot rage acquired from that time almost everyone I [used to trust](https://www.lesswrong.com/posts/wustx45CPL5rZenuo/no-safe-defense-not-even-science) insisted on, on ... Well. That's a _long story_—for another time, perhaps. For _now_, I want to explain how my robot cult's foundational texts had an enormous influence on my self-concept in relation to sex and gender. It all started in summer 2007 (I was nineteen years old), when I came across _Overcoming Bias_, a blog on the theme of how to achieve more accurate beliefs. (I don't remember exactly how I was referred, but I think it was likely to have been [a link from Megan McArdle](https://web.archive.org/web/20071129181942/http://www.janegalt.net/archives/009783.html), then writing as "Jane Galt" at _Asymmetrical Information_.) [Although](http://www.overcomingbias.com/author/hal-finney) [technically](http://www.overcomingbias.com/author/james-miller) [a](http://www.overcomingbias.com/author/david-j-balan) [group](http://www.overcomingbias.com/author/andrew) [blog](http://www.overcomingbias.com/author/anders-sandberg), the vast majority of posts on _Overcoming Bias_ were by Robin Hanson or Eliezer Yudkowsky. I was previously acquainted in passing with Yudkowsky's [writing about future superintelligence](http://yudkowsky.net/obsolete/tmol-faq.html). (I had [mentioned him in my Diary once in 2005](/ancillary/diary/42/), albeit without spelling his name correctly.) Yudkowsky was now using _Overcoming Bias_ and the medium of blogging [to generate material for a future book about rationality](https://www.lesswrong.com/posts/vHPrTLnhrgAHA96ko/why-i-m-blooking). Hanson's posts I could take or leave, but Yudkowsky's sequences of posts about rationality (coming out almost-daily through early 2009, eventually totaling hundreds of thousands of words) were _amazingly great_, drawing on fields from [cognitive psychology](https://www.lesswrong.com/s/5g5TkQTe9rmPS5vvM) to [evolutionary biology](https://www.lesswrong.com/s/MH2b8NfWv22dBtrs8) to explain the [mathematical](https://www.readthesequences.com/An-Intuitive-Explanation-Of-Bayess-Theorem) [principles](https://www.readthesequences.com/A-Technical-Explanation-Of-Technical-Explanation) [governing](https://www.lesswrong.com/posts/eY45uCCX7DdwJ4Jha/no-one-can-exempt-you-from-rationality-s-laws) _how intelligence works_—[the reduction of "thought"](https://www.lesswrong.com/posts/p7ftQ6acRkgo6hqHb/dreams-of-ai-design) to [_cognitive algorithms_](https://www.lesswrong.com/posts/HcCpvYLoSFP4iAqSz/rationality-appreciating-cognitive-algorithms). Intelligent systems that use [evidence](https://www.lesswrong.com/posts/6s3xABaXKPdFwA3FS/what-is-evidence) to construct [predictive](https://www.lesswrong.com/posts/a7n8GdKiAZRX86T5A/making-beliefs-pay-rent-in-anticipated-experiences) models of the world around them—that have "true" "beliefs"—can _use_ those models to compute which actions will best achieve their goals. You simply won't believe how much this blog will change your life; I would later frequently [joke](https://en.wiktionary.org/wiki/ha_ha_only_serious) that Yudkowsky rewrote my personality over the internet. There are a few things about me that I need to explain before I get into the topic-specific impact the blog had on me. The first thing—the chronologically first thing. Ever since I was fourteen years old— (and I _really_ didn't expect to be blogging about this eighteen years later) (I _still_ don't want to be blogging about this, but unfortunately, it actually turns out to be central to the intellectual–political project I've been singlemindedly focused on for the past four years because [somebody has to and no one else will](https://unsongbook.com/chapter-6-till-we-have-built-jerusalem/)) —my _favorite_—and basically only—masturbation fantasy has always been some variation on me getting magically transformed into a woman. I ... need to write more about the phenomenology of this. In the meantime, just so you know what I'm talking about, the relevant TVTrope is ["Man, I Feel Like a Woman."](https://tvtropes.org/pmwiki/pmwiki.php/Main/ManIFeelLikeAWoman) (Or search "body swap" on PornHub.) So, there was that erotic thing, which I was pretty ashamed of at the time, and _of course_ knew that I must never tell a single soul about. (It would have been about three years since the fantasy started that I even worked up the bravery to [tell my Diary about it](/ancillary/diary/53/#first-agp-confession).) But within a couple years, I also developed this beautiful pure sacred self-identity thing, where I was also having a lot of _non_-sexual thoughts about being a girl. Just—little day-to-day thoughts. Like when I would write in my pocket notebook in the persona of my female analogue. Or when I would practice swirling the descenders on all the lowercase letters that had descenders [(_g_, _j_, _p_, _y_, _z_)](/images/handwritten_phrase_jazzy_puppy.jpg) because I thought it made my handwriting look more feminine. Or the time when track and field practice split up into boys and girls, and I ironically muttered under my breath, "Why did I even join this team?—boys, I mean." And so on. The beautiful pure sacred self-identity thing doesn't _feel_ explicitly erotic. The thing I did in the day in class about writing in my notebook about being a girl, was _very different_ from the thing I did in my room at night about _visualizing_ girls with this abstract sense of "But what if that were _me_?" while furiously masturbating. The former activity was my beautiful pure happy romantic daydream, whereas the latter activity was not beautiful or pure at all! Now I am not a cognitive scientist, and can't claim to _know_ exactly what my beautiful pure sacred self-identity thing is, or where it comes from—that's [not the kind of thing I would expect people to _know_ from introspection alone](/2016/Sep/psychology-is-about-invalidating-peoples-identities/). But it has always seemed like a pretty obvious guess that there must have been _some sort of causal relationship_ between the erotic thing, and the beautiful pure sacred self-identity thing, even if the two things don't _feel_ the same: the overlap in subject matter is too much to be a coincidence. And the erotic thing definitely came _first_. Maybe this story reads differently in 2020 from how it was to live in 2005? I think that teenage boys in today's world having the kind of feelings I was having then, upon referencing or hinting at the beautiful pure sacred self-identity thing— (and the beautiful pure sacred self-identity thing is _much_ easier to talk about than the erotic thing) (I mean, the beautiful pure sacred self-identity thing is much harder to talk about _clearly_, but talking about it _un_-clearly is less shameful and requires much less bravery) —are immediately provided with "Oh, that means you're not a cis boy; you're a trans girl" as the definitive explanation. But it was a different time, then. Of course I had _heard of_ transsexualism as a thing, in the form of the "woman trapped in a man's body" trope, but it wasn't something I expected to actually encounter in real life. At the time, I had _no reason to invent the hypothesis_ that I might somehow literally be a woman in some unspecified psychological sense. I knew I was a boy _because_ boys are the ones with penises. That's what the word _means_. I was a boy who had a weird _sex fantasy_ about being a girl. That was just the obvious ordinary straightforward plain-language description of the situation. It _never occured to me_ to couch it in the language of "dysphoria", or actually possessing some innate "gender identity". The beautiful pure sacred self-identity thing was about identifying _with_ women, not identifying _as_ a woman. It's roughly analogous to how a cat lover might be said to "identify with" cats, without claiming to somehow _be_ a cat, because _that would be crazy_. This brings me to the other thing I need to explain about my teenage years, which is that I became very passionate about—well, in retrospect I call it _psychological-sex-differences denialism_, but at the time I called it _antisexism_. Where sometimes people in the culture would make claims about how women and men are psychologically different, and of course I knew this was _bad and wrong_. [TODO: sentence about how "gender identity" was offensive to me, link to "A Beacon"] So, you know, I read a lot about feminism. I remember checking out _The Feminine Mystique_ and Susan Faludi's _Backlash_ from the school library. Before I found my home on _Overcoming Bias_, I would read the big feminist blogs—_Pandagon_, _Feministe_, _Feministing_. It also seems like a pretty obvious guess that there must have been _some sort of causal relationship_ between my antisexism and the erotic and beautiful-pure-sacred self-identity things. True, the blank slate doctrine has been ideologically fashionable my entire life. In the sense that progressivism has been likened to a nontheistic state religion, I was a _very_ religious teenager. But there was presumably a _reason_ I cared so much about being a good pro-feminist, and hardly spent any time at all thinking about, _e.g._, racial justice. So, that's some background about where I was at, personally and ideologically, _before_ Eliezer Yudkowsky rewrote my personality over the internet. My ideological committment to psychological-sex-differences denialism made me uncomfortable when the topic of sex differences happened to come up on the blog—which wasn't particularly often at all, but in such a _vast_ body of work as the Sequences, it did happen to come up a few times (and those few times are the subject of this blog post). For example, as part of [an early explanation of why the values we would want to program into an artificial superintelligence don't reduce to any one simple principle](https://www.lesswrong.com/posts/NnohDYHNnKDtbiMyp/fake-utility-functions), Yudkowsky remarks that "the love of a man for a woman, and the love of a woman for a man, have not been cognitively derived from each other or from any other value." From the perspective of axiomatic antisexism, this assertion is at least somewhat cringe-inducing. Of course most people are straight, but is it not all the _same love_? Well ... think about it. [...] The short story ["Failed Utopia #4-2"](https://www.lesswrong.com/posts/ctpkTaqTKbmm6uRgC/failed-utopia-4-2) portrays an almost-aligned superintelligence constructing a happiness-maximizing utopia for humans—except that because [evolution didn't design women and men to be optimal partners for each other](https://www.lesswrong.com/posts/Py3uGnncqXuEfPtQp/interpersonal-entanglement), and the AI is prohibited from editing people's minds, the happiness-maximizing solution ends up splitting up the human species by sex and giving women and men their own _separate_ utopias, complete with artificially-synthesized romantic partners. At the time, [I expressed horror](https://www.greaterwrong.com/posts/ctpkTaqTKbmm6uRgC/failed-utopia-4-2/comment/PhiGnX7qKzzgn2aKb) at the idea in the comments section, because my quasi-religious psychological-sex-differences denialism required that I be horrified. But looking back eleven years later (my deconversion from my teenage religion being pretty thorough at this point, I think), the _argument makes sense_ (though you need an additional [handwave](https://tvtropes.org/pmwiki/pmwiki.php/Main/HandWave) to explain why the AI doesn't give every _individual_ their separate utopia—if existing women and men aren't optimal partners for each other, so too are individual men not optimal same-sex friends for each other). On my reading of the text, it is _significant_ that the AI-synthesized complements for men are given their own name, the _verthandi_, rather than just being referred to as women. The _verthandi_ may _look like_ women, they may be _approximately_ psychologically human, but the _detailed_ psychology of "superintelligently-engineered optimal romantic partner for a human male" is not going to come out of the distribution of actual human females, and judicious exercise of the [tenth virtue of precision](http://yudkowsky.net/rational/virtues/) demands that a _different word_ be coined for this hypothetical science-fictional type of person. Calling the _verthandi_ "women" would be _worse writing_; it would _fail to communicate_ the impact of what has taken place in the story. Another post in this vein that had a huge impact on me was ["Changing Emotions"](https://www.lesswrong.com/posts/QZs4vkC7cbyjL9XA9/changing-emotions). As an illustration of how [the hope for radical human enhancement is fraught with](https://www.lesswrong.com/posts/EQkELCGiGQwvrrp3L/growing-up-is-hard) technical difficulties, the Great Teacher sketches a picture of just how difficult an actual male-to-female sex change would be. It would be hard to overstate how much of an impact this post had on me. I've previously linked it on this blog eight times. In June 2008, half a year before it was published, I encountered the [2004 mailing list post](http://lists.extropy.org/pipermail/extropy-chat/2004-September/008924.html) that was its predecessor. (The fact that I was trawling through old mailing list archives searching for Yudkowsky content that I hadn't already read, tells you something about what a fanboy I am.) I immediately wrote to a friend: "[...] I cannot adequately talk about my feelings. Am I shocked, liberated, relieved, scared, angry, amused?" The argument goes: it might be easy to _imagine_ changing sex and refer to the idea in a short English sentence, but the real physical world has implementation details, and the implementation details aren't filled in by the short English sentence. The human body, including the brain, is an enormously complex integrated organism; there's no [plug-and-play](https://en.wikipedia.org/wiki/Plug_and_play) architecture by which you can just swap your brain into a new body and have everything work without re-mapping the connections in your motor cortex. And even that's not _really_ a sex change, as far as the whole integrated system is concerned— > Remapping the connections from the remapped somatic areas to the pleasure center will ... give you a vagina-shaped penis, more or less. That doesn't make you a woman. You'd still be attracted to girls, and no, that would not make you a lesbian; it would make you a normal, masculine man wearing a female body like a suit of clothing. But from the standpoint of my secret erotic fantasy, this is actually a _great_ outcome. [...] > If I fell asleep and woke up as a true woman—not in body, but in brain—I don't think I'd call her "me". The change is too sharp, if it happens all at once. In the comments, [I wrote](https://www.greaterwrong.com/posts/QZs4vkC7cbyjL9XA9/changing-emotions/comment/4pttT7gQYLpfqCsNd)— > Is it cheating if you deliberately define your personal identity such that the answer is _No_? (To which I now realize the correct answer is: Yes, it's fucking cheating! The map is not the territory! You can't change the current _referent_ of "personal identity" with the semantic mind game of declaring that "personal identity" now refers to something else! How dumb do you think we are?! But more on this later.)