-I had his email address from previous contract work I had done for MIRI back in '12, so I wrote him offering $1K to talk about sometime between [January 2009](https://www.lesswrong.com/posts/QZs4vkC7cbyjL9XA9/changing-emotions) and [March of the current year](https://www.facebook.com/yudkowsky/posts/10154078468809228) what kind of _massive_ update he made on the topics of human psychological sex differences and MtF transsexuality, mentioning that I had been "feeling baffled and disappointed (although I shouldn't be) that the rationality community is getting this _really easy_ scientific question wrong."
+Second, $1000 isn't actually real money to a San Francisco software engineer.
+
+Third—yes. Yes, it _absolutely_ was a little cultish. There's a sense in which, _sociologically and psychologically speaking_, Yudkowsky is a religious leader, and I was—am—a devout adherent of the religion he made up.
+
+[TODO: explain religion]
+
+One of my emails included the sentence, "I feel awful writing _Eliezer Yudkowsky_ about this, because my interactions with you probably have disproportionately more simulation-measure than the rest of my life, and do I _really_ want to spend that on _this topic_?"
+
+(Referring to the idea that, in a sufficiently large universe where many subjectively-indistinguishable copies of everyone exists, including inside of future superintelligences running simulations of the past, there would plausibly be _more_ copies of my interactions with Yudkowsky than of other moments of my life, on account of that information being of greater decision-relevance to those superintelligences.)