From a2b8774420593c5f98538fa6651208700e929a24 Mon Sep 17 00:00:00 2001 From: "M. Taylor Saotome-Westlake" Date: Sat, 24 Sep 2022 11:19:04 -0700 Subject: [PATCH] check in --- ...-hill-of-validity-in-defense-of-meaning.md | 38 +++++++++---------- notes/a-hill-email-review.md | 36 +++++++++++++++++- notes/a-hill-of-validity-sections.md | 8 +++- 3 files changed, 60 insertions(+), 22 deletions(-) diff --git a/content/drafts/a-hill-of-validity-in-defense-of-meaning.md b/content/drafts/a-hill-of-validity-in-defense-of-meaning.md index f294089..699a289 100644 --- a/content/drafts/a-hill-of-validity-in-defense-of-meaning.md +++ b/content/drafts/a-hill-of-validity-in-defense-of-meaning.md @@ -196,23 +196,25 @@ Anna didn't reply, but I apparently did interest Michael, who chimed in on the e As for the attempt to intervene on Yudkowsky—here I need to make a digression about the constraints I'm facing in telling this Whole Dumb Story. _I_ would prefer to just tell the Whole Dumb Story as I would to my long-neglected Diary—trying my best at the difficult task of explaining _what actually happened_ in a very important part of my life, without thought of concealing anything. But a lot of _other people_ seem to have strong intuitions about "privacy", which bizarrely impose constraints on what _I'm_ allowed to say about my own life: in particular, it's considered unacceptable to publicly quote or summarize someone's emails from a conversation that they had reason to expect to be private. I feel obligated to comply with these widely-held privacy norms, even if _I_ think they're paranoid and [anti-social](http://benjaminrosshoffman.com/blackmailers-are-privateers-in-the-war-on-hypocrisy/). -So I would _think_ that the commonsense rule I should hold myself to while telling this Whole Dumb Story is that I obviously have an inalienable right to blog about _my own_ actions, but that I'm not allowed to refer to private conversations in cases where I don't think I'd be able to get the consent of the other party. (I don't think I'm required to go through the ritual of asking for consent in cases where the revealed information couldn't reasonably be considered "sensitive", or if I know the person doesn't have hangups about privacy.) Specifically, I'm allowed to talk about _me_ emailing Yudkowsky (because that was _my_ action), but I'm not allowed to talk about anything he might have said in reply, or whether he replied. +So I would _think_ that the commonsense privacy-norm-compliance rule I should hold myself to while telling this Whole Dumb Story is that I obviously have an inalienable right to blog about _my own_ actions, but that I'm not allowed to refer to private conversations in cases where I don't think I'd be able to get the consent of the other party. (I don't think I'm required to go through the ritual of asking for consent in cases where the revealed information couldn't reasonably be considered "sensitive", or if I know the person doesn't have hangups about this weird "privacy" thing.) In this case, I'm allowed to talk about _me_ emailing Yudkowsky (because that was _my_ action), but I'm not allowed to talk about anything he might have said in reply, or whether he replied. Unfortunately, there's a potentially serious loophole in the commonsense rule: what if some of my actions (which I would have _hoped_ to have an inalienable right to blog about) _depend on_ content from private conversations? You can't, in general, only reveal one side of a conversation. -Supppose Alice messages Bob at 5 _p.m._, "Can you come to the party?", and also, separately, that Alice messages Bob at 6 _p.m._, "Gout isn't contagious." Should Alice be allowed to blog about the messages she sent at 6 _p.m._ and 7 _p.m._, because she was only describing her own actions, and didn't confirm or deny whether Bob replied at all, let alone quote him? +Supppose Alice messages Bob at 5 _p.m._, "Can you come to the party?", and also, separately, that Alice messages Bob at 6 _p.m._, "Gout isn't contagious." Should Alice be allowed to blog about the messages she sent at 6 _p.m._ and 7 _p.m._, because she's only describing her own messages, and not confirming or denying whether Bob replied at all, let alone quoting him? -I think commonsense privacy-norm-adherence intuitions actually say _No_ here: the text of Alice's messages makes it too easy to guess that sometime between 5 and 6, Bob probably said that he couldn't come to the party because he has gout. It would seem that Alice's right to talk about her own actions in her own life _does_ need to take into account some commonsense judgement whether that leaks "sensitive" information about Bob. +I think commonsense privacy-norm-adherence intuitions actually say _No_ here: the text of Alice's messages makes it too easy to guess that sometime between 5 and 6, Bob probably said that he couldn't come to the party because he has gout. It would seem that Alice's right to talk about her own actions in her own life _does_ need to take into account some commonsense judgement of whether that leaks "sensitive" information about Bob. -In part of the Dumb Story that follows, I'm going to describe several times when I and others emailed Yudkowsky to try to argue with what he said in public, without saying anything about whether Yudkowsky replied, or what he might have said if he did reply. I maintain that I'm within my rights here, because I think commonsense judgement will agree that me talking about the arguments _I_ made, does not in this case leak any sensitive information about the other side of a conversation that may or may not have happened: the story comes off about the same whether Yudkowsky didn't reply at all, or whether he replied in a way that I found sufficiently unsatisfying as to merit the futher emails with followup arguments that I describe. (Talking about later emails _does_ rule out the possible world where Yudkowsky had said, "Please stop contacting me," because I would have respected that, but the fact that he didn't say that isn't "sensitive".) +In part of the Dumb Story that follows, I'm going to describe several times when I and others emailed Yudkowsky to try to argue with what he said in public, without saying anything about whether Yudkowsky replied, or what he might have said if he did reply. I maintain that I'm within my rights here, because I think commonsense judgement will agree that me talking about the arguments _I_ made, does not in this case leak any sensitive information about the other side of a conversation that may or may not have happened: I think the story comes off relevantly the same whether Yudkowsky didn't reply at all (_e.g._, because he was busy with more existentially important things and didn't check his email), or whether he replied in a way that I found sufficiently unsatisfying as to occasion the futher emails with followup arguments that I describe. (Talking about later emails _does_ rule out the possible world where Yudkowsky had said, "Please stop emailing me," because I would have respected that, but the fact that he didn't say that isn't "sensitive".) -It seems particularly important to lay out these principles of adherence to privacy norms in connection to my attempts to contact Yudkowsky, because part of what I'm trying to accomplish in telling this Whole Dumb Story is to deal reputational damage to Yudkowsky, which I claim is deserved. (We want reputations to track reality. If you see Carol exhibiting a pattern of intellectual dishonesty, and she keeps doing it even after you try talking to her about it privately, you might want to write a blog post describing the pattern in detail—not to _hurt_ Carol, particularly, but so that everyone else can make higher-quality decisions about whether they should believe the things that Carol says.) Given that motivation of mine, it seems important that I only try to hang Yudkowsky with the rope of what he said in public, where you can click the links and read the context for yourself. In the Dumb Story that follows, I _also_ describe some of my correspondence with Scott Alexander, but that doesn't seem sensitive in the same way, because I'm not particularly trying to deal reputational damage to Scott. (Not because Scott performed well, but because one wouldn't really have _expected_ Scott to perform well in this situation; a public reputation-update isn't called for in the same way.) +It seems particularly important to lay out these principles of adherence to privacy norms in connection to my attempts to contact Yudkowsky, because part of what I'm trying to accomplish in telling this Whole Dumb Story is to deal reputational damage to Yudkowsky, which I claim is deserved. (We want reputations to track reality. If you see Carol exhibiting a pattern of intellectual dishonesty, and she keeps doing it even after you try talking to her about it privately, you might want to write a blog post describing the pattern in detail—not to _hurt_ Carol, particularly, but so that everyone else can make higher-quality decisions about whether they should believe the things that Carol says.) Given that motivation of mine, it seems important that I only try to hang Yudkowsky with the rope of what he said in public, where you can click the links and read the context for yourself. In the Dumb Story that follows, I _also_ describe some of my correspondence with Scott Alexander, but that doesn't seem sensitive in the same way, because I'm not particularly trying to deal reputational damage to Alexander in the same way. (Not because Scott performed well, but because one wouldn't really have _expected_ Scott to perform well in this situation.) In accordance with the privacy-norm-adherence policy just described, I don't think I should say whether Yudkowsky replied to Michael's and my emails, nor ([again](/2022/TODO/blanchards-dangerous-idea-and-the-plight-of-the-lucid-crossdreamer/#cheerful-price-privacy-constraint)) whether he accepted the cheerful price money, because any conversation that may or may not have occured would have been private. But what I _can_ say, because it was public, is that we saw [this addition to the Twitter thread](https://twitter.com/ESYudkowsky/status/1068071036732694529): > I was sent this (by a third party) as a possible example of the sort of argument I was looking to read: [http://unremediatedgender.space/2018/Feb/the-categories-were-made-for-man-to-make-predictions/](/2018/Feb/the-categories-were-made-for-man-to-make-predictions/). Without yet judging its empirical content, I agree that it is not ontologically confused. It's not going "But this is a MAN so using 'she' is LYING." -Look at that! The great Eliezer Yudkowsky said that my position is not ontologically confused. That's _probably_ high praise coming from him! You might think that that should have been the end of the story. Yudkowsky denounced a particular philosophical confusion, I already had a related objection written up, and he acknowledged my objection as not being the confusion he was trying to police. I _should_ be satisfied, right? +Look at that! The great Eliezer Yudkowsky said that my position is "not ontologically confused." That's _probably_ high praise coming from him! + +You might think that that should have been the end of the story. Yudkowsky denounced a particular philosophical confusion, I already had a related objection written up, and he acknowledged my objection as not being the confusion he was trying to police. I _should_ be satisfied, right? I wasn't, in fact, satisfied. This little "not ontologically confused" clarification buried deep in the replies was _much less visible_ than the bombastic, arrogant top level pronouncement insinuating that resistance to gender-identity claims _was_ confused. (1 Like on this reply, _vs._ 140 Likes/21 Retweets on start of thread.) I expected that the typical reader who had gotten the impression from the initial thread that Yudkowsky thought that gender-identity skeptics didn't have a leg to stand on, would not, actually, be disabused of this impression by the existence of this little follow-up. Was it greedy of me to want something _louder_? @@ -222,23 +224,23 @@ The _issue_ is that category boundaries are not arbitrary (if you care about int It's true that [the reason _I_ was continuing to freak out about this](/2019/Jul/the-source-of-our-power/) to the extent of sending him this obnoxious email telling him what to write (seriously, who does that?!) had to with transgender stuff, but wasn't the reason _Scott_ should care. -The other year, Alexander had written a post, ["Kolmogorov Complicity and the Parable of Lightning"](http://slatestarcodex.com/2017/10/23/kolmogorov-complicity-and-the-parable-of-lightning/), explaining the consequences of political censorship by means of an allegory about a Society with the dogma that thunder occurs before lightning. The problem isn't so much the sacred dogma itself (it's not often that you need to _directly_ make use of the fact that lightning comes first), but that the need to _defend_ the sacred dogma _destroys everyone's ability to think_. +The other year, Alexander had written a post, ["Kolmogorov Complicity and the Parable of Lightning"](http://slatestarcodex.com/2017/10/23/kolmogorov-complicity-and-the-parable-of-lightning/), explaining the consequences of political censorship by means of an allegory about a Society with the dogma that thunder occurs before lightning. The problem isn't so much the sacred dogma itself (it's not often that you need to _directly_ make use of the fact that lightning comes first), but that [the need to _defend_ the sacred dogma](https://www.lesswrong.com/posts/wyyfFfaRar2jEdeQK/entangled-truths-contagious-lies) [_destroys everyone's ability to think_](https://www.lesswrong.com/posts/XTWkjCJScy2GFAgDt/dark-side-epistemology). It was the same thing here. It wasn't that I had any direct practical need to misgender anyone in particular. It still wasn't okay that trying to talk about the reality of biological sex to so-called "rationalists" got you an endless deluge of—polite! charitable! non-ostracism-threatening!—_bullshit nitpicking_. (What about [complete androgen insensitivity syndrome](https://en.wikipedia.org/wiki/Complete_androgen_insensitivity_syndrome)? Why doesn't this ludicrous misinterpretation of what you said [imply that lesbians aren't women](https://thingofthings.wordpress.com/2018/06/18/man-should-allocate-some-more-categories/)? _&c. ad infinitum_.) With enough time, I thought the nitpicks could and should be satisfactorily answered. (Any ones that couldn't would presumably be fatal criticisms rather than bullshit nitpicks.) But while I was in the process of continuing to write all that up, I hoped Alexander could see why I feel somewhat gaslighted. (I had been told by others that I wasn't using the word "gaslighting" correctly. _Somehow_ no one seemed to think I had the right to define _that_ category boundary for my convenience.) -If our vaunted rationality techniques result in me having to spend dozens of hours patiently explaining why I don't think that I'm a woman and that [the person in this photograph](https://daniellemuscato.startlogic.com/uploads/3/4/9/3/34938114/2249042_orig.jpg) isn't a woman, either (where "isn't a woman" is a _convenient rhetorical shorthand_ for a much longer statement about [naïve Bayes models](https://www.lesswrong.com/posts/gDWvLicHhcMfGmwaK/conditional-independence-and-naive-bayes) and [high-dimensional configuration spaces](https://www.lesswrong.com/posts/WBw8dDkAWohFjWQSk/the-cluster-structure-of-thingspace) and [defensible Schelling points for social norms](https://www.lesswrong.com/posts/Kbm6QnJv9dgWsPHQP/schelling-fences-on-slippery-slopes)), then our techniques are _worse than useless_. +If our vaunted rationality techniques resulted in me having to spend dozens of hours patiently explaining why I didn't think that I was a woman and that [the person in this photograph](https://daniellemuscato.startlogic.com/uploads/3/4/9/3/34938114/2249042_orig.jpg) wasn't a woman, either (where "isn't a woman" is a _convenient rhetorical shorthand_ for a much longer statement about [naïve Bayes models](https://www.lesswrong.com/posts/gDWvLicHhcMfGmwaK/conditional-independence-and-naive-bayes) and [high-dimensional configuration spaces](https://www.lesswrong.com/posts/WBw8dDkAWohFjWQSk/the-cluster-structure-of-thingspace) and [defensible Schelling points for social norms](https://www.lesswrong.com/posts/Kbm6QnJv9dgWsPHQP/schelling-fences-on-slippery-slopes)), then our techniques were _worse than useless_. -If Galileo ever muttered "And yet it moves", there's a long and nuanced conversation you could have about the consequences of using the word "moves" in Galileo's preferred sense or some other sense that happens to result in the theory needing more epicycles. It may not have been obvious in November 2014, but in retrospect, _maybe_ it was a _bad_ idea to build a [memetic superweapon](https://archive.is/VEeqX) that says that the number of epicycles _doesn't matter_. +If Galileo ever muttered "And yet it moves", there's a long and nuanced conversation you could have about the consequences of using the word "moves" in Galileo's preferred sense, or some other sense that happens to result in the theory needing more epicycles. It may not have been obvious in November 2014, but in retrospect, _maybe_ it was a _bad_ idea to build a [memetic superweapon](https://archive.is/VEeqX) that says that the number of epicycles _doesn't matter_. -And the reason to write this as a desperate email plea to Scott Alexander when I could be working on my own blog, was that I was afraid that marketing is a more powerful force than argument. Rather than good arguments propagating through the population of so-called "rationalists" no matter where they arise, what actually happens is that people like Alexander and Yudkowsky rise to power on the strength of good arguments and entertaining writing (but mostly the latter), and then everyone else sort-of absorbs most of their worldview (plus noise and [conformity with the local environment](https://thezvi.wordpress.com/2017/08/12/what-is-rationalist-berkleys-community-culture/)). So for people who didn't [win the talent lottery](http://slatestarcodex.com/2015/01/31/the-parable-of-the-talents/) but think they see a flaw in the _Zeitgeist_, the winning move is "persuade Scott Alexander." +And the reason to write this as a desperate email plea to Scott Alexander when I could be working on my own blog, was that I was afraid that marketing is a more powerful force than argument. Rather than good arguments propagating through the population of so-called "rationalists" no matter where they arise, what actually happens is that people like Alexander and Yudkowsky rise to power on the strength of good arguments and entertaining writing (but mostly the latter), and then everyone else sort-of absorbs some of their worldview (plus noise and [conformity with the local environment](https://thezvi.wordpress.com/2017/08/12/what-is-rationalist-berkleys-community-culture/)). So for people who didn't [win the talent lottery](http://slatestarcodex.com/2015/01/31/the-parable-of-the-talents/) but think they see a flaw in the _Zeitgeist_, the winning move is "persuade Scott Alexander." -Back in 2010, the rationalist community had a shared understanding that the function of language is to describe reality. Now, we didn't. If Scott didn't want to cite my creepy blog about my creepy fetish, that was _totally fine_; I _liked_ getting credit, but the important thing is that this "No, the Emperor isn't naked—oh, well, we're not claiming that he's wearing any garments—it would be pretty weird if we were claiming _that!_—it's just that utilitarianism implies that the _social_ property of clothedness should be defined this way because to do otherwise would be really mean to people who don't have anything to wear" gaslighting maneuver needed to _die_, and he alone could kill it. +Back in 2010, the rationalist community had a shared understanding that the function of language is to describe reality. Now, we didn't. If Scott didn't want to cite my creepy blog about my creepy fetish, that was _totally fine_; I liked getting credit, but the important thing is that this "No, the Emperor isn't naked—oh, well, we're not claiming that he's wearing any garments—it would be pretty weird if we were claiming _that!_—it's just that utilitarianism implies that the _social_ property of clothedness should be defined this way because to do otherwise would be really mean to people who don't have anything to wear" gaslighting maneuver needed to _die_, and he alone could kill it. -... Scott didn't get it. We agreed that self-identity, natal-sex, and passing-based gender categories each had their own pros and cons, and that it's uninteresting to focus on whether something "really" belongs to a category, rather than on communicating what you mean. Scott took this to mean that what convention to use is a pragmatic choice that we can make on utilitarian grounds, and that being nice to trans people is worth a little bit of clunkiness. +... Scott didn't get it. We agreed that self-identity-, natal-sex-, and passing-based gender categories each had their own pros and cons, and that it's uninteresting to focus on whether something "really" belongs to a category, rather than on communicating what you mean. Scott took this to mean that what convention to use is a pragmatic choice that we can make on utilitarian grounds, and that being nice to trans people is worth a little bit of clunkiness. -But I considered myself to be prosecuting _not_ the object-level question of which gender categories to use, but the meta-level question of what laws of thought underlie the cognitive function of categorization, for which, "whatever, it's a pragmatic choice, just be nice" wasn't an answer. I didn't have a simple, [mistake-theoretic](https://slatestarcodex.com/2018/01/24/conflict-vs-mistake/) characterization of the language and social conventions that everyone should use such that anyone who defected from the compromise would be wrong. The best I could do was try to objectively predict the consequences of different possible conventions—and of _conflicts_ over possible conventions. +But I considered myself to be prosecuting _not_ the object-level question of which gender categories to use, but the meta-level question of what _laws of thought_ underlie the cognitive function of categorization, for which, "whatever, it's a pragmatic choice, just be nice" wasn't an answer. I didn't have a simple, [mistake-theoretic](https://slatestarcodex.com/2018/01/24/conflict-vs-mistake/) characterization of the language and social conventions that everyone should use such that anyone who defected from the compromise would be wrong. The best I could do was try to objectively predict the consequences of different possible conventions—and of _conflicts_ over possible conventions. ["... Not Man for the Categories"](https://slatestarcodex.com/2014/11/21/the-categories-were-made-for-man-not-man-for-the-categories/) had concluded with a section on Emperor Norton, a 19th century San Francisco resident who declared himself Emperor of the United States. Certainly, it's not difficult or costly for the citizens of San Francisco to _address_ Norton as "Your Majesty" as a courtesy or a nickname. But there's more to being the Emperor of the United States than people calling you "Your Majesty." Unless we abolish Congress and have the military enforce Norton's decrees, he's not _actually_ functioning in the role of emperor—at least not according to the currently generally-understood meaning of the word "emperor." @@ -248,9 +250,7 @@ You're not really going to bring him Rutherford B. Hayes's head. So what are you To be sure, words can be used in many ways depending on context, but insofar as Norton _is_ interpreting "emperor" in the traditional sense, and you keep calling him your emperor without caveats or disclaimers, _you are lying to him_. -... Scott still didn't get it. - -Anyway, I _did_ end up in more conversation with Michael Vassar, Ben Hoffman, and Sarah Constantin, who were game to help me with reaching out to Yudkowsky again to explain the problem in more detail—and to appeal to the conscience of someone who built their career on [higher standards](https://www.lesswrong.com/posts/DoLQN5ryZ9XkZjq5h/tsuyoku-naritai-i-want-to-become-stronger). +... Scott still didn't get it. But I _did_ soon end up in more conversation with Michael Vassar, Ben Hoffman, and Sarah Constantin, who were game to help me with reaching out to Yudkowsky again to explain the problem in more detail—and to appeal to the conscience of someone who built their career on [higher standards](https://www.lesswrong.com/posts/DoLQN5ryZ9XkZjq5h/tsuyoku-naritai-i-want-to-become-stronger). Yudkowsky probably didn't think much of _Atlas Shrugged_ (judging by [an offhand remark by our protagonist in _Harry Potter and the Methods_](http://www.hpmor.com/chapter/20)), but I kept thinking of the part where our heroine Dagny Taggart entreats the great Dr. Robert Stadler to denounce [an egregiously deceptive but technically-not-lying statement](https://www.lesswrong.com/posts/MN4NRkMw7ggt9587K/firming-up-not-lying-around-its-edge-cases-is-less-broadly) by the State Science Institute, whose legitimacy derives from its association with his name. Stadler has become cynical in his old age and demurs, disclaiming all responsibility: "I can't help what people think—if they think at all!" ... "How can one deal with truth when one deals with the public?" @@ -264,7 +264,7 @@ There were times during these weeks where it felt like my mind shut down with th Except ... I had been raised back in the 'aughts to believe that you're you're supposed to concede arguments on the basis of encountering a superior counterargument that makes you change your mind, and I couldn't actually point to one. "Maybe I'm making a fool out of myself by picking fights with all these high-status people" is _not a counterargument_. -Meanwhile, Anna continued to be disinclined to take a side in the brewing Category War, and it was beginning to put a strain on our friendship, to the extent that I kept ending up crying at some point during our occasional meetings. She told me that my "You have to pass my philosophy-of-language litmus test or I lose all respect for you as a rationalist" attitude was psychologically coercive. I agreed—I was even willing to go up to "violent"—in the sense that I'd cop to [trying to apply social incentives towards an outcome rather than merely exchanging information](http://zackmdavis.net/blog/2017/03/an-intuition-on-the-bayes-structural-justification-for-free-speech-norms/). But sometimes you need to use violence in defense of self or property, even if violence is generally bad. If we think of the "rationalist" brand name as intellectual property, maybe it's property worth defending, and if so, then "I can define a word any way I want" isn't obviously a terrible time to start shooting at the bandits? +Meanwhile, Anna continued to be disinclined to take a side in the brewing Category War, and it was beginning to put a strain on our friendship, to the extent that I kept ending up crying at some point during our occasional meetings. She said that my "You have to pass my philosophy-of-language litmus test or I lose all respect for you as a rationalist" attitude was psychologically coercive. I agreed—I was even willing to go up to "violent"—in the sense that I'd cop to [trying to apply social incentives towards an outcome rather than merely exchanging information](http://zackmdavis.net/blog/2017/03/an-intuition-on-the-bayes-structural-justification-for-free-speech-norms/). But sometimes you need to use violence in defense of self or property, even if violence is generally bad. If we think of the "rationalist" brand name as intellectual property, maybe it was property worth defending, and if so, then "I can define a word any way I want" isn't obviously a terrible time to start shooting at the bandits? My _hope_ was that it was possible to apply just enough "What kind of rationalist are _you_?!" social pressure to cancel out the "You don't want to be a Bad ([Red](https://slatestarcodex.com/2014/09/30/i-can-tolerate-anything-except-the-outgroup/)) person, do you??" social pressure and thereby let people look at the arguments—though I wasn't sure if that actually works, and I was growing exhausted from all the social aggression I was doing about it. (If someone tries to take your property and you shoot at them, you could be said to be the "aggressor" in the sense that you fired the first shot, even if you hope that the courts will uphold your property claim later.) @@ -284,7 +284,7 @@ The problem with this kind of exchange is not that anyone is being shouted down, It's not that anyone explicitly said, "Biological sex isn't real" in those words. ([The elephant in the brain](https://en.wikipedia.org/wiki/The_Elephant_in_the_Brain) knows it wouldn't be able to get away with _that_.) But if everyone correlatedly plays dumb whenever someone tries to _talk_ about sex in clear language in a context where that could conceivably hurt some trans person's feelings, I think what you have is a culture of _de facto_ biological sex denialism. ("'The original thing that already exists without having to try' sounds fake to me"!!) It's not hard to get people to admit that trans women are different from cis women, but somehow they can't (in public, using words) follow the implication that trans women are different from cis women _because_ trans women are male. -Ben thought I was wrong to think of this kind of behavior as non-ostracisizing. The deluge of motivated nitpicking _is_ an implied marginalization threat, he explained: the game people are playing when they do that is to force me to choose between doing arbitarily large amounts of interpretive labor, or being cast as never having answered these construed-as-reasonable objections, and therefore over time losing standing to make the claim, being thought of as unreasonable, not getting invited to events, _&c._ +Ben thought I was wrong to think of this kind of behavior as non-ostracisizing. The deluge of motivated nitpicking _is_ an implied marginalization threat, he explained: the game people are playing when they do that is to force me to choose between doing arbitarily large amounts of [interpretive labor](https://acesounderglass.com/2015/06/09/interpretive-labor/), or being cast as never having answered these construed-as-reasonable objections, and therefore over time losing standing to make the claim, being thought of as unreasonable, not getting invited to events, _&c._ I saw the dynamic he was pointing at, but as a matter of personality, I was more inclined to respond, "Welp, I guess I need to write faster and more clearly", rather than to say, "You're dishonestly demanding arbitrarily large amounts of interpretive labor from me." I thought Ben was far too quick to give up on people who he modeled as trying not to understand, whereas I continued to have faith in the possibility of _making_ them understand if I just ... never gave up. Not to be _so_ much of a scrub as to play chess with a pigeon (which shits on the board and then struts around like it's won), or wrestle with a pig (which gets you both dirty, and the pig likes it), or dispute [what the Tortise said to Achilles](https://en.wikipedia.org/wiki/What_the_Tortoise_Said_to_Achilles)—but to hold out hope that people in "the community" could only be _boundedly_ motivatedly dense, and anyway that giving up wouldn't make me a stronger writer. @@ -940,7 +940,7 @@ The battle that matters—and I've been _very_ explicit about this, for years— This is a battle between Feelings and Truth, between Politics and Truth. -In order to take the side of Truth, you need to be able to tell Joshua Norton that he's not actually Emperor of the United States (even if it hurts him). You need to be able to tell a prideful autodidact that the fact that he's failing quizzes in community college differential equations class, is evidence that his study methods aren't doing what he thought they were (even if it hurts him). And you need to be able to say that trans women are male and trans men are female _with respect to_ a female/male "sex" concept that encompasses the many traits that aren't affected by contemporary surgical and hormonal interventions (even if it hurts someone who does not like to be tossed into a Male Bucket or a Female Bucket as it would be assigned by their birth certificate, and—yes—even if it probabilistically contributes to that person's suicide). +In order to take the side of Truth, you need to be able to tell Joshua Norton that he's not actually Emperor of the United States (even if it hurts him). You need to be able to tell a prideful autodidact that the fact that he's failing quizzes in community college differential equations class, is evidence that his study methods aren't doing what he thought they were (even if it hurts him). And you need to be able to say, in public, that trans women are male and trans men are female _with respect to_ a female/male "sex" concept that encompasses the many traits that aren't affected by contemporary surgical and hormonal interventions (even if it hurts someone who does not like to be tossed into a Male Bucket or a Female Bucket as it would be assigned by their birth certificate, and—yes—even if it probabilistically contributes to that person's suicide). If you don't want to say those things because hurting people is wrong, then you have chosen Feelings. diff --git a/notes/a-hill-email-review.md b/notes/a-hill-email-review.md index 9267ab9..b4e9cee 100644 --- a/notes/a-hill-email-review.md +++ b/notes/a-hill-email-review.md @@ -145,8 +145,42 @@ Michael: Once people think they know what the political sides are, the political 15 Apr: me—concern about bad faith nitpicking, Jessica says I'm not 15 Apr: Jessica's comment, the lesson of Funk-tunul 15 Apr: trying to recruit Steven Kaas -15 Apr: Signal message to Ben— +15 Apr: Signal conversation with Ben— +I'm having some ideas on how I could write my post about disillusionment with the rationalists well still not being too hostile and not leaking info from private conversations + +> Yeah, I think the right target to keep in mind is people like you a few years ago who haven't yet had the experiences you've had +> So they don't actually have to freak out to the point of being imprisoned and then demand help from community leaders and not get it, they can just learn from you +> Like, the actual sympathetic-but-naive people +> Not the people fucking with you + +[...] + +people also deserve credit for the help they did give me (care resources from Alicorn et al.) even if I don't think there intellectual performance was good enough to deserve the Mandate of Heaven implied by the "rationalist" brand name + +> Yes, but we need to keep separate accounts for _personal_ gratitude, and institutional fraud +> Like, I think that if somehow Alicorn gets hurt by the Rationalists dissolving, and we're in a position to help her, we should try to help in any way that _doesn't_ work via continuing to validate the collective narcissistic fantasy that the Rationalists are a thing +> Insofar as that makes things hard, we have to think of it as a hostage situations +> But, it's not obvious that this imposes undue costs on the individuals that have helped you a lot, except maybe Anna +[...] +> And, _not_ doing this is imposing costs on people like Michael (and all of humanity) instead, which is a lot less fair than harming Anna's interests incidentally by breaking up a scam she's drawing power from + +[...] +should I just reread you posts about EA, or is there more? + +> People are systematically conflating corruption, accumulation of dominance, and theft, with getting things done. +[...] + +> These both seem like evidence that CFAR and MIRI see themselves as building an army, not a truth-seeking intellectual community. +[...] + +> This isn't about direct benefit vs harm +> This is about what, substantively, they're doing +> And they claimed to be cultivating an epistemically rational community while in fact building an army of loyalists + +[...] + last night I was up until 2 a.m. and wrote Anna two paper postscards asking for her to engage with my "... Boundaries?" post when she ahs time and to give me an estimate of when she might have time [...] I figured that spamming people with hysterical and somewhat demanding physical postcards is more polite (and funnier) than my usual habit of spamming people with hysterical and somewhat demanding emails + 16 Apr: I feel sad and gaslighted 16 Apr: Jessica sparks engagement by appealing to anti-lying Ben— I think a commitment to lying is always wrong just straightforwardly leads to deciding that anything power demands you say is true. diff --git a/notes/a-hill-of-validity-sections.md b/notes/a-hill-of-validity-sections.md index 150b4eb..554a8d7 100644 --- a/notes/a-hill-of-validity-sections.md +++ b/notes/a-hill-of-validity-sections.md @@ -1,4 +1,5 @@ with internet available— +_ Gallileo "And yet it moves" _ Discord logs before Austin retreat _ examples of snarky comments about "the rationalists" _ screenshot Rob's Facebook comment which I link @@ -10,7 +11,6 @@ _ weirdly hostile comments on "... Boundaries?" far editing tier— _ rephrase "gamete size" discussion to make it clearer that Yudkowsky's proposal also implicitly requires people to be agree about the clustering thing _ smoother transition between "deliberately ambiguous" and "was playing dumb"; I'm not being paranoid for attributing political motives to him, because he told us that he's doing it -_ it's worth an extra two sentences to explain context of Robert Stadler comparison better _ I'm sure Eliezer Yudkowsky could think of some relevant differences _ clarify why Michael thought Scott was "gaslighting" me, include "beeseech bowels of Christ" _ conversation with Ben about physical injuries (this is important because it explains where the "cut my dick off rhetoric" came from) @@ -1224,4 +1224,8 @@ Still citing it (31 Jul 22): https://www.reddit.com/r/slatestarcodex/comments/wb Still citing it (19 Sep 22): https://twitter.com/ingalala/status/1568391691064729603 ------- \ No newline at end of file +------ + +If you _have_ intent-to-inform and occasionally end up using your megaphone to say false things (out of sloppiness or motivated reasoning in the passion of the moment), it's actually not that big of a deal, as long as you're willing to acknowledge corrections. (It helps if you have critics who personally hate your guts and therefore have a motive to catch you making errors, and a discerning audience who will only reward the critics for finding real errors and not fake errors.) In the long run, the errors cancel out. + +If you _don't_ have intent-to-inform, but make sure to never, ever say false things (because you know that "lying" is wrong, and think that as long as you haven't "lied", you're in the clear), but you don't feel like you have an obligation to acknowledge criticisms (for example, because you think you and your flunkies are the only real people in the world, and anyone who doesn't want to become one of your flunkies can be disdained as a "post-rat"), that's potentially a much worse situation, because the errors don't cancel. -- 2.17.1