Building Cooperative Epistemology (Response to "EA has a Lying Problem", among other things)

by Raemon11th Jan 2017160 comments

18

Frontpage

This is in response to Sarah Constantin's recent post about intellectual dishonesty within the EA community.

I roughly agree with Sarah's main object level points, but I think this essay doesn't sufficiently embody the spirit of cooperative discourse it's trying to promote. I have a lot of thoughts here, but they are building off a few existing essays. (There's been a recent revival over on Less Wrong attempting to make it a better locus for high quality discussion. I don't know if it's especially succeeded, but I think the concepts behind that intended revival and very important)

  1. Why Our Kind Can't Cooperate (Eliezer Yudkowsky)
  2. A Return to Discussion (Sarah Constantin)
  3. The Importance of [Less Wrong, OR another Single Conversational Locus] (Emphasis mine) (Anna Salamon)
  4. The Four Layers of Intellectual Conversation (Eliezer Yudkowsky)

    I think it's important to have all three concepts in context before delving into:

  5. EA has a lying problem (Sarah Constantin)

I recommend reading all of those. But here's a rough summary of what I consider the important bits. (If you want to actually argue with these bits, please read the actual essays before doing so, so you're engaging with the full substance of the idea)

  • Intellectuals and contrarians love to argue and nitpick. This is valuable - it produces novel insights, and keeps us honest. BUT it makes it harder to actually work together to achieve things. We need to understand how working-together works on a deep enough level that we can do so without turning into another random institution that's lost it's purpose. (See Why Our Kind... for more)

  • Lately, people have tended to talk on social media (Facebook, Tumblr, etc) rather than in formal blogs or forums that encourage longform discussion. This has a few effects. (See A Return to Discussion for more)
    1. FB discussion is fragmented - it's hard to find everything that's been said on a topic. (And tumblr is even worse)
    2. It's hard to know whether OTHER people have read a given thing on a topic.
    3. A related point (not necessarily in "A Return to Discussion" is that social media incentives some of the worst kinda of discussion. People share things quickly, without reflection. People read and respond to things in 5-10 minute bursts, without having time to fully digest them. 

  • Having a single, long form discussion area that you can expect everyone in an intellectual community to have read, makes it much easier to building knowledge. (And most of human progress is due, not to humans being smart, but being able to stand on the shoulders of giants). Anna Salamon's "Importance of a Single Conversational Locus" is framed around x-risk, but I think it applies to all aspects of EA: the problems the world faces are so huge that they need a higher caliber of thinking and knowledge-building than we currently have in order to solve.

  • In order to make true intellectual progress, you need people to be able to make critiques. You also need those critics to expect their criticism to in turn be criticized, so that the criticism is high quality. If a critique turns out to be poorly thought out, we need shared, common knowledge of that so that people don't end up rehashing the same debates.

  • And finally, (one of) Sarah's points in "EA has a lying problem" is that, in order to be different from other movements and succeed where they failed, EA needs to hold itself to a higher standard than usual. There's been much criticism of, say, Intentional Insights for doing sketchy, truth-bendy things to gain prestige and power. But that plenty of "high status" people within the EA community do things that are similar, even if to a different degree. We need to be aware of that.

    I would not argue as strongly as Sarah does that we shouldn't do it at all, but it's worth periodically calling each other out on it.

Cooperative Epistemology

So my biggest point here, is that we need to be more proactive and mindful about how discussion and knowledge is built upon within the EA community.

To succeed at our goals:

  • EA needs to hold itself to a very high intellectual standard (higher than we currently have, probably. In some sense anyway)
  • Factions within EA needs to be able to cooperate, share knowledge. Both object level knowledge (i.e. how cost effective is AMF?) and meta/epistemic knowledge like:
    1. How do we evaluate messy studies
    2. How do we discuss things online so that people actually put effort into reading and contributing the discussion.
    3. What kinds of conversational/debate norms lead people to be more transparent.
  • We need to be able to apply all the knowledge to go out and accomplish things, which will probably involve messy political stuff.

I have specific concerns about Sarah's post, which I'll post in a comment when I have a bit more time.

 

77 comments, sorted by Highlighting new comments since Today at 12:29 PM
New Comment
Some comments are truncated due to high volume. (⌘F to expand all)Change truncation settings
9Carl_Shulman4yundefined
5EricHerboso4yundefined
2JBeshir4yundefined
7EricHerboso4yundefined
5Telofy4yundefined
4JBeshir4yundefined
1Jeff_Kaufman4yundefined
1Raemon4yundefined
9Brian_Tomasik4yundefined
1Raemon4yundefined
1Brian_Tomasik4yundefined
5erikaalonso4yundefined
4Brian_Tomasik4yundefined
8RyanCarey4yundefined
3Evan_Gaensbauer4yundefined
8Ben_Todd4yundefined
4jsteinhardt4yundefined
3Daniel_Dewey4yundefined
5Brian_Tomasik4yundefined
5Peter_Hurford4yundefined
4jsteinhardt4yundefined
5Daniel_Dewey4yundefined
0RyanCarey4yundefined
0Daniel_Dewey4yundefined
9kbog4yundefined
0Richard_Batty4yundefined
1kbog4yundefined
5atucker4yundefined
2TruePath4yundefined
2JBeshir4yundefined
1Telofy4yundefined
3Fluttershy4yundefined
1Telofy4yundefined
1Linch4yundefined
1Telofy4yundefined
2Peter_Hurford4yundefined
0Tom_Ash4yundefined
0John_Maxwell_IV4yundefined
9Peter_Hurford4yundefined
7JBeshir4yundefined
5DavidNash4yundefined
7Ben_West4yundefined
1JBeshir4yundefined
5JBeshir4yundefined
2Ben_West4yundefined
2Peter_Hurford4yundefined
2atucker4yundefined
1kbog4yundefined
1atucker4yundefined
1kbog4yundefined
2JBeshir4yundefined
-3Gleb_T4yundefined