Oh, it sounds like you might be confused about the context I'm talking about this occurring in, and I'm not sure that explaining it more fully is on-topic enough for this post. I'm going to leave this thread here for now to not detract from the main conversation. But I'll consider making a separate post about this and welcome feedback there.
I really appreciate you sharing your perspective on this. I think these are extremely hard calls, as evidenced by the polarity of the discussion on this post, and to some extent it feels like a lose-lose situation. I don't think these decisions should be made in a vacuum and want other people's input, which is one reason I'm flagging how this affects my work and the larger involvement funnels in EA.
Thanks for spelling this out.
I think to give some color to how this affects my work in particular (speaking strictly for myself as I haven't discussed this with others on my team):
One of our organization priorities is ensuring we are creating a welcoming and hopefully safe community for people to do good better, regardless of people's identities. A large part of our work is familiarizing people with and connecting them to organizations and resources, including ones that aren't explicitly EA-branded. We are often one of their first touch points within EA and its niches, including forecasting. Many factors shape whether people decide to continue and deepen their involvement, including how positive they find these early touch points. When we're routing people toward organizations and individuals, we know that their perception of our recommendations in turn affects their perception of us and of EA as a whole.
Good, smart, ambitious people usually have several options for professional communities to spend their time within. EA and its subcommunities are just one option and an off-putting experience can mean losing people for good.
With this in mind, I will feel much more reluctant to direct community members to Manifold in particular and (EA-adjacent) forecasting spaces more broadly, especially if the community member is an underrepresented group in EA. I think Manifold brings a lot of value, but I can't in good conscience recommend they plug into communities I believe most people I am advising would find notably morally off-putting.
This is of course a subjective judgement call, I understand there are strong counterarguments here, and what repels one person also attracts another. But I hope this gives a greater sense of the considerations/trade-offs I (and probably many others) will have to spend time thinking about and reaching decisions around as a result of Manifest.
Austin, I'm gathering there might be a significant and cruxy divergence in how you conceptualize Manifold's position in and influence on the EA community and how others in the community conceptualize this. Some of the core disagreements discussed here are relevant regardless, but it might help clarify the conversation if you describe your perspective on this.
Commenting just to encourage you to make this its own post. I haven't seen a (recent) standalone post about this topic, it seems important, and though I imagine many people are following this comment section it also seems easy for this discussion to get lost and for people with relevant opinions to miss it/not engage because it's off-topic.
You might be interested in the information security sphere, which some in the EA community focus on, especially in the context of safe AI. This 80,000 Hours podcast from 2022 is a good overview: https://80000hours.org/podcast/episodes/nova-dassarma-information-security-and-ai-systems/
The application deadline has been extended and now closes on July 28 at 11:59 pm ET.