We should put all possible changes/reforms in a big list, that everyone can upvote/downvote, agree disagree.
EA is governed but a set of core EAs, so if you want change, I suggest that giving them less to read and a strong signal of community consensus is good.
The top-level comments should be a short clear explanation of a possible change. If you want to comment on a change, do it as a reply to the top level comment
This other post gives a set of reforms, but they are a in a big long list at the bottom. Instead we can have a list that changes by our opinions! https://forum.effectivealtruism.org/posts/54vAiSFkYszTWWWv4/doing-ea-better-1
Note that I do not agree with all comments I post here.
Making EA appeal to a wider range of moral views
EA is theoretically compatible with a wide range of moral views, but our own rhetoric often conflates EA with utilitarianism. Right now, if you hold moral views other than utilitarianism (including variants of utilitarianism such as negative utilitarianism), you often have to do your own homework as to what those views imply you should do to achieve the greatest good. Therefore, we should spend more effort making EA appeal to a wider range of moral views besides utilitarianism.
What this could entail:
I actually think EA is inherently utilitarian, and a lot of the value it provides is allowing utilitarias to have a conversation among ourselves without having to argue the basic points of utilitarianism with every other moral view. For example, if a person is a nativist (prioritizing the well being of their own country-people), then they definitionally aren't an EA. I don't want EA to appeal to them, because I don't want every conversation to be slowed down by having to argue with them, or at least find another way to filter them out. EA is supposed to be the mechanism to filter the nativists out of the conversation.