I finally finished my series of 5 essays on my life philosophy (Valuism)! In a nutshell, Valuism suggests that you work to figure out what you intrinsically value, and then that you try to use effective methods to create more of what you intrinsically value.
While simple and straightforward at first glance, few people seem to approach life this way, and Valuism ends up with a number of surprising implications and, I think, provides a perspective that can help shed light on a number of different domains.
Interestingly, Effective Altruism is implied by Valuism plus a specific set of strong intrinsic values that most Effective Altruists have (reducing suffering + truth).
Here is the sequence of essays, if you feel like checking them out:
Part 1: Doing what you value as a life philosophy – an introduction to Valuism
Part 2: What to do when your values conflict?
Part 3: Should Effective Altruists be Valuists instead of utilitarians?
Part 4: What would a robot value? An analogy for human values
Part 5: Valuism and X: how Valuism sheds light on other domains
A big shoutout goes to Amber Dawn Ace who wrote these essays with me.
Preference utilitarianism and valuism don't have much in common.
Preference utilitarianism: maximize the interests/preferences of all beings impartially.
First, preferences and intrinsic values are not the same thing. For instance, you may have a preference to eat Cheetos over eating nachos, but that doesn't mean you intrinsically value eating Cheetos or that eating Cheetos necessarily gets you more of what you intrinsically value than eating nachos will. Human choice is driven by a lot of factors other than just intrinsic values (though intrinsic values play a role).
Second, preference utilitarianism is not about your own preferences, it's about the preferences of all beings impartially.