BSc Econometrics and Operations Research (focus: econometrics)
MSc Systems Biology (focus: evolutionary game theory for adaptive cancer treatment)
MSc Econometrics and Operations Research (focus: operations research, and math. econ. thesis regarding social influence and change)
Premaster Executive Master in Actuarial Science
Premaster and master Amsterdam Master in Actuarial Science
Teaser essay bundle "Ideas to Secure Our Future":
Hi Joe,
I much appreciated your post on deep atheism, and will still finish that. I also found this above post, and I thought I could contribute to the understanding at some points (as a Thesean myself).
Best,
Indra
“So holding lifespans fixed, a greater capacity for synchronic welfare does entail a greater capacity for diachronic welfare.” I’m missing here a discussion of adaptation, e.g. I might really like my first donut, but with more donuts my welfare capacity from another donut rapidly declines. The rate of this declining might differ across species. As such momentary peaks might be higher in one species while less rate of decline and less variance in another species inclunes larger diachronic welfare, despite lower synchronic welfare, at times, or, sustainably.
In addition, we might also want to use - and take in account - our abilities to look ahead. Suppose for example a worthwhile task that requires two people to engage in it. The first person to engage in it gains zero marginal returns, while the latter gets everything (all of the returns as marginal returns). The first person might however predict the second person's behavior and based on the expectation that results engage with this task. By contrast, chimpanzees are not able to do this; You would never see two of them cooperate to together e.g. carry a log (research by Joseph Henrich).
Producing reasoning transparancy would I think yield echo-chamber-reduction effects, and also inform the powerful person practicing it how to (to them, for starters) weigh pros and cons of transferring power. Moreover, without it, I don't see using reason and evidence to do the most good practiced, nor with that a license to be a powerful EA, as opposed to simply powerful. And if EA membership would instead just be about trying to do the most good, that would include all of humanity minus some deviants.
I appreciate your point that people that donate are under no obligation. As such an advisory (instead of instructing) role to them seems fitting. On the other hand, the intellectual EA community should however also have the freedom to: not take on certain money, or not take on certain money coupled to certain actions, or disassociate with people, e.g. when this otherwise puts the community's (intellectual) integrity at risk, e.g. their reasoning transparancy. (And even chosen intransparancy one can be transparant about at a higher level.) In that, also being that much of EA charity work is research-based, an analogy to the scientific community, where such potent integrity risk is also tantamount, seems quite fitting.
All in all there should I think be some balance in the democratic power at both ends, including on the burden of proof, instead of this being fully one-sided. Take in FTX maybe as another (historical) example. And ideally both sides are practicing (reasoning transparancy and are) getting better in being informed by reason and evidence to do good better. Potentially this identifies (and resolves?) some (but not all) cruxes, and fleshes out new ones, while also responding to some of your encouragements, to move the conversation (or reasoning transparancy) forward?