Dr. Ellen Quigley, Special Advisor to the CFO of the University of Cambridge who previously has given talks at EA Finance and the Stanford Existential Risk Initiative, recently was featured on the Bogleheads podcast. The late Jack Bogle was the founder of Vanguard, the multi-trillion dollar money management firm. Vanguard was the first firm to provide a low cost passive investing index fund that tracked the stock market. The mutual fund industry was dead set against it when it began, because they feared it would disrupt their profits through expensive fees. That said, there had been a small contingent of luminaries, such as Economics Nobel Prize winner Paul Samuelson and Warren Buffett, who long contended that for the everyday investor, with relatively little or no information about the markets, index investing that tracked the market average were by far the best way to invest in securities. In the decades that have followed since Bogle started Vanguard, research has consistently showed that index investing beats the vast majority of other professional investment strategies in the long run. What is interesting in the work of Quigley and her colleagues is that they show that not only is index investing profitable, it is also better at incentivizing socially responsible corporate behavior than traditional ESG and SR investing.

Advocates of index investing organized themselves as "Bogleheads" in honor of Jack Bogle. Before his death, Bogle started the Bogle Center, which promotes index investing and financial literacy.  He handpicked its current leader, Rick Ferri, who hosts the Bogleheads podcast. I encourage everyone to catch Ellen's appearance on the Bogleheads podcast to learn about index investing and how it is more effective than traditional forms of socially responsible investing.

Also, for EAers who are looking for ways to passively increase their wealth to increase the amount they can donate, I encourage you to learn more about index investing. Some places to start include The Bogleheads' Guide to Investing, The Little Book of Common Sense Investing, and the Bogleheads Forum. The community of Bogleheads is a relatively wealthier community primarily organized around the idea that index investing produces a good financial return. If EAers became more engaged with that community, and helped make the ethical ethos of index investing become more prevalent within the community, then EAers may be able to convince them to direct more of their financial gains to EA causes minimizing suffering in the world. 

10

0
0

Reactions

0
0
Comments


No comments on this post yet.
Be the first to respond.
Curated and popular this week
 ·  · 5m read
 · 
The AI safety community has grown rapidly since the ChatGPT wake-up call, but available funding doesn’t seem to have kept pace. However, there’s a more recent dynamic that’s created even better funding opportunities, which I witnessed as a recommender in the most recent SFF grant round.[1]   Most philanthropic (vs. government or industry) AI safety funding (>50%) comes from one source: Good Ventures. But they’ve recently stopped funding several categories of work (my own categories, not theirs): * Many Republican-leaning think tanks, such as the Foundation for American Innovation. * “Post-alignment” causes such as digital sentience or regulation of explosive growth. * The rationality community, including LessWrong, Lightcone, SPARC, CFAR, MIRI. * High school outreach, such as Non-trivial. In addition, they are currently not funding (or not fully funding): * Many non-US think tanks, who don’t want to appear influenced by an American organisation (there’s now probably more than 20 of these). * They do fund technical safety non-profits like FAR AI, though they’re probably underfunding this area, in part due to difficulty hiring for this area the last few years (though they’ve hired recently). * Political campaigns, since foundations can’t contribute to them. * Organisations they’ve decided are below their funding bar for whatever reason (e.g. most agent foundations work). OP is not infallible so some of these might still be worth funding. * Nuclear security, since it’s on average less cost-effective than direct AI funding, so isn’t one of the official cause areas (though I wouldn’t be surprised if there were some good opportunities there). This means many of the organisations in these categories have only been able to access a a minority of the available philanthropic capital (in recent history, I’d guess ~25%). In the recent SFF grant round, I estimate they faced a funding bar 1.5 to 3 times higher. This creates a lot of opportunities for other donors
Nikola
 ·  · 1m read
 · 
 ·  · 1m read
 ·