If you work for a frontier AI company, either because you think they care about saving the world or especially if you think that you will be the one to influence them, you are deluded. Wake up and quit.
If you care about protecting the world, you will quit, even though it will be hard to give up the money and the prestige and the hope that they would fix the problem. The actual path to reducing AI risk is not as glamorous or as clear at this point as following the instructions of a wealthy and well-organized corporation, but at least you will be going in the right direction.
The early 80k-style advice to work at an AI lab was mainly to make technical discoveries for safety that e.g. academia didn't have the resources for. When they were small, it also made some sense to try to influence the industry culture. Now, this advice is crazy-- there is no way 1 EA joining a 1000 person company with duties to their investors and locked in a death race is going to "influence" it. The influence goes entirely the other way. If you weren't frogboiled, you would never have selected this path for influence.
There's a lot more to say on this, but I think this is the crux. Your chance for positive marginal impact for AI Safety is not with the labs. If you work for the labs, you're probably just a henchman for a supervillain megaproject, and you can have some positive counterfactual impact right now by quitting. Don't sell out.
I agree there are some possible attitudes that society could have towards AI development which could put us in a much safer position.
I think that the degree of consensus you'd need for the position that you're outlining here is practically infeasible, absent some big shift in the basic dynamics. I think that the possible shifts which might get you there are roughly:
I think there's potentially something to each of these. But I think the GDM paper is (in expectation) actively helpful for 1 and probably 3, and doesn't move the needle much either way on 2.
(My own view is that 3 is the most likely route to succeed. There's some discussion of the pragmatics of this route in AI Tools for Existential Security or AI for AI Safety (both of which also discuss automation of safety research, which is another potential success route), and relevant background views on the big-picture strategic situation in the Choice Transition. But I also feel positive about people exploring routes 1 and 2.)