Hide table of contents

Do experts agree on what labs should do to make AI safer? Schuett et al's (2023) survey ("Towards best practices in AGI safety and governance: A survey of expert opinion") finds a broad consensus.

We organized two evaluations of this paper (link: summary and ratings, click through to the evaluations). Both evaluators rated the paper highly & found this work valuable & meaningful for policy & discussions of AI safety. They also highlighted important limitations,  including sampling bias, classification of practices, interpretation of results, and abstract agreement vs. real-world implementation and gave suggestions for improvements and future work.

 

For example

Sampling/statement selection bias concerns:

Evaluator 1: "... whether the sample of respondents is disproportionately drawn from individuals already aligned with AGI safety priorities"

Evaluator 2: ..."where statements selected from labs practices are agreed on by labs themselves"

 

Abstract agreement vs. real-world implementation:

Evaluator 1: "items ~capture agreement in principle, rather than ... [the] grappling with... tradeoffs ... real-world governance inevitably entails."

Evaluator 2: "agreement on practices could indicate a host of different things"; these caveats should should be incorporated more in the stated results.

 

The evaluation summary highlights other issues that we (as evaluator managers) believe merit further evaluation. We also point to a recent work that is related and complementary.  

The Unjournal has published 37 evaluation packages (and 20 are in progress), mainly targeting impactful work in quantitative social science and economics across a range of outcomes and cause areas. See our output at https://unjournal.pubpub.org. See the list of over 200 papers with potential for impact (which we have/or are currently considering or evaluating) here

9

0
0

Reactions

0
0
Comments
No comments on this post yet.
Be the first to respond.
Curated and popular this week
Relevant opportunities