I have written a new GovAI blog post - link here.
How should labs should share large AI models? I argue for a "structured access" approach, where outsiders interact with the model at arm's length. The aim is to both (a) prevent misuse, and (b) enable safety-relevant research on the model. The GPT-3 API is a good early example, but I think we can go even further. This could be a promising direction for AI governance.
I would be interested to hear people's thoughts :)
Good question. A few possible strategies:
(1) Make it really easy. Have accessible software tools out there, so labs don't have to build everything from scratch.
(2) Sponsor relevant technical research. I'm especially thinking of research falling under "AI security". E.g. how easy is model-stealing, given different forms of access?
(3) Have certain labs act as early adopters. They experiment with the best setup and set an example for other labs.
(4) More public advocacy in favour of structured access.
(5) Set up a conference track where there's a specific role for labs sharing large models in a structured way. The expectations of the content of the paper would be different, e.g. they don't need to have scientifically interesting findings already. The authors explain everything included, e.g. "we have model checkpoints corresponding to XYZ different points in the training run". Analogous to a paper that introduces a new dataset.