This is a special post for quick takes by WillPearson. Only they can create top-level comments. Comments here also appear on the Quick Takes page and All Posts page.
So I've been trying to think of ways to improve the software landscape. If we do this it might make traditional software more aligned with human values and it's models for building more advanced systems too.
One piece I've been looking at is software licensing.
Instead of traditional open source, have an easy to get license for a version of software, based on a cryptographic identity. This could make it less frictional to be a bad actor.
This license is checked on startup that it matches the version of the software running (git sha stored somewhere). If it doesn’t the software fails to start. It can also be used by clients and servers to identify each other but does not have to marry up one to one with a person’s identity.
The license is acquired as easily as a let’s encrypt style certificate, but the identity has to part of the reputation system (which might require a fee).
The software might require a license from one of many reputation monitoring systems. So that no monitoring system becomes a single point of failure.
Edit: effective altruism might decide to fund awards for work of software ecosystem engineering with non software engineers as the judges to bring this digital infrastructure to the publics consciousness and incentivise making it understandable as well
I had an idea for a new concept in alignment that might allow nuanced and human like goals (if it can be fully developed).
Has anyone explored using neural clusters found by mechanistic interpretability as part of a goal system?
So that you would look for clusters for certain things e.g. happiness or autonomy and have that neural clusters in the goal system. If the system learned over time it could refine that concept.
This was inspired by how human goals seem to have concepts that change over time in them.
I've got an idea for a business that could help biosecurity by helping stop accidental leaks of data to people that shouldn't have it. I'm thinking about proving the idea with personal identifiable information. Looking for feedback and collaborators.
How should important ideas around topics like AI and biorisk be shared? Is there a best practice, or government departments that specialise in handling that?
So I've been trying to think of ways to improve the software landscape. If we do this it might make traditional software more aligned with human values and it's models for building more advanced systems too.
One piece I've been looking at is software licensing.
Instead of traditional open source, have an easy to get license for a version of software, based on a cryptographic identity. This could make it less frictional to be a bad actor.
This license is checked on startup that it matches the version of the software running (git sha stored somewhere). If it doesn’t the software fails to start. It can also be used by clients and servers to identify each other but does not have to marry up one to one with a person’s identity.
The license is acquired as easily as a let’s encrypt style certificate, but the identity has to part of the reputation system (which might require a fee).
The software might require a license from one of many reputation monitoring systems. So that no monitoring system becomes a single point of failure.
Edit: effective altruism might decide to fund awards for work of software ecosystem engineering with non software engineers as the judges to bring this digital infrastructure to the publics consciousness and incentivise making it understandable as well
I had an idea for a new concept in alignment that might allow nuanced and human like goals (if it can be fully developed).
Has anyone explored using neural clusters found by mechanistic interpretability as part of a goal system?
So that you would look for clusters for certain things e.g. happiness or autonomy and have that neural clusters in the goal system. If the system learned over time it could refine that concept.
This was inspired by how human goals seem to have concepts that change over time in them.
I've got an idea for a business that could help biosecurity by helping stop accidental leaks of data to people that shouldn't have it. I'm thinking about proving the idea with personal identifiable information. Looking for feedback and collaborators.
How should important ideas around topics like AI and biorisk be shared? Is there a best practice, or government departments that specialise in handling that?
My blog might be of interest to people