Compute accounting technical details |
---|
A detailed method for both “ground truth” as well as good approximations for the total compute used in training and inference is required for meaningful compute-based controls. Here is an example of how the “ground truth” could be tallied at a technical level. |
Definitions: |
Compute causal graph: For a given output O of an AI model, there is a set of digital computations for which changing the result of that computation could potentially change O. (This should be conservatively assumed, i.e. there should be a clear reason to believe that a computation is independent of a precursor that both occurs earlier in time and has a physical potential causal path of effect.) This includes computation done by the AI model during inference, as well as computations that went into input, data preparation, and training of the model. Because any of these may itself be output from an AI model, this is computed recursively, cut off where a human has provided a significant change to the input. |
Training Compute: The total compute, in FLOP or other units, entailed by the compute causal graph of a neural network (including data preparation, training, and fine-tuning, and any other computations.) |
Output Compute: The total compute in the compute causal graph of a given AI output, including all neural networks (and including their Training Compute) and other computations going into that output. |
Inference Compute Rate: In a series of outputs, the rate of change (in FLOP/s or other units) of Output Compute between outputs, i.e. the compute used to produce the next output, divided by the timed interval between the outputs. |
Examples and approximations: |
|
Implementation Example: Here is one example of how a gate closure could work, given a limit of 1027 FLOP for training and 1020 FLOP/s for inference (running the AI): |
---|
1. Pause: For reasons of national security, the US Executive branch asks all companies based in the US, doing business in the US, or using chips manufactured in the US, to cease and desist from any new AI training runs that might exceed the 1027 FLOP Training Compute limit. The US should commence discussions with other countries hosting AI development, strongly encouraging them to take similar steps and indicating that the US pause may be lifted should they choose not to comply.
2. US oversight and licensing: By executive order or action of an existing regulatory agency, the US requires that within (say) one year:
3. International oversight:
4. International verification and enforcement:
|
Details for a strict AGI liability regime |
---|
|
A tiered approach to AGI safety & security standards | |||
---|---|---|---|
Risk Tier | Trigger(s) | Requirements for training | Requirement for deployment |
RT-0 | AI weak in autonomy, generality, and intelligence | none | none |
RT-1 | AI strong in one of autonomy, generality, and intelligence | none | Based on risk and use, potentially safety cases approved by national authorities wherever the model can be used |
RT-2 | AI strong in two of autonomy, generality, and intelligence | Registration with national authority with jurisdiction over the developer | Safety case bounding risk of major harm below authorized levels plus independent safety audits (including black-box and white-box redteaming) approved by national authorities wherever the model can be used |
RT-3 | AGI strong in autonomy, generality, and intelligence | Pre-approval of safety and security plan by national authority with jurisdiction over the developer | Safety case guaranteeing bounded risk of major harm below authorized levels as well as required specifications, including cybersecurity, controllability, a non-removable killswitch, alignment with human values, and robustness to malicious use. |
RT-4 | Any model that also exceeds either 1027 FLOP Training or 1020 FLOP/s Inference | Prohibited pending international agreed lift of compute cap | Prohibited pending international agreed lift of compute cap |