-
Notifications
You must be signed in to change notification settings - Fork 1
/
aises_8_7
310 lines (304 loc) · 20.7 KB
/
aises_8_7
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
<h1 id="sec:compute-gov">8.7 Compute Governance</h1>
<p>A common shorthand for computational resources or computing power
used for AI is <em>compute</em>. In this section, we will discuss how
compute governance enables AI governance. First, we will examine how
since compute is indispensable for AI development, governing it would
help us govern AIs. Then, we will examine the key properties of compute
that make it governable—–physicality, excludability, and
quantifiability. These features make it more feasible to track, monitor
and, if appropriate, restrict the development of potentially dangerous
AI systems, and more generally facilitate the enforcement of AI
governance measures. We also consider why governing compute is more
promising than governing other factors used in AI production.<p>
</p>
<h2
id="compute-is-indispensable-for-ai-development-and-deployment">8.7.1 Compute
Is Indispensable for AI Development and Deployment</h2>
<p>Compute enables the development of more capable AIs. In addition,
compute is necessary to deploy AIs. If we restrict someone’s access to
compute, they cannot develop or deploy any AIs. As a result, we can use
compute to govern AIs, determining how and when they are deployed.</p>
<p><strong>Hardware is necessary for AI systems.</strong> Like the
uranium in a nuclear weapon, compute is fundamental to running AIs. In
its simplest form, we can think of compute as a select group of
high-performance chips like GPUs and TPUs that are designed to run
modern AIs. These are often the latest chips, tailor-made for AI tasks
and found in large data centers. As AI technology changes, so too will
the hardware, adapting to new tech developments, regulations, and the
evolving requirements of AI applications.</p>
<p><strong>The metric FLOP/s is common across forms of compute.</strong>
To measure compute, we often use the metric <em>FLOP/s</em>, which is
the number of floating-point operations (such as addition or
multiplication) a computer can do in a second. When we talk about
“increasing” compute, we’re referring to using more processors, using
better processors, or allowing these processors to run for extended
periods, effectively increasing the number of floating-point operations
done in total. An analogous escalation of this is improving a nuclear
arsenal by adding more weapons to it, developing more dangerous weapons
like H-bombs, or creating bigger bombs by using more uranium.</p>
<p><strong>More compute allows for the development of more AI
capabilities.</strong> Compute plays a pivotal role in the evolution of
AI capabilities. More compute means that AI systems can be built with
more parameters and effectively utilize larger datasets. In the Artificial Intelligence Fundamentals chapter,
we looked at scaling laws, which show us that many AIs have their
performance increase reliably with an increase in model size and dataset
size. Richard Sutton’s “The Bitter Lesson” states that general methods
in AI that harness computation tend to be the most effective by a
significant margin <span class="citation"
data-cites="Sutton-bitter">[1]</span>. Having more compute means
training AI systems that are more capable and advanced, which means that
knowing how much compute an AI uses lets us approximate its
capabilities.<p>
Often, pushing the boundaries in AI development requires having vast
compute. AIs can require training on large supercomputers that cost
hundreds of millions or even billions of dollars. Moreover,
computational demands for these AI models are constantly intensifying,
with their compute requirements doubling roughly every six months. This
growth rate surpasses the 2.5-year doubling time we see for the
price-performance of AI chips <span class="citation"
data-cites="amodei2018ai">[2]</span>. Given this trend, it’s likely that
future AI models will demand even greater investment in computational
resources and, in turn, possess greater capabilities.</p>
<p><strong>More compute enables better results.</strong> Compute is not
only essential in training AI-—it is also necessary to run powerful AI
models effectively. Just as we rely on our brains to think and make
decisions even after we’ve learned, AI models need compute to process
information and execute tasks even after training. If developers have
access to more compute, they can run bigger models. Since bigger models
usually yield better results, having more compute can enable better
results.</p>
<p><strong>Large-scale compute isn’t a strict requirement for all future
AI applications.</strong> AI efficiency research aims to reduce compute
requirements while preserving performance by improving other factors
like algorithmic efficiency. If algorithms become so refined that
high-capability systems can train on less powerful devices, compute’s
significance for governance might diminish. Some existing systems like
AI-powered drug discovery tools do not require much compute, but it has
been demonstrated that they be can repurposed to create chemical weapons
<span class="citation" data-cites="Urbina2022DualUO">[3]</span>.<p>
Additionally, there’s continued interest towards creating efficient AI
models capable of running on everyday devices such as smartphones and
laptops. Though projections from current trends suggest it will be
decades before data center-bound systems like GPT-4 could train on a
basic GPU <span class="citation"
data-cites="epoch2023mltrends">[4]</span>, the shift towards greater
efficiency might speed up dramatically with just a few breakthroughs. If
AI models require less compute, especially to the point that they become
commonplace on consumer devices, regulating AI systems based on compute
access might not be the most effective approach.</p>
<h2 id="compute-is-physical-excludable-and-quantifiable">8.7.2 Compute Is
Physical, Excludable, and Quantifiable</h2>
<p>To produce AI, developers need three primary factors: data,
algorithms, and compute. In this section, we will explore why governing
compute appears to be a more promising avenue than governing the other
factors. A resource is governable when the entity with legitimate claims
to control it—–such as a government–—has the ability to control and
direct it. Compute is governable because</p>
<ol>
<li><p>It can be determined who has access to compute and how they
utilize it.</p></li>
<li><p>It is possible to establish and enforce specific rules about
compute.</p></li>
</ol>
<p>These are true because compute is <em>physical</em>,
<em>excludable</em>, and <em>quantifiable</em>. These characteristics
allow us to govern compute, making it a potential point of control in
the broader domain of AI governance. We will now consider each of these
in turn.</p>
<h3 id="compute-is-physical">Compute Is Physical</h3>
<p>The first key characteristic that makes compute governable is its
physicality. Compute is physical, unlike datasets, which are virtual, or
algorithms, which are intangible ideas. This makes compute rivalrous and
enables tracking and monitoring, both of which are crucial to
governance.</p>
<p><strong>Since compute is physical, it is rivalrous.</strong> Compute
is rivalrous: it cannot be used by multiple entities simultaneously.
This is unlike other factors in AI production, such as algorithms which
can be used by anyone who knows them or data which can be acquired from
the same source or even stolen or copied and used simultaneously
(although this may be difficult due to information security and the size
of training datasets). Because compute cannot be simultaneously accessed
by multiple users or easily duplicated, regulators can be confident that
when it is being used by an approved entity, it is not also being used
by someone else. This makes it easier to regulate and control the use of
compute. GPUs can’t be downloaded but instead must be fabricated,
purchased, and shipped.</p>
<div>
<figure>
<img
src="https://raw.githubusercontent.com/WilliamHodgkins/AISES/main/images/compute.png" style="width:70%"
alt="image" />
<p class="tb-caption">Figure 8.2: Advanced semiconductor manufacturing
tools (such as the ASML Twinscan NXE) are large, highly specialized machines. <span class="citation"
data-cites="asml">[5]</span></p>
</figure>
</div>
<p><strong>Since compute is physical, it is trackable.</strong> Compute
is trackable, from chip fabrication to its use in data centers. This is
because compute is tangible and often sizable: Figure 8.2 shows a
cutting-edge semiconductor tool used as compute that costs $200 million
and requires 40 freight containers, 20 trucks, and 3 cargo planes to
ship anywhere.<p>
Unlike uranium, which is difficult to procure but not impossible to
steal and transport small amounts of, acquiring large-scale compute
requires the investment of resources in a relatively complicated and
visible process. Stakeholders, whether semiconductor firms, regulatory
bodies, or other involved entities, can accurately evaluate and trace
the overall quantity of these assets. For instance, if we monitor the
sales and distribution of chips, we know who possesses which
computational capacities and their intended applications. The complete
supply chain, from the semiconductor origins to the extensive data
centers harboring vast AI computational power, can be monitored, which
means it can be governed. By contrast, data acquisition and algorithmic
improvements can be done discreetly: possession of these within a
computing infrastructure can be concealed more easily than the
possession of the infrastructure itself.</p>
<h3 id="compute-is-excludable">Compute Is Excludable</h3>
<p>The second key characteristic that makes compute governable is its
excludability. Something is excludable if it is feasible to stop others
from using it. Most privately produced goods like automobiles are
excludable whereas others, such as clean air or street lighting, are
difficult to exclude people from consuming even if a government or
company doesn’t want to let them use it. Compute is excludable because a
few entities, such as the US and the EU, can control crucial parts of
its supply chain. This means that these actors can monitor and prevent
others from using compute.</p>
<p><strong>The compute supply chain makes monitoring easier.</strong> In
2023, the vast majority of advanced AI chips globally are crafted by a
single firm, Taiwan Semiconductor Manufacturing Company (TSMC). These
chips are based on designs from a few major companies, such as Nvidia
and Google, and TSMC’s production processes rely on photolithography
machines from a similarly monopolistic industry led by ASML <span
class="citation" data-cites="arnold2022eto">[6]</span>. Entities such as
the US and EU can, therefore, regulate these companies to control the
supply of compute—if the supply chain dynamics do not change
dramatically over time <span class="citation"
data-cites="khan2021semiconductor">[7]</span>. This simplifies the
tracking of frontier AI chips and enforcing of regulatory guidelines;
it’s what made the US export ban of cutting-edge AI chips to China in
2022 feasible. This example illustrates that these chips can be
governed. By contrast, data can be purchased from anywhere or found
online, and algorithmic advances are not excludable either, especially
given the open science and collaborative norms in the AI community.</p>
<p><strong>Frequent chip replacements means governance is effective
quickly.</strong> The price performance of AI chips is increasing
exponentially. With new chips frequently making recent products
obsolete, compute becomes more excludable. Historical trends show that
GPUs double their price performance approximately every 2.5 years <span
class="citation" data-cites="epoch2023mltrends">[8]</span>. In
conjunction with the rapidly increasing demand for more compute, data
centers frequently refresh their chips and purchase vast quantities of
new compute regularly to retain competitiveness. This frequent chip
turnover offers a significant window for governance since regulations on
new chips will be relevant quickly.</p>
<h3 id="compute-is-quantifiable">Compute Is Quantifiable</h3>
<p>The third key characteristic that makes compute governable is its
quantifiability. Quantifiability refers to the ability to measure and
compare both the quantity and quality of resources. Metrics such as
FLOP/s serve as a yardstick for comparing computational capabilities
across different entities. If a developer has more chips of the same
type, we can accurately deduce that they have access to more compute,
which means we can use compute to set thresholds and monitor
compliance.</p>
<p><strong>Quantifiability facilitates clear threshold setting.</strong>
While chips and other forms of compute differ in many ways, they can all
be quantified in FLOP/s. This allows regulators to determine how
important it is to regulate a model that is being developed: models that
use large amounts of compute are likely more important to regulate.
Suppose a regulator aims to regulate new models that are large and
highly capable. A simple way to do this is to set a FLOP/s threshold,
above which more regulations, permissions, and scrutiny take effect. By
contrast, setting a threshold on dataset size is less meaningful:
quality of data varies enough that 25 GB of data could contain all the
text in Wikipedia or one high-definition photo album. Even worse,
algorithms are difficult to quantify at all.</p>
<p><strong>Quantifiability is key to monitoring compliance.</strong>
Beyond the creation of thresholds, quantifiability also helps us monitor
compliance. Given the physical nature and finite capacity of compute, we
can tell whether an actor has sufficient computational power from the
type and quantity of chips they possess. A regulator might require
organizations with at least 1000 chips at least as good as A100s to
submit themselves for additional auditing processes. A higher number of
chips directly correlates to more substantial computational
capabilities, unlike with algorithms where there is no comparable metric
and data for which metrics are much less precise. In addition to compute
being physical and so traceable, this enables the enforcement of rules
and thresholds.</p>
<h3 id="forms-of-compute-governance">Forms of Compute Governance</h3>
<p><strong>Compute governance can focus on individual chips or on clusters of such chips.</strong> As alluded to previously, GPUs used for training and running AI systems are typically housed in purpose-built data center facilities which are equipped to handle their high demands for power and cooling. These data centers usually also have high levels of physical security to protect this valuable hardware and the AI models that are run on it. One approach to monitoring compute which is imperfect but requires limited new technologies or regulation is to track these data center facilities. However, such an approach faces challenges in terms of identifying which facilities are housing chips used for training AI as opposed to other types of computing hardware, or understanding how many chips they house and what they are being used for. Other approaches rely more heavily on tracking the individual chips suitable for AI training and inference, for example via some form of registry of chip sales, or via ``on-chip mechanisms.''</p>
<p><strong>{Providers of cloud services can support compute governance.</strong> Data centers can be owned and operated by the companies that are developing AI, but more commonly AI developers rent access to computing hardware from providers of cloud computing services which own large pools of computing hardware rented out to third parties. These cloud providers could be required to support relevant authorities with compute governance. For example, they could implement know-your-customer (KYC) policies to ensure that AI developers subject to export controls or other sanctions are not able to use cloud providers' hardware to train powerful AI systems.</p>
<p><strong>On-chip mechanisms.</strong> On-chip governance mechanisms have been proposed as a way to make monitoring and verification schemes for compute more robust to attempts at circumventing them. These could take the form of secure modules implemented within the chip's hardware and firmware that make it easier for authorized parties such as regulators to detect the location of chips suitable for training powerful AI systems. Such features do not exist today, but could likely be implemented in forms that leverage existing secure hardware features and do not require re-designing chips or compromising users' privacy by monitoring what software is being run on the chips. More ambitious forms of compute governance might aim to provide information on whether chips are actually being used to train large AI systems. This could be done at the hardware level by monitoring whether GPUs are being interconnected into the large clusters of chips required for such purposes.</p>
<h3 id="conclusions-about-compute-governance">Conclusions About Compute
Governance</h3>
<p><strong>Compute governance is a promising route to AI
governance.</strong> Compute is necessary for the development and
deployment of AIs, as well as being well-suited to governance. Relative
to governing algorithms and datasets, the other factors used to produce
AIs, governing compute is promising because it is physical, excludable,
and quantifiable.<p>
These features are useful for national governments, as they make it more
feasible to monitor the development and deployment of potentially
dangerous AI systems. This is a precondition for effective enforcement
of AI regulations. Similarly, compute’s features can support
verification of compliance with national or international standards and
agreements. This makes it easier to implement regulatory regimes,
particularly at an international level.</p>
<p>Compute governance is currently made simpler by certain factors such
as the controllability of the supply chain or requirement of large-scale
compute for highly capable models. If the supply chain of hardware for
training and running AI systems became much less concentrated, this
would make it harder to enforce potential restrictions on access to
relevant hardware as a governance mechanism. Similarly, if the required
compute resources to train dangerous AI systems were within the means of
a small business or even an individual, this would dramatically increase
the challenge of monitoring relevant compute resources as part of AI
governance.<p>
While describing compute governance, we have touched upon various ideas
about how to use compute to ensure companies produce safe AIs,
governments create and enforce regulations, and international
cooperation can control the compute supply chain. In the rest of this
chapter, we further discuss governance tools aimed at corporate,
national, and international governance.</p>
<br>
<br>
<h3>References</h3>
<div id="refs" class="references csl-bib-body" data-entry-spacing="0"
role="list">
<div id="ref-Sutton-bitter" class="csl-entry" role="listitem">
<div class="csl-left-margin">[1] R.
Sutton, <span>“The bitter lesson.”</span> [Online]. Available: <a
href="http://www.incompleteideas.net/IncIdeas/BitterLesson.html">http://www.incompleteideas.net/IncIdeas/BitterLesson.html</a></div>
</div>
<div id="ref-amodei2018ai" class="csl-entry" role="listitem">
<div class="csl-left-margin">[2] D.
Amodei, D. Hernandez, G. Sastry, J. Clark, G. Brockman, and I.
Sutskever, <span>“AI and compute. OpenAI.”</span> 2018.</div>
</div>
<div id="ref-Urbina2022DualUO" class="csl-entry" role="listitem">
<div class="csl-left-margin">[3] F.
L. Urbina, F. Lentzos, C. Invernizzi, and S. Ekins, <span>“Dual use of
artificial-intelligence-powered drug discovery,”</span> <em>Nature
Machine Intelligence</em>, 2022.</div>
</div>
<div id="ref-epoch2023mltrends" class="csl-entry" role="listitem">
<div class="csl-left-margin">[4] Epoch, <span>“Key trends and figures in machine
learning.”</span> 2023. [Online]. Available: <a
href="https://epochai.org/trends">https://epochai.org/trends</a></div>
</div>
<div id="ref-asml" class="csl-entry" role="listitem">
<div class="csl-left-margin">[5] <span>“The $150m Machine Keeping Moore's Law Alive.”</span> 2021. Available: <a
href="https://www.wired.com/story/asml-extreme-ultraviolet-lithography-chips-moores-law/">Wired</a></div>
</div>
<div id="ref-arnold2022eto" class="csl-entry" role="listitem">
<div class="csl-left-margin">[6] Z.
Arnold, J. VerWey, J. Melot, and N. Signh, <span>“ETO supply chain
explorer.”</span> [Online]. Available: <a
href="https://cset.georgetown.edu/publication/eto-supply-chain-explorer/">https://cset.georgetown.edu/publication/eto-supply-chain-explorer/</a></div>
</div>
<div id="ref-khan2021semiconductor" class="csl-entry" role="listitem">
<div class="csl-left-margin">[7] S.
M. Khan, A. Mann, and D. Peterson, <span>“The semiconductor supply
chain: Assessing national competitiveness,”</span> <em>Center for
Security and Emerging Technology</em>, vol. 8, no. 8, 2021.</div>
</div>
</div>