-
Notifications
You must be signed in to change notification settings - Fork 0
/
posts.xml
2301 lines (1940 loc) · 349 KB
/
posts.xml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
<?xml version="1.0" encoding="UTF-8" ?>
<!-- This is a WordPress eXtended RSS file generated by WordPress as an export of your site. -->
<!-- It contains information about your site's posts, pages, comments, categories, and other content. -->
<!-- You may use this file to transfer that content from one site to another. -->
<!-- This file is not intended to serve as a complete backup of your site. -->
<!-- To import this information into a WordPress site follow these steps: -->
<!-- 1. Log in to that site as an administrator. -->
<!-- 2. Go to Tools: Import in the WordPress admin panel. -->
<!-- 3. Install the "WordPress" importer from the list. -->
<!-- 4. Activate & Run Importer. -->
<!-- 5. Upload this file using the form provided on that page. -->
<!-- 6. You will first be asked to map the authors in this export file to users -->
<!-- on the site. For each author, you may choose to map to an -->
<!-- existing user on the site or to create a new user. -->
<!-- 7. WordPress will then import each of the posts, pages, comments, categories, etc. -->
<!-- contained in this file into your site. -->
<!-- generator="WordPress/4.9.10" created="2019-04-01 03:42" -->
<rss version="2.0"
xmlns:excerpt="http://wordpress.org/export/1.2/excerpt/"
xmlns:content="http://purl.org/rss/1.0/modules/content/"
xmlns:wfw="http://wellformedweb.org/CommentAPI/"
xmlns:dc="http://purl.org/dc/elements/1.1/"
xmlns:wp="http://wordpress.org/export/1.2/"
>
<channel>
<title>Bits and Atoms</title>
<link>https://bitsandatoms.co</link>
<description></description>
<pubDate>Mon, 01 Apr 2019 03:42:25 +0000</pubDate>
<language>en-US</language>
<wp:wxr_version>1.2</wp:wxr_version>
<wp:base_site_url>https://bitsandatoms.co</wp:base_site_url>
<wp:base_blog_url>https://bitsandatoms.co</wp:base_blog_url>
<wp:author><wp:author_id>1</wp:author_id><wp:author_login><![CDATA[[email protected]]]></wp:author_login><wp:author_email><![CDATA[[email protected]]]></wp:author_email><wp:author_display_name><![CDATA[[email protected]]]></wp:author_display_name><wp:author_first_name><![CDATA[Nik]]></wp:author_first_name><wp:author_last_name><![CDATA[Dawson]]></wp:author_last_name></wp:author>
<generator>https://wordpress.org/?v=4.9.10</generator>
<item>
<title>The Perils of Progress</title>
<link>https://bitsandatoms.co/the-perils-of-progress/</link>
<pubDate>Thu, 17 Aug 2017 21:40:05 +0000</pubDate>
<dc:creator><![CDATA[[email protected]]]></dc:creator>
<guid isPermaLink="false">https://bitsandatoms.co/?p=55</guid>
<description></description>
<content:encoded><![CDATA[<h4><span style="font-weight: 400;">On the implications of intelligent machines and the transition to the knowledge economy.</span></h4>
<h5>INTRODUCTION</h5>
<span style="font-weight: 400;">Knowledge has become currency in the global economy. The knowledge economy is elevating Information as the primary source of value, overtaking material resources from our Industrial past. This is because Information feeds and defines intelligent machines.[note]'Intelligent Machines' or 'Artificial Intelligence' refers to a non-organic autonomous entities that are able to sense and act upon an environment to achieve specific goals. Intelligent agents may also learn or use knowledge to achieve these goals, which are governed by algorithms that are made by people.[/note]</span><span style="font-weight: 400;"> And it’s these intelligent technologies that leverage human knowledge, which help us to achieve more. </span>
<span style="font-weight: 400;">So, as human knowledge is supplemented by intelligent machines, more humans are becoming a function of their ability to interact with intelligent machines. Therefore, equipping people with technical skills is of growing importance. </span>
<span style="font-weight: 400;">Particularly important are the technical skills that enable people to build, manage, and improve the software and hardware powering intelligent machines. This includes skills such as design engineering, software development, and robotics. But it equally extends to areas such as Data Analysis, Human-Computer Interaction, or other technical disciplines helping steer intelligent technologies. Ultimately, it’s technical skillsets such as these that enhance human knowledge and help unlock idle capacity.</span>
<span style="font-weight: 400;">However, these technical disciplines demand high levels of skill. They require training and preparation, often with a strong foundation in Science, Technology, Engineering and Mathematics (STEM). As the demand for high-skilled labour outpaces lower-skilled work,[note]Pew Research Center (2016), <a href="http://www.pewsocialtrends.org/2016/10/06/1-changes-in-the-american-workplace/"><i>The State of American Jobs</i></a><i>: The changing demand for job skills and preparation</i>.[/note]</span><span style="font-weight: 400;"> industry is unable to meet these demands because people aren’t upskilling fast enough. This is concerning because low and medium-skilled jobs are at the greatest risk of automation.[note]‘Low and medium-skilled labour’ refers to routine and rules-based physical and cognitive tasks. Furman, Jason (2016), <a href="https://obamawhitehouse.archives.gov/sites/default/files/page/files/20160707_cea_ai_furman.pdf"><i>Is This Time Different? The Opportunities and Challenges of Artificial Intelligence</i></a>, Council of Economic Advisers to the White House, New York University, pg. 5-6.[/note]</span>
<span style="font-weight: 400;">There are two central problems underlying these labour market dynamics:</span>
<ol>
<li style="font-weight: 400;"><span style="font-weight: 400;">The majority of the world’s labour force are low and medium-skilled workers, with a disproportionate amount of whom live in the Global South[note]‘Global South refers to developing countries, which are located primarily in the Southern Hemisphere. United Nations <a href="http://ssc.undp.org/content/dam/ssc/documents/exhibition_triangular/SSCExPoster1.pdf">South-South Cooperation</a> (2007).[/note]</span><span style="font-weight: 400;">; </span></li>
<li style="font-weight: 400;"><span style="font-weight: 400;">Not nearly enough is being done to upskill and prepare workers for the demands of the knowledge economy, particularly in the Global South.</span></li>
</ol>
<img class="aligncenter wp-image-82" src="https://bitsandatoms.co/wp-content/uploads/2017/08/Global-South.png" alt="" width="718" height="368" />
<p style="text-align: center;"><span style="font-weight: 400; color: #999999;">Map of the Global South in red and Global North in blue</span></p>
<p style="text-align: center;"><span style="font-weight: 400; color: #999999;">Source: Wikimedia Foundation</span></p>
<span style="font-weight: 400;">The implications of widespread skill shortages are significant. Income inequality will rise; social unrest will ensue; and entire populations could lose their opportunity to contribute. While this dystopian outlook is intentionally hyperbolic, there are signs that this is already happening.[note]McKinsey Global Institute (2016), <i>Poorer than their parents? Flat or Falling Incomes in Advanced Economies</i>.[/note]</span>
<blockquote>It doesn’t have to be this way. This doesn’t have to be our future.</blockquote>
<span style="font-weight: 400;">Through systematic education and training we can raise the hidden talents and untapped potential of current and future workers. However, it will require a different conception of how skills are developed and education is delivered. We’ll need a strategy that follows a demand-led approach, which engages industry as central partners in the design and delivery of skills education. A movement away from the tired education model of ‘credentialing’, and a reorientation towards in-demand skill acquisition. Governments and International Institutions are central to helping facilitate these solutions.</span>
<span style="font-weight: 400;">A core focus needs to be placed upon the development of greater technical skills. This will enable people to build, manage, and interact with intelligent machines in the workforce. Equally important are the social and interpersonal skills that are the basis of human cooperation and the ability to navigate quickly changing labour markets. These are the broad skillsets that are in demand now and in the future. More specifically, Governments and Educational Institutions would do well to emphasise and develop the following three main skill areas:</span>
<ul>
<li style="list-style-type: none;">
<ul>
<li style="font-weight: 400;"><b>Software Development</b><span style="font-weight: 400;">, by equipping people with the skills to build and manage software applications;</span></li>
<li style="font-weight: 400;"><b>Design Engineering</b><span style="font-weight: 400;">, by learning to design, improve, and build objects through disciplines such as Robotics, 3D printing, and Mechatronics;</span></li>
<li style="font-weight: 400;"><b>Work readiness<span style="font-weight: 400;">, by training and mentoring people to navigate modern labour markets, pursue in-demand education, or start their own businesses.</span></b></li>
</ul>
</li>
</ul>
This alone is not the answer, but building creative responses around these skill development areas should be part of the solution. Building an inclusive future takes work. To leverage intelligent machines that harness the full power of human creativity requires careful planning, technical skills, and a deliberate view of the future.
<h5>THIS IS A BIG DEAL</h5>
<span style="font-weight: 400;">As intelligent technologies intersect with more parts of our lives, it’s changing how we live and how we work, regardless of geography. It’s shifting the primary sources of value from material assets to knowledge. This transition is most obviously seen in the stock market. </span><a href="https://en.wikipedia.org/wiki/List_of_public_corporations_by_market_capitalization#1997"><span style="font-weight: 400;">In</span><span style="font-weight: 400;"> 1997</span></a><span style="font-weight: 400;">, technology companies made up 2 out of the top 10 companies by market capitalisation; fast forward 20 years, it’s jumped to 5 out of 10. </span>
<img class="aligncenter wp-image-83" src="https://bitsandatoms.co/wp-content/uploads/2017/08/largest-companies-by-market-cap-chart.jpg" alt="" width="425" height="438" />
<p style="text-align: center;"><span style="font-weight: 400; color: #999999;">Source: Visual Capitalist, August 2016</span></p>
<span style="font-weight: 400;">This is significant because economic value is an indication of social authority. It represents the growing role that technology plays in our lives. And it highlights the greater authority we grant technology to inform our social dynamics.</span>
<span style="font-weight: 400;">When we need to travel from one place to another, we seldom ask a friend or close passerby for detailed directions. Instead, we pull out our smart phones and ask Google Maps. For countless routine tasks, we’ve elevated intelligent machines as the higher source of authority over people. Google knows best.</span>
<blockquote><span style="font-weight: 400;">The ability of intelligent machines to complete cognitive tasks is revolutionary. Cognitive tasks have historically been reserved to human labour; now they’re routinely performed by machines at breakneck speeds.</span></blockquote>
<span style="font-weight: 400;">Amazon’s book recommendations is a familiar example of an intelligent machine. While browsing, Amazon recommends books based on information like your purchase history, search history, and user profile, with the goal of selling more books. Your recommendations adapt as new information is provided by you and others. The more information Amazon collects, the more accurate their recommendations become, the higher their book sales rise. The wise local bookstore owner doesn’t stand a chance.</span>
<span style="font-weight: 400;">All of this speaks to the higher value we’re placing on knowledge. We see this in the growing market value of technology firms, the rapid wage growth of software developers,[note]Burning Glass Technologies (2016), <i>Beyond Point and Click: The expanding demand for coding skills</i> pg. 3.[/note]</span><span style="font-weight: 400;"> and the average 10 hours per day[note]Ernst & Young (2016), <i>Digital Australia: State of the Nation 2015-16</i> pg. 13.[/note]</span><span style="font-weight: 400;"> we spend on internet connected devices. All of these indicators show that we’ve entered the ‘Information Age’. And information is transferred through applied technologies that feeds ‘knowledge’ in the knowledge economy.</span>
<h5>THE IMPORTANCES OF CONTRIBUTION</h5>
<span style="font-weight: 400;">Equipping people with technical skills is important for two main reasons:</span>
<ul>
<li><b>Personal contributions<span style="font-weight: 400;"> - This applies to both economic and social contributions. Acquiring the technical skills of a software developer places an individual in high demand. They’re able to attract a higher salary, which can help improve the well-being of themselves and their family. Less measurable, but no less significant, are the well-being and identity benefits that arise from contributing to important work. This isn’t to say that all important work is reserved to people with technical skills. They’re not. But as technology is increasingly applied to help solve our most pressing problems, like climate change, healthcare, and education, a strong technical skillset is often required to be part of these solutions. This requirement for technical skills to help solve hard problems will continue to increase.</span></b></li>
<li><strong>Collective contributions</strong><span style="font-weight: 400;"> - The degree of change instigated by technology should not be dismissed by the magnitude of what’s left to do. While Twitter may feel like a passing fad, it’s still global instant communication, for</span><span style="font-weight: 400;"> free</span><span style="font-weight: 400;">. That’s a significant social advancement. What’s even more significant is how small the group of people behind these changes are. They’re a tiny proportion of the world’s population, with highly technical skills. So, imagine what we could collectively achieve if we broadened the talent pool, even just a little. A rising tide lifts all boats.</span></li>
</ul>
<span style="font-weight: 400;">The point is, enabling people to meaningfully contribute in a future with more intelligent technologies is good for the individual and the collective.</span>
<h5>THE TIME IS RIPE</h5>
<span style="font-weight: 400;">All signs point to a future where intelligent technologies intersect with more parts of our lives. This greater role will largely be made possible by progress in Artificial Intelligence (AI).[note]Artificial Intelligence is the broad discipline of non-organic intelligent technologies, which includes subset disciplines such as Machine Learning, Natural Language Processing, and Computer Vision.[/note]</span><span style="font-weight: 400;"> We’ve seen false starts and similarly large claims</span> <span style="font-weight: 400;">about AI before,</span><span style="font-weight: 400;">[note]Chen, Frank (2016), <i>AI, Deep Learning, and Machine Learning: A Primer </i>[Video], Andreessen Horowitz.[/note]</span><span style="font-weight: 400;"> but we’re on the precipice of something big. </span>
<span style="font-weight: 400;">The breakneck advancements in AI are being brought about through a confluence of developments. The driving factors are:</span>
<ul>
<li><b>Greater computational power</b></li>
<li style="list-style-type: none;">
<ul>
<li style="list-style-type: none;">
<ul>
<li style="font-weight: 400;"><span style="font-weight: 400;">Moore’s Law has largely held constant since 1971, with CPU transistor counts doubling every two years. The same applies for quality adjusted microprocessor prices, memory capacity, sensors, and pixels in digital cameras.[note]The Economist (2016) <i>After Moore’s Law: Technology Quarterly</i> June Quarterly Edition.[/note]</span></li>
</ul>
</li>
</ul>
</li>
<li><b>Increased quantities and access to training data</b></li>
<li style="list-style-type: none;">
<ul>
<li style="list-style-type: none;">
<ul>
<li style="font-weight: 400;"><span style="font-weight: 400;">The proliferation of devices and applications have exploded, which has caused a 50% compound annual growth rate (CAGR) from 2010.[note]Purdy, Mark; & Daugherty, Paul (2016) <i>Why Artificial Intelligence is the Future of Growth</i> Accenture publications pg. 11.[/note]</span></li>
</ul>
</li>
</ul>
</li>
<li><b>Better algorithms</b></li>
<li style="list-style-type: none;">
<ul>
<li style="list-style-type: none;">
<ul>
<li style="font-weight: 400;"><span style="font-weight: 400;">The processes and techniques that construct algorithms have become more sophisticated. Algorithm methods such as convolutional, feedforward, and adversarial networks are improving large-scale data processing.</span></li>
</ul>
</li>
</ul>
</li>
<li><b>Broad investment</b></li>
<li style="list-style-type: none;">
<ul>
<li style="list-style-type: none;">
<ul>
<li style="font-weight: 400;"><span style="font-weight: 400;">Growth in private AI investment has increased by 6X between 2011 to the beginning of 2016 alone.[note]CB Insights (2016) <i>Artificial Intelligence Explodes: New Deal Activity Record for AI Startups</i> [Blog][/note]</span></li>
</ul>
</li>
</ul>
</li>
<li><b>Open source frameworks and libraries</b></li>
<li style="list-style-type: none;">
<ul>
<li style="font-weight: 400;"><span style="font-weight: 400;">The open sourcing of AI tools such as Google’s Tensor Flow, Facebook’s Torch, and Amazon’s DSSTNE have made it more accessible than ever for companies and individuals to apply their data to some of the most powerful AI technologies.</span></li>
</ul>
</li>
</ul>
<img class="aligncenter wp-image-85" src="https://bitsandatoms.co/wp-content/uploads/2017/08/AlphaGo-1024x683.jpg" alt="" width="586" height="391" />
<p style="text-align: center;"><span style="font-weight: 400; color: #999999;">AlphaGo plays Lee Sedol in 2016</span></p>
<p style="text-align: center;"><span style="font-weight: 400; color: #999999;">Source: New Scientist, 4 January 2017</span></p>
<span style="font-weight: 400;">These advancements are driving the progress of intelligent machines. As they continue to develop, they will increasingly intersect with more parts of our lives, including work.</span>
<h5>1984 OR UTOPIA?</h5>
<span style="font-weight: 400;">Too often, the impact of AI on the future of work is debated from two extreme ends of the spectrum. On one side, the Dystopians anticipate humanity’s obsolescence; on the other, the Utopians predict Heaven on Earth. </span>
<blockquote><span style="font-weight: 400;">Both make headlines. Both Brave New Worlds with more AI. Neither are particularly helpful.</span></blockquote>
<span style="font-weight: 400;">While these predictions are still possibilities that shouldn’t be ruled out, they offer little explanation with how we get there. More helpful is a view of history, rooted in the present. </span>
<span style="font-weight: 400;">Let’s consider the effects of the Industrial Revolution on employment. The labour market has historically been split into three main sectors: agriculture, industry, and services. Until around 1800, the majority of Americans were employed in the agricultural sector. Just one century later, the employment rate in agriculture had halved. Significant portions of agricultural labour were automated, people moved to urban centres for industry-based jobs, and as education improved, a growing populace took up service professions. By 1900, the three sectors were almost evenly distributed.[note]Gallman, Robert E.; & Weiss, Thomas J. (1969) "The Service Industries in the Nineteenth Century." In Production and Productivity in the Service Industries, ed. Victor R. Fuchs, 287-352. New York: Columbia University Press.[/note]</span><span style="font-weight: 400;"> This was a seismic shift from the fields and herds, to the cities and factories.</span>
<span style="font-weight: 400;">And the pace of change continued to accelerate. By 2010, only 2 percent of Americans were employed in agriculture, 20 percent in industry, and the 78 percent majority assumed service-based employment.[note]1950–2010: <a href="http://www.bea.gov/">Bureau of Economic Analysis</a>, National Income and Product Accounts.[/note]</span>
<img class="aligncenter wp-image-86" src="https://bitsandatoms.co/wp-content/uploads/2017/08/Distribution-of-labour-force-by-sector-1024x633.png" alt="" width="656" height="405" />
<p style="text-align: center;">Source: Lewis Johnston, MinnPost[note]<span style="font-weight: 400;">Johnston, Lewis (2012) </span><i><span style="font-weight: 400;">History lessons: Understanding the decline in manufacturing</span></i><span style="font-weight: 400;"> [Blog], MinnPost.[/note]</span></p>
<span style="font-weight: 400;">So, how do the changes of the Industrial Revolution help inform the transition we’re experiencing with the Information Revolution?</span>
<span style="font-weight: 400;">There are similarities and differences. The similarities revolve around the human response and public perception to these changes; the differences concern the fundamental technologies driving these changes and their effects.</span>
<h5>THE PROMETHEAN MYTH</h5>
<span style="font-weight: 400;">The awakening of the Industrial Revolution evoked fears of mass-unemployment through mechanical automation. People worried about the obsolescence of humanity and the Luddites screamed about the unravelling of social order. </span>
<span style="font-weight: 400;">This response has been a constant throughout history. It goes as far back to Greek Mythology with the ‘</span><a href="http://www.theoi.com/Titan/TitanPrometheus.html"><span style="font-weight: 400;">Promethean Myth</span></a><span style="font-weight: 400;">’: Man acquires fire from the Gods; fire becomes the source of Man’s pain and suffering. We hear variations of the same narrative today: humans build and deploy robots; robots take our jobs and our purpose. </span>
<blockquote><span style="font-weight: 400;">Yet, the fears of the Promethean Myth have been consistently unfounded. To take the position that ‘this time is different’ is a complete abnegation of economic history. </span></blockquote>
<span style="font-weight: 400;">Of course new jobs will be created. Ideas, industries, and fields will be conceived that we haven’t even imagined. The economic process of ‘Creative Destruction’ will upend markets, replacing jobs from the ‘old’ economy and ushering in the ‘new’. </span>
<span style="font-weight: 400;">A 2011 study by </span><a href="http://owni.fr/files/2011/03/internet_impact_rapport_mcKinseycompany.pdf"><span style="font-weight: 400;">McKinsey</span></a><span style="font-weight: 400;"> found that the Internet had destroyed 500,000 jobs in France in the previous 15 years. However, over the same period, the Internet had created 1.2 million jobs. That’s a net employment addition of 700,000 and a rate of 2.4 jobs created for every job destroyed.[note]McKinsey Global Institute (2017) <i>Technology, Jobs, and the Future of Work</i>.[/note]</span>
<span style="font-weight: 400;">This has been a constant occurrence across generations. We’ve seen technology over the past 144 years create more jobs than it’s destroyed.[note]Deloitte (2015) <i>Technology and People: The great job-creating machine</i>.[/note]</span><span style="font-weight: 400;"> And it continues to do so, with the net rate of job creation increasing over the past two decades.[note]World Economic Forum (2016) <i>The Future of Jobs Report</i>.[/note]</span><span style="font-weight: 400;"> </span>
<span style="font-weight: 400;">This isn’t wishful thinking or wilful ignorance; it’s pragmatic reasoning centred around the history and potential of human creativity. </span>
<span style="font-weight: 400;">So what’s all the fuss about?</span>
<h5>IT'S ALL ABOUT SKILLS</h5>
<span style="font-weight: 400;">If we project that the trend of ‘job replacement’ by new technologies will continue, then the next step is to consider what </span><span style="font-weight: 400;">types of jobs</span><span style="font-weight: 400;"> will be demanded.</span>
<span style="font-weight: 400;">This is where we begin to approach unfamiliar territory.</span>
<span style="font-weight: 400;">The transition from the Agricultural Revolution to the Industrial Revolution saw entire populations move from farmlands to factories. This transition, however, was still predominantly a movement between low-skilled vocations.[note]Harari, Yuval N. (2016) <i>Homo Deus: A Brief History of Tomorrow</i> Chapter 9: ‘The Great Decoupling’.[/note]</span><span style="font-weight: 400;"> People shifted from the physical labour of toiling the earth to joining the assembly line of a steel mill. Very different work, but still low-skilled jobs to low-skilled jobs.</span>
<span style="font-weight: 400;">The fundamental difference with the transition to the Information Age is the greater demand for high-skilled labour. This demand is driven by the growing proliferation of intelligent machines. And the ability to interact with these intelligent machines requires strong technical skills. </span>
<span style="font-weight: 400;">The core problem is that people aren’t developing these skills at a sufficient pace.</span>
<span style="font-weight: 400;">More jobs today require higher levels of preparation and this rate is increasing. According to the US Department of Labor, American workers requiring above-average skills increased by 68% from 1980 to 2015. This is more than double the demand for workers with below-average skills, which only increased by 31% over the same period.[note]Pew Research Center (2016), <a href="http://www.pewsocialtrends.org/2016/10/06/1-changes-in-the-american-workplace/"><i>The State of American Jobs</i></a><i>: The changing demand for job skills and preparation</i>.[/note]</span>
<img class="aligncenter wp-image-87" src="https://bitsandatoms.co/wp-content/uploads/2017/08/Employment-growth-by-skill-level-1024x633.png" alt="" width="644" height="398" />
<p style="text-align: center;"><span style="font-weight: 400; color: #999999;">Note: based on employed US citizens aged 16 years and older. </span></p>
<p style="text-align: center;"><span style="color: #999999;"><span style="font-weight: 400;">Source: Pew Research Center[note]Ibid[/note]</span></span></p>
<span style="font-weight: 400;">This growth in demand for workers with higher levels of preparation is a function of the higher levels of skillsets demanded for these jobs. The modern economy is disproportionately demanding and valuing advanced cognitive skills.</span>
<span style="font-weight: 400;">Occupations requiring higher levels of social or analytical skills grew significantly between 1980 to 2015. Jobs with above-average social skills, like communications and people management, grew by 83%. Similarly, employment in jobs that require above-average analytical skills, such as non-routine cognitive and highly technical tasks, increased by 77%. This dwarfs the employment growth of labour with higher levels of physical skills, which only grew by 18% over the same period.[note]Ibid[/note]</span>
<img class="aligncenter wp-image-88" src="https://bitsandatoms.co/wp-content/uploads/2017/08/Stronger-employment-growth-for-higher-skills-1024x596.png" alt="" width="641" height="373" />
<p style="text-align: center;"><span style="font-weight: 400; color: #999999;">Note: based on employed US citizens aged 16 years and older.</span></p>
<p style="text-align: center;"><span style="color: #999999;">Source: Pew Research Center[note]Ibid[/note]</span></p>
<span style="font-weight: 400;">This growing demand for higher levels of preparation and skills is indicative of the transition to the knowledge economy. As physical labour is increasingly automated, the demand for cognitive labour grows, particularly non-routine cognitive labour. </span>
<span style="font-weight: 400;">Physical labour and routine manufacturing are being automated because it’s now more efficient for machines to complete these tasks. This is not the case for social and analytical skills aforementioned. It’s extremely difficult for intelligent machines to complete abstract social tasks and non-routine cognitive exercises. Machines can’t yet deal with the social idiosyncrasies of people management, or the analytical rigours of strategy development - at least as effectively as humans. There are certainly intelligent machines that supplement these areas. And it’s crucial for people in these roles to be able to interact with these intelligent machines. However, these machines are applied to what they do best: computation and data processing.</span>
<span style="font-weight: 400;">Therefore, the concern shouldn’t be that there won’t be jobs created on the foundations of AI; the concern should be that we won’t have enough people with the skills to do these jobs.</span>
<h5>SKILL SHORTAGES AND THEIR DISCONTENTS</h5>
<span style="font-weight: 400;">The implications of skill shortages result in unfavourable economic and social outcomes. The mismatch between the growth rates of rising high-skilled labour demand and sluggish high-skilled labour supply impact us all. It affects us economically, socially, and morally.</span>
<span style="font-weight: 400;">If only a small and shrinking proportion of the world’s population can fulfil these high-skilled jobs, it places downward pressure on everyone else. More people enter the pools of lower skilled work and wage rates decrease as more people slide down the skill curve. Meanwhile, wages in higher-skilled labour disproportionately rise. </span>
<span style="font-weight: 400;">This growing gap between income advancement and employment opportunities has widened over the past few decades. For instance, US college graduates in 1981 earned a wage premium of 48% over high school graduates. By 2005, the wage premium had risen to 97% - college graduates earn almost double that of high school graduates.[note]Autor, David (2014) <i>Skills, education, and the rise of earnings inequality among the ‘other 99 percent</i>, Science, Volume 344, Issue 6186, May 2014. See also: McKinsey Global Institute (2016), <i>Poorer than their parents? Flat or Falling Incomes in Advanced Economies</i>.[/note]</span><span style="font-weight: 400;"> This growing income inequality is reflected across the world, and felt most acutely in the Global South.</span>
<span style="font-weight: 400;">Disparities in wealth and earnings potential have been a constant in capitalist economies. However, automation and the subsequent rising demand for high skilled labour could accelerate this inequality. The White House noted in 2016 that the effects of automation through AI on labour markets will likely grow over the forthcoming decade.[note]Executive Office of the President (2016), <i>Artificial intelligence, automation, and the economy</i>.[/note]</span>
<span style="font-weight: 400;">This means that more jobs, or parts of jobs, will experience automation by intelligent machines. McKinsey Global Institute researched automation potential by examining over 2,000 work activities across 46 countries, which represents around 80% of the global workforce. Their research found that the proportion of occupations that could be </span><i><span style="font-weight: 400;">fully</span></i><span style="font-weight: 400;"> automated by demonstrated technologies is small - less than 5%. The potential for </span><i><span style="font-weight: 400;">partial</span></i><span style="font-weight: 400;"> automation, however, is much higher. Around 60% of all occupations have at least 30% of activities that are technically automatable, based on current technologies.[note]McKinsey Global Institute (2017) <i>Technology, Jobs, and the Future of Work</i>.[/note]</span><span style="font-weight: 400;"> Therefore, the majority of occupations will change, and more people will have to interact with intelligent technologies as part of their work.</span>
<p style="text-align: center;"><strong><span style="color: #808080;">Automation potential based on demonstrated technology of occupation titles in the US (cumulative)</span></strong></p>
<img class="aligncenter wp-image-89" src="https://bitsandatoms.co/wp-content/uploads/2017/08/Automation-potential.png" alt="" width="698" height="413" />
<p style="text-align: center;">Source:<span style="font-weight: 400;"> McKinsey Global Institute[note]McKinsey Global Institute (2017) </span><i><span style="font-weight: 400;">A Future that Works: Automation, Employment, and Productivity</span></i><span style="font-weight: 400;">, Executive Summary pg. 5.[/note]</span></p>
<blockquote><span style="font-weight: 400;">If people fail to sufficiently upskill, this will disproportionately favour the highly skilled. Conversely, it will disfavour the low and medium-skilled workers who are unprepared for the demands of the knowledge economy.</span></blockquote>
<span style="font-weight: 400;">Not only could the labour supply of similarly lower-skilled workers increase, placing downward pressure on wages, but lower skilled and cheaper labour is more likely to come under pressure from automation.</span>
<span style="font-weight: 400;">The Council of Economic Advisers to the White House ranked the probability of automation according to wages. They found that 83% of jobs making less than $20 per hour would come under pressure from automation. This is over twenty times more likely than jobs earning $40 per hour.[note]Furman, Jason (2016), <a href="https://obamawhitehouse.archives.gov/sites/default/files/page/files/20160707_cea_ai_furman.pdf"><i>Is This Time Different? The Opportunities and Challenges of Artificial Intelligence</i></a>, Council of Economic Advisers to the White House, New York University, pg. 4-5.[/note]</span>
<img class="aligncenter size-full wp-image-91" src="https://bitsandatoms.co/wp-content/uploads/2017/08/chart.png" alt="" width="446" height="506" />
<p style="text-align: center;"><span style="color: #999999;"><span style="font-weight: 400;">Source: Jason Furman[note]Ibid</span><span style="font-weight: 400;">[/note]</span></span></p>
<span style="font-weight: 400;">Regardless of whether these probabilities are exactly accurate, the magnitude of this variance is immense. Assuming that wages are correlated with skill-levels, these projections show that automation could cause a disproportionately large decline in the demand for less-skilled jobs, and a minimal decline in demand for high-skilled jobs. As a result, wage pressures rise and inequality gaps widen.</span>
<span style="font-weight: 400;">This has been playing out for decades. Wages as a share of GDP in advanced economies has dropped sharply since the 1970s.[note]McKinsey Global Institute (2016), <i>Poorer than their parents? Flat or Falling Incomes in Advanced Economies</i>.[/note]</span><span style="font-weight: 400;"> Yet, wages for the highest income earners over the same period have risen consistently.[note]Saez, Emmanuel (2015) <i>Striking it Richer: The Evolution of Top Incomes in the United States</i>.[/note]</span><span style="font-weight: 400;"> While it’s wrong to attribute this growing inequality as a pure function of technology, people have not acquired the necessary skills at a sufficient rate to meet these emerging labour demands.</span>
<span style="font-weight: 400;">All of this speaks to a systematic failure to adequately prepare people for the future of work.</span>
<h5>DEVELOPING COUNTRIES WILL BE HARDEST HIT</h5>
<span style="font-weight: 400;">As the majority of low and medium skilled workers live in developing countries, the developing countries are at the greatest risk.</span>
<span style="font-weight: 400;">Cheap labour, improved logistics, and internet connectivity enabled offshoring of manufacturing in advanced economies. This ‘deindustrialisation’ helped lift incomes and living standards in the developing countries that assumed this outsourced labour. It also guided the transition of advanced economies toward more service-based labour (as previously discussed). </span>
<blockquote><span style="font-weight: 400;">Therefore the fundamental concerns for developing economies are the same as advanced economies, they’re just more protracted. The central issue is still the rate of skill acquisition.</span></blockquote>
<span style="font-weight: 400;">At the current rate, the projections look bleak. A World Bank report found that two-thirds of all jobs in the developing world face significant automation.[note]World Bank Group (2016) <a href="http://documents.worldbank.org/curated/en/896971468194972881/pdf/102725-PUB-Replacement-PUBLIC.pdf"><i>Digital Dividends</i></a> pg. 23.[/note]</span><span style="font-weight: 400;"> Interestingly, the majority of these are likely to be middle-skilled, middle-paying occupations (for e.g. clerks, plant and machine operators). Low-skilled jobs are still at significant risk of displacement. However, the more immediate concern for low-skilled workers is where the medium-skilled workers look for their next job. </span>
<span style="font-weight: 400;">If they’re not equipped for the new demands of the knowledge economy, medium-skilled workers move down the skill curve and the low-skilled labour supply increases. This heightens the bargaining power of employers, wages are pressured down, and inequality widens. In some ways, we’re seeing this today with the movement towards the ‘on demand’ economy. Due to abundant supply of low and moderately skilled labour, employers dictate wages and only pay for discrete periods of work. Think: Uber or your favourite food delivery service.</span>
<span style="font-weight: 400;">The effect of growing displacement of medium-skilled labour is referred to as ‘employment polarisation’. This is where labour supply becomes concentrated at either ends of the skill spectrum. The main issue with this are the obstructions to upward social mobility.[note]Santos, Indhira (2016) <a href="http://blogs.worldbank.org/developmenttalk/labor-market-polarization-developing-countries-challenges-ahead">Labor market polarization in developing countries: challenges ahead</a> [Blog], World Bank Group.[/note]</span><span style="font-weight: 400;"> If employment polarisation worsens, there are fewer opportunities for people to climb the skill ladder, as the medium-skilled rung is weakened or transformed.</span>
<span style="font-weight: 400;">This process of turnover, accelerated by automation through intelligent machines, could lead to sustained periods of underemployment or unemployment.[note]Furman, Jason (2016), <a href="https://obamawhitehouse.archives.gov/sites/default/files/page/files/20160707_cea_ai_furman.pdf"><i>Is This Time Different? The Opportunities and Challenges of Artificial Intelligence</i></a>, Council of Economic Advisers to the White House, New York University, pg. 6.[/note]</span><span style="font-weight: 400;"> Not all workers will have the training or skills to fulfil the new jobs created by AI. Developing countries are particularly at risk, given higher numbers of low-skilled and medium skilled workers, fewer training opportunities, and less comprehensive safety nets. This has not historically been a recipe for peace.</span>
<span style="font-weight: 400;">The extent of labour displacement, however, will be determined by people’s abilities to upskill and prepare for the demands of the knowledge economy.</span>
<h5>THE EDUCATION BUBBLE</h5>
<span style="font-weight: 400;">The gulf between what’s happening and what needs to happen is intimidatingly large. Public Policy and Education Institutions have not kept apace with the advancements of AI and other technologies. Consequently, people are unprepared for the demands of the workplace, today and tomorrow.</span>
<span style="font-weight: 400;">In a 2013 study of youth and employers across nine countries, 40% of employers stated ‘lack of skills’ as the primary reason for entry-level vacancies.[note]McKinsey On Society (2013) <i>Education to Employment: Designing a System that Works,</i> pg. 18-21.[/note]</span><span style="font-weight: 400;"> Further to this, 60% of employers said graduates were not adequately prepared for the skill requirements of work. They noted particular gaps in technical, analytical, and communication skills. This reinforces the growing and unmet demand for higher-level skills previously discussed. </span>
<span style="font-weight: 400;">Yet, perceptions of ‘readiness’ vary between Education providers, employers, and youth. In this same research, 72% of Education providers claimed their graduates are adequately prepared. Whereas Employers and Youth stated rates of ‘readiness’ only 42% and 45%, respectively.</span>
<img class="aligncenter wp-image-92" src="https://bitsandatoms.co/wp-content/uploads/2017/08/graduate-readiness-1024x633.png" alt="" width="656" height="405" />
<p style="text-align: center;"><span style="color: #999999;"><span style="font-weight: 400;">Source: McKinsey On Society[note]McKinsey On Society (2013) </span><i><span style="font-weight: 400;">Education to Employment: Designing a System that Works,</span></i><span style="font-weight: 400;"> pg. 18-21.[/note]</span></span></p>
<span style="font-weight: 400;">The disconnect between Education providers and modern workplace requirements conveys the shortcomings of Higher Education. And a lot of it has to do with misdirected incentives.</span>
<span style="font-weight: 400;">Higher Education prioritise ‘credentialing’ over skill acquisition. Their incentives are skewed towards ‘bums on seats’ over teaching relevant skills and achieving learning outcomes. As such, Higher Education has increasingly become a ‘tick the box’ exercise for students, particularly in developed countries. This is apparent through the skyrocketing University fee growth across almost all advanced economies. For instance, US College fees for a four year course have risen almost 20 times since 1971.[note]The College Board (2017) <a href="https://trends.collegeboard.org/college-pricing/figures-tables/tuition-fees-room-and-board-over-time"><i>Tuition and Fees and Room and Board over Time</i></a>.[/note]</span>
<blockquote><span style="font-weight: 400;">The message here is as clear as it is concerning: it doesn’t matter what you learn, so long as you’re here.</span>
<span style="font-weight: 400;">But it does matter. It matters a great deal.</span></blockquote>
<h5>WHAT NEEDS TO BE DONE?</h5>
<span style="font-weight: 400;">There are three key areas that we see as important levers for preparing people for a future with more intelligent machines.</span>
<ol>
<li><strong> Education - <i>Adapt school and tertiary education systems to help prepare students for the changing workplace demands.</i></strong></li>
</ol>
<span style="font-weight: 400;">At a school-level, this means placing a greater emphasis on developing STEM skills, creativity, and critical thinking amongst students. These are the foundational skills that help prepare people for the demands of high-skilled labour. Schools should and will use more technology to achieve learning outcomes. This will help teachers to deliver personalised learning to help enhance students’ strengths and target their weakness for further development.</span>
<span style="font-weight: 400;">As teaching quality has the greatest in-school impact on student outcomes,[note]Hattie, John (2013) <i>Teachers Make a Difference. What is the Research Evidence?</i> ACER Research Conference.[/note]</span><span style="font-weight: 400;"> equipping our educators with these technical and abstract skills is critical. So, it’s important to adapt Initial Teacher Preparation to best prepare educators with the knowledge and pedagogies to foster these skills in our students.</span>
<span style="font-weight: 400;">At a higher-education level, there must be a reorientation towards skill acquisition in priority skill areas. Tertiary providers would do well to partner more readily with industry to inform their curriculum and add practical insight into their content. No longer are universities preparing their students for careers in academia. They’ve become the gateway to professional industry, so they should focus on industry preparation.</span>
<span style="font-weight: 400;">Students undergoing any tertiary education should be able to point to a demonstrable set of relevant and applicable skills to future employers. The internet and digital platforms are making skill development and demonstration more accessible than ever before. The opportunities to build scalable training programs for developing countries is immense, as more people come online. </span>
<span style="font-weight: 400;">The problem is that tertiary providers aren’t incentivised to do so. They’re attracting exorbitant fees because society places irrational value on credentials. Perhaps it will be necessary to adjust these incentives to help reorient education systems towards prioritising skills over credentials. Or perhaps it will instigate an influx of private training providers to fill this void.</span>
<ol start="2">
<li><strong> Policy - <i>Rethink transition support for affected workers and create favourable environments for entrepreneurs to create digitally enabled jobs.</i></strong></li>
</ol>
<span style="font-weight: 400;">Transition support is more than just Universal Basic Income (UBI). While UBI is a nice ideal, supporters aren’t clear on what constitutes ‘Universal’ or ‘Basic’. ‘Universal’ also presumes international cooperation at a time where nationalism is on the rise.[note]Onder, Harun (2016) <a href="https://www.brookings.edu/blog/future-development/2016/07/18/the-age-factor-and-rising-nationalism/"><i>The age factor and rising nationalism</i></a>, Brookings Institution.[/note]</span><span style="font-weight: 400;"> However, fiscal and education safety nets should be provided to workers that are displaced from employment.</span>
<span style="font-weight: 400;">More immediate, and from my perspective, more important, is how governments and international institutions help support the creation of new jobs, develop necessary skills in their workforces, and successfully match workers to these jobs.</span>
<span style="font-weight: 400;">Firstly, it’s important that governments expand the distribution channels of education and training. This will allow more people to acquire the skills that complement and benefit from innovations. Governments should determine how private enterprise can drive skill development. Considering the clear disconnect between traditional education providers and industry, governments would do well to support training at the source of where skill shortages are most acutely felt: private enterprises. This could mean providing incentives for in-work training programs or supporting partnerships with private training providers. In any case, industry should be incentivised to develop their workers and supported to help inform education programs.</span>
<span style="font-weight: 400;">Secondly, governments should be proactive in supporting innovation in their cities, states, and countries. Replication is not necessarily the formula here. If governments are trying to copy successes like Silicon Valley, they’re not learning from them. Silicon Valley has been successful precisely because they took a different approach and created a unique ecosystem. Sure, reducing regulatory barriers for innovation and creating tax incentives for Angel investment are important strategies that should be adopted. But the core strategy should be focused on building a unique and supportive ecosystem. </span>
<span style="font-weight: 400;">To do this, policymakers should determine the innovation domain that is (or could be) their regional competitive advantage. This could be cryptocurrency, biotech, or even a subcategory like computer vision. Once this has been determined, then the process begins of creating ‘regulatory competitive advantage’, according to that specific domain. This will help attract talent, investment, and potentially spur new domains of innovation in the ecosystem.</span>
<ol start="3">
<li><strong> Research - <i>More research in AI needs to occur to ensure that it’s safe and optimal for people.</i></strong></li>
</ol>
<span style="font-weight: 400;">The magnitude of developments in AI should not be understated. It will change the way we work, interact, and live. Indeed, it already is. </span>
<blockquote><span style="font-weight: 400;">So, you would think that being on the precipice of such significant changes, people would want to have as strong a grasp on the forces stirring these changes. The statistics paint a different a story.</span></blockquote>
<span style="font-weight: 400;">While billions are invested[note]"Spending on AI technologies by companies is expected to grow to $47 billion in 2020 from a projected $8 billion in 2016, according to IDC." Norton, Steven (2017) <i>Artificial Intelligence Looms Larger in the Corporate World</i>, The Wall Street Journal. Dow Jones & Company.[/note]</span><span style="font-weight: 400;"> in advancing the powers of AI, it’s estimated that fewer than 100 people in the world are researching ways to make AI safe[note]Farquhar, Seb (2017) <a href="http://effective-altruism.com/ea/16s/changes_in_funding_in_the_ai_safety_field/"><i>Changes in funding in the AI safety field</i></a>. The estimate of 'under 100' is based on an informal count of people doing directly relevant work at the organisations in this article, which is significantly below 100. This was cross-checked by estimating the cost per full time staff member: the forecast spending of $9m in 2017 would not be enough to sustain more than 100 staff members given the high cost of hiring machine learning researchers. Note that this figure could be inaccurate if there is a large and non-public AI safety project.[/note]</span><span style="font-weight: 400;"> and optimal for humanity. Organisations like </span><a href="https://openai.com/"><span style="font-weight: 400;">OpenAI</span></a><span style="font-weight: 400;"> and </span><a href="http://www.nickbostrom.com/"><span style="font-weight: 400;">Professor Nick Bostrom</span></a><span style="font-weight: 400;"> from University of Oxford are the leading the way.</span>
<span style="font-weight: 400;">Most of the concerns for AI surround the ‘unintended consequences’. When programmed correctly, intelligent machines do </span><i><span style="font-weight: 400;">exactly</span></i><span style="font-weight: 400;"> what you tell them, often faster, more accurately, and cheaper than any human. The smarter the system becomes, the harder it is for a person to exercise meaningful oversight. While positive opportunities abound, the negative consequences should not be dismissed.</span>
<span style="font-weight: 400;">Take financial credit and Machine Learning systems as an example. Imagine you go to the bank for loan. You’re met by a bank teller (human or otherwise!) and he requests access to your financial information. You feed him the standard details required and it’s entered into the system. The intelligent machine then looks through ALL your available data; where you live, where you work, who’s in your social network, the energy bill you forgot to pay four years ago. All of it. You awkwardly wait with the overdressed bank teller and the system runs its analysis. PING!</span>
<span style="font-weight: 400;">‘Credit Denied’.</span>
<span style="font-weight: 400;">‘What? How come?’ You cry.</span>
<span style="font-weight: 400;">The bank teller shrugs. ‘I dunno. The system said so.’</span>
<span style="font-weight: 400;">The problem with this scenario is that it presumes that the intelligent machine considers all the information, explicit and hidden. But what if you come from a poor area, you’re classified in a racial minority group, and you’ve had limited schooling opportunities? You might be an honest, hardworking, and intelligent person, but the system places unjust weightings on your inherited circumstances.</span>
<span style="font-weight: 400;">So, if intelligent machines do not account for the social biases and legacy injustices that weigh on our societies today, then we’re at risk of further perpetuating them tomorrow. </span>
<span style="font-weight: 400;">We still haven’t developed adequate solutions to a lot of social biases without machines. So blindly extending authority to intelligent machines to make these abstract decisions could be problematic.</span>
<span style="font-weight: 400;">This is not an argument to reduce progress in AI. It’s an argument to dramatically increase research in AI.</span>
<h5>PROGRESS</h5>
<span style="font-weight: 400;">While we face very real challenges with the development of intelligent machines, the application of these technologies are still our best hope. </span>
<span style="font-weight: 400;">Let me be clear, slowing technological progress in the name of ‘saving jobs’ will punish consumers and stall advancements to quality of life. Applied technologies have cured diseases, reduced famines, and helped lift entire populations out of poverty. But these technological solutions have been successful because of their </span><i><span style="font-weight: 400;">application</span></i><span style="font-weight: 400;">, not because of their technological capabilities. </span>
<span style="font-weight: 400;">This is an important distinction because it shows that technologies are never deterministic. The utopian and dystopian arguments both assume an inevitable worldview. A determinism that the arc of technological progress bends on its own.</span>
<span style="font-weight: 400;">But this reasoning is just lazy. Technology has never been deterministic. Positive outcomes depend on the applied efforts and cooperation of smart people. Entropy is not on our side.</span>
<span style="font-weight: 400;">The point is not that technology is harmful; the point is that the progress of technology does not always align neatly to the march of humanity. </span>
<span style="font-weight: 400;">So, the progress of humanity is neither guaranteed nor hopeless. Instead, it’s up to us.</span>]]></content:encoded>
<excerpt:encoded><![CDATA[]]></excerpt:encoded>
<wp:post_id>55</wp:post_id>
<wp:post_date><![CDATA[2017-08-17 21:40:05]]></wp:post_date>
<wp:post_date_gmt><![CDATA[2017-08-17 21:40:05]]></wp:post_date_gmt>
<wp:comment_status><![CDATA[open]]></wp:comment_status>
<wp:ping_status><![CDATA[open]]></wp:ping_status>
<wp:post_name><![CDATA[the-perils-of-progress]]></wp:post_name>
<wp:status><![CDATA[publish]]></wp:status>
<wp:post_parent>0</wp:post_parent>
<wp:menu_order>0</wp:menu_order>
<wp:post_type><![CDATA[post]]></wp:post_type>
<wp:post_password><![CDATA[]]></wp:post_password>
<wp:is_sticky>0</wp:is_sticky>
<category domain="post_tag" nicename="ai-policy"><![CDATA[AI Policy]]></category>
<category domain="post_tag" nicename="ai-safety"><![CDATA[AI Safety]]></category>
<category domain="category" nicename="artificial-intelligence"><![CDATA[Artificial Intelligence]]></category>
<category domain="post_tag" nicename="future-of-work"><![CDATA[Future of Work]]></category>
<category domain="post_tag" nicename="labour-markets"><![CDATA[Labour Markets]]></category>
<wp:postmeta>
<wp:meta_key><![CDATA[_edit_last]]></wp:meta_key>
<wp:meta_value><![CDATA[1]]></wp:meta_value>
</wp:postmeta>
<wp:postmeta>
<wp:meta_key><![CDATA[_oembed_0c55dd6777eada5ad00bd041ef55c5e5]]></wp:meta_key>
<wp:meta_value><![CDATA[<iframe width="1165" height="655" src="https://www.youtube.com/embed/ZxnoBfUYy04?feature=oembed" frameborder="0" allowfullscreen></iframe>]]></wp:meta_value>
</wp:postmeta>
<wp:postmeta>
<wp:meta_key><![CDATA[_oembed_time_0c55dd6777eada5ad00bd041ef55c5e5]]></wp:meta_key>
<wp:meta_value><![CDATA[1503005209]]></wp:meta_value>
</wp:postmeta>
<wp:postmeta>
<wp:meta_key><![CDATA[_themify_builder_settings_json]]></wp:meta_key>
<wp:meta_value><![CDATA[[{"row_order":"0","gutter":"gutter-default","equal_column_height":"","column_alignment":"col_align_top","cols":[{"column_order":"0","grid_class":"col-full first last","grid_width":"","modules":[],"styling":[]}],"styling":[]}]]]></wp:meta_value>
</wp:postmeta>
<wp:postmeta>
<wp:meta_key><![CDATA[builder_switch_frontend]]></wp:meta_key>
<wp:meta_value><![CDATA[0]]></wp:meta_value>
</wp:postmeta>
<wp:postmeta>
<wp:meta_key><![CDATA[_thumbnail_id]]></wp:meta_key>
<wp:meta_value><![CDATA[80]]></wp:meta_value>
</wp:postmeta>
</item>
<item>
<title>AI Governance - 11 Key Issues</title>
<link>https://bitsandatoms.co/ai-governance-11-key-issues/</link>
<pubDate>Fri, 01 Sep 2017 21:10:27 +0000</pubDate>
<dc:creator><![CDATA[[email protected]]]></dc:creator>
<guid isPermaLink="false">https://bitsandatoms.co/?p=136</guid>
<description></description>
<content:encoded><![CDATA[<b>TL;DR</b>
<ul>
<li style="font-weight: 400;"><span style="font-weight: 400;">As Artificial Intelligence (AI) is broadly applied to social and economic domains, measured oversight becomes increasingly important.</span></li>
<li style="font-weight: 400;"><span style="font-weight: 400;">But determining appropriate regulatory frameworks for AI is complex. Eleven key issues include:</span></li>
</ul>
<ol>
<li><span style="font-weight: 400;">Defining AI</span></li>
<li><span style="font-weight: 400;">Articulating</span><span style="font-weight: 400;"> ethical standards and social norms</span></li>
<li>
<p style="display: inline !important;">Accountability when AI causes harm</p>
</li>
<li><span style="font-weight: 400;">Appropriate </span><span style="font-weight: 400;">degree of oversight</span></li>
<li><span style="font-weight: 400;">Measurement & evaluation </span><span style="font-weight: 400;">of the impact</span></li>
<li>
<p style="display: inline !important;">The control problem</p>
</li>
<li>
<p style="display: inline !important;"><span style="font-weight: 400;">Openness</span></p>
</li>
<li>
<p style="display: inline !important;"><span style="font-weight: 400;">Privacy & security</span></p>
</li>
<li>
<p style="display: inline !important;"><span style="font-weight: 400;">Projections</span></p>
</li>
<li>
<p style="display: inline !important;"><span style="font-weight: 400;">Assessing institutional competence</span></p>
</li>
<li>
<p style="display: inline !important;"><span style="font-weight: 400;">The political problem</span></p>
</li>
</ol>
<ul>
<li>In the absence of robust policies, <a href="http://www.lawandai.com/about/">Matt Scherer</a> has proposed a voluntary AI certification system. AI-certified programs would be granted limited liability privileges and would provide incentives to meet safety standards. The certification standards would be established and monitored by an independent government agency.</li>
</ul>
<strong>AI Governance</strong>
<span style="font-weight: 400;">Knowingly or unknowingly, Artificial Intelligence systems are intersecting with more parts of our lives. And not just in areas of trivial importance. AI systems are being applied to essential areas of society. From analysisng Electronic Health Records that improve diagnosis rates; to balancing power supply for energy grids. AI can help us achieve more and raise standards of living.</span>
<span style="font-weight: 400;">So, as AI systems are deployed at scale within fundamental societal structures, then measured oversight at scale becomes necessary.</span>
<span style="font-weight: 400;">Such public interest roles are typically assumed by the arms of national governments. But AI public policy has been met with almost radio silence across the world. As a result, the development and applications of AI continue to exist in a policy vacuum. </span>
<b>Issues with Governing AI</b>
<span style="font-weight: 400;">The unique challenges and complexities of AI do not fit neatly into existing governance frameworks. Safety standards are fluid. Accountability is opaque. And policy-makers lack expertise. The amount of investment in developing AI has exceeded investments in making AI safe by an order of magnitude[note]Farquhar, Seb (2017) <a href="http://effective-altruism.com/ea/16s/changes_in_funding_in_the_ai_safety_field/"><i>Changes in funding in the AI safety field</i></a>.[/note].</span><span style="font-weight: 400;"> This is fueling the immense growth in AI applications with almost unfettered regulatory oversight. The surprising thing is that many of the most prominent Tech leaders, such as Elon Musk[note]Kurt Wagner, <a href="https://www.recode.net/2017/7/15/15976744/elon-musk-artificial-intelligence-regulations-ai"><i>Elon Musk just told a group of America’s governors that we need to regulate AI before it’s too late</i></a>, Recode (July. 15, 2017).[/note]</span><span style="font-weight: 400;"> and Bill Gates[note]<em>For example</em>, see Eric Mack, <a href="http://www.forbes.com/sites/ericmack/2015/01/28/bill-gates-also-%20worries-artificial-intelligence-is-a-threat/"><i>Bill Gates Says You Should Worry About Artificial Intelligence</i></a>, FORBES (Jan. 28, 2015)[/note]</span><span style="font-weight: 400;">, think that a degree of regulatory oversight is important. </span>
<span style="font-weight: 400;">Meanwhile, policy-makers sit idle. AI is viewed as a black box. Most are unclear about what AI actually is, let alone instituting appropriate governance.</span>
<span style="font-weight: 400;">So, what are the main issues with governing AI?</span>
<span style="font-weight: 400;">Like with all major public policy areas, the issues extend beyond just hard technical problems. There are conceptual issues, as well as practical problems.</span>
<span style="text-decoration: underline;"><span style="font-weight: 400;">Conceptual Policy Issues</span></span>
<ul>
<li style="font-weight: 400;"><i><span style="font-weight: 400;">Defining AI</span><span style="font-weight: 400;"> - </span></i><span style="font-weight: 400;">The problems with defining AI for regulatory purposes </span><span style="font-weight: 400;">centre around the conceptual ambiguities of ‘intelligence’. The definitions of intelligence vary widely. Intellectual characteristics like ‘consciousness’ and ‘the ability to learn’ are at best nebulous. So, arriving at an agreed definition for Artificial Intelligence is difficult. The subjective nature of AI terminology means that it becomes a moving target for policy-makers. Definitions of AI range from: ‘the ability to act ‘humanly’’[note]A. M. Turing (1950) Computing Machinery and Intelligence. Mind 49: 433-460.[/note];</span><span style="font-weight: 400;"> to ‘performing intellectual tasks’[note]<em>See</em> BRUCE PANDOLFINI, KASPAROV AND DEEP BLUE: THE HISTORIC CHESS MATCH BETWEEN MAN AND MACHINE 7–8 (1997).[/note];</span><span style="font-weight: 400;"><span style="font-weight: 400;"> and the modern definition of ‘acting rationally to achieve goals’[note]</span></span>'Intelligent Machines' or 'Artificial Intelligence' refers to a non-organic autonomous entities that are able to sense and act upon an environment to achieve specific goals. Intelligent agents may also learn or use knowledge to achieve these goals, which are governed by algorithms that are made by people. Russell, Stuart J.; Norvig, Peter (2003), <i>Artificial Intelligence: A </i>Modern<i> Approach </i>(2nd ed.), Upper Saddle River, New Jersey: Prentice Hall, ISBN 0-13-790395-2, chpt. 2.; <i>See also </i>Stephen M. Omohundro, The Basic AI Drives, in ARTIFICIAL GENERAL INTELLIGENCE 2008 483, 483 (2008) (defining AI as a system that “has goals which it tries to accomplish by acting in the world”).[/note]. However, even a ‘goal-oriented’ approach doesn’t provide clarity for a regulatory definition.</li>
<li style="font-weight: 400;"><em><span style="font-weight: 400;">Ethical Standards & Social Norms</span></em><span style="font-weight: 400;"> - For autonomous systems to operate effectively in society, they need to so ethically and in alignment with social norms. But what is good behaviour? What is just? Any attempt to develop AI governance structures will inevitably confront such philosophical questions. </span></li>
<li style="font-weight: 400;"><em><span style="font-weight: 400;">Accountability</span></em><span style="font-weight: 400;"> - Assigning liability for when autonomous systems negatively perform is a difficult conceptual and practical challenge. This will be particularly important in social and economic domains. For instance, to what degree can a physician rely on intelligent diagnosis systems without increasing exposure to malpractice claims in the case of a systems error? Precedent in case law is sparse. And the applications of AI systems are rapidly expanding in the absence of </span><span style="font-weight: 400;">ex-ante </span><span style="font-weight: 400;">accountability frameworks.</span></li>
<li style="font-weight: 400;"><em><span style="font-weight: 400;">The Degree of Oversight</span></em><span style="font-weight: 400;"> - The extent of regulation is always a delicate balance. Ultimately, an ideal AI governance structure would help maximise the opportunities for positive outcomes, while minimising the negative risks. The advantage is that AI development is still in its infancy. However, a failure to institute appropriate oversight could yield unfavourable outcomes. If regulations go too far, innovation could be inhibited and societal benefits lost. If regulations don’t go far enough, negative outcomes at scale could result and knee-jerk policy reactions ensue. We’ve seen this before in other sectors, such as Bioengineering and Biomedicine. For instance, the impact that Thalidomide had on tightening FDA regulations on drug classifications in the US.[note]Bren L (2001-02-28). "Frances Oldham Kelsey: FDA Medical Reviewer Leaves Her Mark on History". <i>FDA Consumer</i>. U.S. Food and Drug Administration.[/note]</span></li>
</ul>
<span style="text-decoration: underline;"><span style="font-weight: 400;">Practical Policy Issues</span></span>
<ul>
<li style="font-weight: 400;"><i><span style="font-weight: 400;">Measurement & Evaluation</span></i><span style="font-weight: 400;"> - While technical progress is being made in the emerging field of AI Safety, we currently lack agreed upon methods to assess the social and economic impacts of AI systems. Robust M&E methods are important as they support investigative, regulatory, and enforcement functions. They help set benchmarks, so we can know AI applications are producing positive outcomes.</span></li>
<li><em><span style="font-weight: 400;">The Control Problem</span></em><span style="font-weight: 400;"> - The risks associated with control of autonomous systems is a core problem across all segments of AI. In the case of autonomous Machine Learning systems, there are risks that as they continue to learn and adapt, the potential for human control is inhibited. Once control is lost, it may be difficult to regain control. From a policy perspective, there are obvious public risks. So, if the potential for such scenarios is in any way more than theoretical, then the assurance of human control and public alignment will be necessary.</span></li>
<li style="font-weight: 400;"><em><span style="font-weight: 400;">Openness </span></em><span style="font-weight: 400;">- Determining the desirability of openness in </span><span style="font-weight: 400;">AI research & development is a key issue for policy-makers (including openness about source code, science, data, safety techniques, capabilities, and goals). Types and degrees of</span><span style="font-weight: 400;"> openness result in complex societal tradeoffs, particularly in the long-term. While higher levels of openness will likely accelerate AI development, it may also exacerbate a racing dynamic: a situation where competitors race to develop the first General Artificial Intelligence. Such a dynamic may result in inadequate safety measures in order to accelerate progress.[note]Bostrom, Nick (2017) “Strategic Implications of Openness in AI Development.” Global Policy 8 (2): 135–48.[/note]</span><span style="font-weight: 400;"> This scenario increases the public exposure to systemic risks. It’s important to note that technology and policy decisions are never deterministic. We can’t know for certain that any scenarios will come to pass. It’s plausible, however, that the lever of Openness will have significant second, third, and fourth-order effects. Therefore, it’s an important policy consideration.</span></li>
<li style="font-weight: 400;"><i><span style="font-weight: 400;">Privacy & Security</span></i><span style="font-weight: 400;"> - Data is the life source of AI systems. Maintaining standards that uphold privacy and ensure the security of the data accessed by AI systems is a key technical and policy challenge. People should have the right to access, manage, and control the data they generate, given AI systems’ power to analyse and utilise the data. Moreover, it’s imperative that personal data is securely stored and not unscrupulously accessed or used without expressed consent.</span></li>
<li style="font-weight: 400;"><em><span style="font-weight: 400;">Projections</span></em><span style="font-weight: 400;"> - The decision-making processes of AI systems are diametrically different to those of humans. That’s why AI systems generate solutions that humans never considered.[note]<i>For e.g. see:</i> Cade Metz, <a href="https://www.wired.com/2016/03/two-moves-alphago-lee-sedol-redefined-future/"><i>IN TWO MOVES, ALPHAGO AND LEE SEDOL REDEFINED THE FUTURE</i></a>. WIRED (16. March, 2017).[/note]</span><span style="font-weight: 400;"> This ability to create value through unexpected solutions is a fundamental point of attraction towards AI systems. It’s also a risk. Accurately projecting adverse effects from AI systems is difficult, precisely because outcomes can be unexpected. As AI increasingly enters into social and economic domains, policy-makers will seek reassurance from projections as part of due diligence. But there aren’t clear projection methods.</span></li>
<li style="font-weight: 400;"><em><span style="font-weight: 400;">Assessing Institutional Competence</span></em><span style="font-weight: 400;"> - Even if it were decided that regulatory oversight should be instituted for broad-scope AI, governance structures still need to be determined. There are notable issues at hand: legislators lack expertise; courts can’t act quickly enough on a case-by-case basis to establish precedent; and international institutions can be perceived as toothless tigers. While challenging, there are lessons in history of effective governance structures to oversee powerful technologies. The Treaty on the Nonproliferation of Nuclear Weapons offers relevant insight. While still an underserved research area, Matt Scherer proposes a useful regulatory framework for AI, which is summarised below.</span></li>
<li style="font-weight: 400;"><i><span style="font-weight: 400;">The Political Problem</span></i><span style="font-weight: 400;"> - The current and potential powers of AI are not deterministic. They depend on their applications, which are currently decisions made by humans. Like with any source of power, there’s potential for good and subversion. The political challenge with AI is to achieve a situation in which individuals or institutions empowered by such AI use it in ways that promote the common good. At a time where nationalism is on the rise,[note]Onder, Harun (2016) <i>The age factor and rising nationalism</i>, Brookings Institution.[/note]</span><span style="font-weight: 400;"> international cooperation is becoming increasingly difficult. Political cooperation, however, is necessary to the safe broad-scale deployment of AI, which transcends national borders.</span></li>
</ul>
<span style="font-weight: 400;">These issues, taken together, highlight the complexities of establishing appropriate AI policies. National governments are still in the early days of their thinking. Last year, the US government held a series of public workshops with industry and research leaders. This resulted in a summary report presented to The White House.[note]US National Science & Technology Committee on Technology (2016) “Preparing for the Future of Artificial Intelligence.” Executive Office of the President.[/note]</span><span style="font-weight: 400;"> Similarly, the UK House of Commons commissioned an inquiry into the opportunities and implications of Robotics and Artificial Intelligence.[note]UK House of Commons (2016) “<a href="https://www.publications.parliament.uk/pa/cm201617/cmselect/cmsctech/145/145.pdf">Robotics and Artificial Intelligence - United Kingdom Parliament.</a>” n.d.[/note]</span><span style="font-weight: 400;"> While the intent is positive, policy positions are still abstract. This demonstrates the elementary understanding of how broad-scale AI might impact society. Let alone the potential roles of public policy.</span>
<b>A Proposed AI Regulatory Framework</b>
<span style="font-weight: 400;">In the absence of robust policies, </span><a href="http://www.lawandai.com/about/"><span style="font-weight: 400;">Matt Scherer</span></a><span style="font-weight: 400;">, an attorney and legal scholar from the US, has presented a useful proposal to regulate AI systems.[note]Scherer, Matthew U. (2016) “Regulating Artificial Intelligence Systems: Risks, Challenges, Competencies, and Strategies,” <i>Harvard Journal of Law and Technology</i> 29 (2): 354-400.[/note]</span><span style="font-weight: 400;"> The centrepiece of this tort-based framework involves an AI certification process. Certification would require designers, manufacturers, and sellers of AI systems to fulfil safety and legal standards. These standards would be developed and monitored by an independent AI Agency that’s appropriately staffed by AI specialists.</span>
<span style="font-weight: 400;">Scherer proposes that rather than creating an AI Agency with ‘FDA-like powers’ to ban products, AI programs that are successfully certified could be granted limited liability. This means that plaintiffs would have to establish </span><i><span style="font-weight: 400;">actual negligence</span></i><span style="font-weight: 400;"> in the design, manufacturing, or operation of an AI system to be successful in a tort claim. The uncertified AI programs would still be available for commercial sale but would be subject to strict joint and several liability. Successful plaintiffs would, therefore, be permitted to ‘</span><i><span style="font-weight: 400;">recover the full amount of their damages from any entity in the chain of development, distribution, sale, or operational of the uncertified AI</span></i><span style="font-weight: 400;">’.[note]Ibid pg. 395.[/note]</span>
<span style="font-weight: 400;">Another advantage to Scherer’s proposal is that it leverages the institutional strengths of legislatures, agencies, and courts. As a summary, this structure would allocate roles in the following ways:</span>
<ul>
<li style="font-weight: 400;"><em>Legislature</em><span style="font-weight: 400;"> - This system would utilise the democratic mandate of the Legislature to determine the goals and purposes that guide AI governance. It would also use the powers of the Legislature to enact legislation (Scherer refers to this as the ‘Artificial Intelligence Development Act’) that would create an independent agency for oversight.</span></li>
<li style="font-weight: 400;"><em>Independent Agency</em><span style="font-weight: 400;"> - As legislators lack the specialist knowledge required, they would delegate the central task of assessing the safety of AI systems to an independent agency of AI specialists. Independence is key, as it will help inculcate the Agency from the jockeying of electoral politics. An independent agency also has the flexibility to act preemptively. This flexibility and responsiveness is particularly important as AI development continues at breakneck speeds.</span></li>
<li style="font-weight: 400;"><em>Courts</em><b><span style="font-weight: 400;"> - The courts would be utilised for their strengths in adjudicating cases and allocating responsibility. This would require the courts to apply the rules governing negligence claims, differentiating between certified-AI with limited liability and uncertified-AI with strict liability. A core role of the courts will be allocating responsibility to the parties that caused harm through the AI program.</span></b></li>
</ul>
<span style="font-weight: 400;">This proposed structure isn’t a panacea to the list of issues above. It does, however, provide a flexible regulatory framework for oversight, without draconian regulations. By leveraging tort systems, the proposed structure would provide strong incentives for AI developers to incorporate safety features and internalise the associated costs. It would also provide a disincentive for distributors to sell uncertified AI programs that haven’t met public safety standards.</span>
<span style="font-weight: 400;">Regardless of whether Scherer’s proposal is considered appropriate, governments will need to develop policy positions for broad-scope AI. This will take careful planning and consideration. It will also require a sense of urgency. Ultimately, the future depends on what we do in the present.</span>]]></content:encoded>
<excerpt:encoded><![CDATA[]]></excerpt:encoded>
<wp:post_id>136</wp:post_id>
<wp:post_date><![CDATA[2017-09-01 21:10:27]]></wp:post_date>
<wp:post_date_gmt><![CDATA[2017-09-01 21:10:27]]></wp:post_date_gmt>
<wp:comment_status><![CDATA[open]]></wp:comment_status>
<wp:ping_status><![CDATA[open]]></wp:ping_status>
<wp:post_name><![CDATA[ai-governance-11-key-issues]]></wp:post_name>
<wp:status><![CDATA[publish]]></wp:status>
<wp:post_parent>0</wp:post_parent>
<wp:menu_order>0</wp:menu_order>
<wp:post_type><![CDATA[post]]></wp:post_type>
<wp:post_password><![CDATA[]]></wp:post_password>
<wp:is_sticky>0</wp:is_sticky>
<category domain="category" nicename="artificial-intelligence"><![CDATA[Artificial Intelligence]]></category>
<category domain="post_tag" nicename="artificial-intelligence-policy"><![CDATA[Artificial Intelligence Policy]]></category>
<wp:postmeta>
<wp:meta_key><![CDATA[_edit_last]]></wp:meta_key>
<wp:meta_value><![CDATA[1]]></wp:meta_value>
</wp:postmeta>
<wp:postmeta>
<wp:meta_key><![CDATA[_themify_builder_settings_json]]></wp:meta_key>
<wp:meta_value><![CDATA[[{"row_order":"0","gutter":"gutter-default","equal_column_height":"","column_alignment":"col_align_top","cols":[{"column_order":"0","grid_class":"col-full first last","grid_width":"","modules":[],"styling":[]}],"styling":[]}]]]></wp:meta_value>
</wp:postmeta>
<wp:postmeta>
<wp:meta_key><![CDATA[builder_switch_frontend]]></wp:meta_key>
<wp:meta_value><![CDATA[0]]></wp:meta_value>
</wp:postmeta>
<wp:postmeta>
<wp:meta_key><![CDATA[_thumbnail_id]]></wp:meta_key>
<wp:meta_value><![CDATA[143]]></wp:meta_value>
</wp:postmeta>
</item>
<item>
<title>Designing Effective Policies for Safety-Critical AI</title>
<link>https://bitsandatoms.co/effective-policies-for-safety-critical-ai/</link>
<pubDate>Fri, 22 Sep 2017 05:02:39 +0000</pubDate>
<dc:creator><![CDATA[[email protected]]]></dc:creator>
<guid isPermaLink="false">https://bitsandatoms.co/?p=152</guid>
<description></description>
<content:encoded><![CDATA[<h3><span style="font-weight: 400;">Key policy considerations for national governments</span></h3>
<span style="font-weight: 400;">There’s growing noise around ‘regulating’ AI.[note]For example, see: Oren Etzioni, ‘<a href="https://www.nytimes.com/2017/09/01/opinion/artificial-intelligence-regulations-rules.html?mcubz=0">How to Regulate Artificial Intelligence</a>’, New York Times (1 Sept 2017).[/note]</span><span style="font-weight: 400;"> Some claim it’s too early,[note]Ahmed, K (2015) “<a href="http://www.bbc.com/news/business-34266425">Google’s Demis Hassabis - misuse of artificial intelligence ‘could do harm’</a>", BBC News.[/note]</span><span style="font-weight: 400;"> citing that precautionary regulations could impede technical developments; others call for action,[note]Scherer, Matthew U. (2015) "Regulating artificial intelligence systems: Risks, challenges, competencies, and strategies." <i>Harvard Journal of Technology and Law</i>.[/note]</span><span style="font-weight: 400;"> advocating measures that could mitigate the risks of AI.</span>
<span style="font-weight: 400;">It’s an important problem. And both ends of the debate make compelling arguments. AI applications have the potential to improve output, productivity, and quality of life. Forestalling AI developments that facilitate these advancements are big opportunity costs. Equally, the risks of broad-scope AI applications shouldn’t be dismissed. There are near-term implications, like job displacement and autonomous weapons, and longer-term risks, like values misalignment and the </span><a href="https://futureoflife.org/2015/11/23/the-superintelligence-control-problem/"><span style="font-weight: 400;">control problem</span></a><span style="font-weight: 400;">.</span>
<span style="font-weight: 400;">Regardless of where one sits on the ‘AI regulation’ spectrum, few would disagree that policymakers should have a firm grasp on the development and implications of AI. It’s unsurprising, given the rapid developments, that most do not. </span>
<b>Asymmetry of Knowledge</b>
<span style="font-weight: 400;">Policymakers are still very much at the beginning of learning about AI. The US government held public hearings late last year to ‘survey the current state of AI’.[note]US National Science & Technology Committee on Technology (2016) “Preparing for the Future of Artificial Intelligence.” Executive Office of the President.[/note]</span><span style="font-weight: 400;"> Similarly, the UK House of Commons undertook an inquiry to identify AIs’ ‘potential value’ and ‘prospective problems’.[note]UK House of Commons (2016) “<a href="https://www.publications.parliament.uk/pa/cm201617/cmselect/cmsctech/145/145.pdf">Robotics and Artificial Intelligence - United Kingdom Parliament.</a>”[/note]</span>
<span style="font-weight: 400;">While these broad inquiries signify positive engagement, they also highlight policymakers’ relatively poor understanding, particularly compared to industry. This is understandable given the majority of A</span><span style="font-weight: 400;">I </span><span style="font-weight: 400;">development and expertise is concentrated in a select few organisations. This handful of organisations (literally, counted on two hands) possess orders of magnitude more data than anyone else. This access to data, coupled with resources and technical expertise, has fueled the rapid and concentrated development of AI. </span>
<span style="font-weight: 400;">Governments lacking in-house expertise is problematic from a policy development perspective. As AI increasingly intersects with safety-critical areas of society, governments hold responsibilities to act in the interests of their citizens. But if they don’t have the ability to formulate measured policies in accordance with these interests, then unintended consequences could arise, placing their citizens at risk. Without belabouring scenarios of misguided policies, governments should prioritise building their own expertise. Whether they’re prepared or not, governments are key stakeholders. They hold Social Contracts with their citizens to act on their behalf. So, as AI is applied to safety-critical industries, like healthcare, energy, and transportation, understanding the opportunities and implications is essential.</span>
<span style="font-weight: 400;">Ultimately, knowledge and expertise are central to effective policy decisions. And independence helps align policies to the public interest. While the spectrum of potential policy actions for safety-critical AI is broad, all with their own effects, inaction is also a policy position. This is where most governments are at. I think it’s important they rigorously consider the implications of these decisions.</span>
<span style="font-weight: 400;">Let me be clear: I’m not advocating for ‘more’ or ‘less’ regulation of AI. I’m advocating for governments to build their capacity to make effective and independent policy decisions. At the moment, few are qualified to do so. That’s a big reason why AI has developed in a policy vacuum. </span>
<b>‘Policy’ rather than ‘Regulation’</b>
<span style="font-weight: 400;">The term ‘regulation’ is not helpful in advancing the discourse on the roles of governments with AI. ‘Regulation’ can evoke perceptions of restriction. A heavy hand impeding growth, rather than nourishing progress. This is counterproductive, particularly in these early stages.</span>
<span style="font-weight: 400;">More useful, and less abrasive, is the term ‘policy’. Policy simply refers to a set of decisions that societies, through their governments, make about what they do and do not want to permit, and what they do or do not want to encourage.[note]Brundage, Miles, and Joanna Bryson, (2016) “<a href="http://arxiv.org/abs/1608.08196">Smart Policies for Artificial Intelligence</a>.” <i>arXiv,</i> pg. 2.[/note]</span><span style="font-weight: 400;"> Policies can be ‘pro-innovation’, helping to accelerate the development and diffusion of technologies. Policies can also decelerate and redirect technological development and diffusion. Science and Technology policy has a long history of both, and everything in between. </span>
<span style="font-weight: 400;">So, the term ‘policy’ is necessarily flexible. It may sound like semantics. But language matters.</span>
<b>Safety-Critical AI</b>
<span style="font-weight: 400;">Safety-critical AI refers to autonomous systems whose malfunction or failure can lead to serious consequences.[note]Nusser, Sebastian, (2009) “<a href="https://www.researchgate.net/publication/40220479_Robust_Learning_in_Safety-Related_Domains_machine_learning_methods_for_solving_safety-related_application_problems">Robust Learning in Safety-Related Domains : machine learning methods for solving safety-related application problems</a>”, OAI.[/note]</span><span style="font-weight: 400;"> This could include</span><span style="font-weight: 400;"> adverse environmental effects, loss or severe damage of equipment, harm or serious injury of people, or even death.</span>
<span style="font-weight: 400;">While there’s no formal definition of what constitutes ‘safety-critical’, governments already identify sectors considered ‘critical infrastructure’. The </span><a href="https://www.dhs.gov/critical-infrastructure-sectors"><span style="font-weight: 400;">US Department of Homeland Security</span></a><span style="font-weight: 400;"> classifies:</span>
<span style="font-weight: 400;">“</span><i><span style="font-weight: 400;">16 critical infrastructure sectors whose assets, systems, and networks, whether physical or virtual, are considered so vital to the United States that their incapacitation or destruction would have a debilitating effect on security, national economic security, national public health or safety, or any combination thereof.</span></i><span style="font-weight: 400;">”</span>
<span style="font-weight: 400;">Similarly, the Australian government (my home country) defines critical infrastructure as:</span>
<span style="font-weight: 400;">“</span><i><span style="font-weight: 400;">… those physical facilities, supply chains, information technologies and communication networks which, if destroyed, degraded or rendered unavailable for an extended period, would significantly impact on the social or economic wellbeing of the nation or affect Australia’s ability to conduct national defence and ensure national security.”</span></i><span style="font-weight: 400;">[note]Australia New Zealand Counter Terrorism Committee, (2015) “<a href="https://www.nationalsecurity.gov.au/Media-and-publications/Publications/Documents/national-guidelines-protection-critical-infrastructure-from-terrorism.pdf">National Guidelines for Protecting Critical Infrastructure from Terrorism</a>”, pg. 3.[/note]</span>
<span style="font-weight: 400;">These include sectors like energy, financial services, and transportation. Scrolling through the list, we can see AI already being applied in all of these sectors. </span>
<span style="font-weight: 400;">This as it should be. AI can improve productivity in the critical sectors of society and help us achieve more.</span>
<span style="font-weight: 400;">The problem is that AI systems designers still face very real challenges in making AI safe. These challenges are exacerbated, and their importance heightened, when AI’s are applied to safety-critical sectors. The stakes are high.</span>
<b>Concrete Problems in AI Safety</b>
<span style="font-weight: 400;">Dario Amodei et al. provide an excellent account of ‘</span><a href="https://arxiv.org/pdf/1606.06565.pdf"><span style="font-weight: 400;">Concrete Problems in AI Safety</span></a><span style="font-weight: 400;">’. This paper lists five practical research problems related to </span><i><span style="font-weight: 400;">accidents</span></i><span style="font-weight: 400;"> in machine learning systems (the most dominant subcategory of AI) that may emerge from poor design of real-world AI systems. A summary of the five problems are as follows:[note]Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané (2016) “<a href="http://arxiv.org/abs/1606.06565">Concrete Problems in AI Safety</a>.” <i>arXiv.</i>[/note]</span>
<ol>
<li style="font-weight: 400;"><b>Avoiding Negative Side Effects:</b><span style="font-weight: 400;"> ensuring that AI systems do not disturb the environment in negative ways while pursuing its goals.</span></li>
<li style="font-weight: 400;"><b>Avoiding Reward Hacking:</b><span style="font-weight: 400;"> ensuring that AI systems do not ‘game’ its reward function by exploiting bugs in its environment and acting in unintended (and potentially harmful) ways to achieve its goal(s). </span></li>
<li style="font-weight: 400;"><b>Scaleable Oversight: </b><span style="font-weight: 400;">ensuring that AI systems do the right things at scale despite limited information.</span></li>
<li style="font-weight: 400;"><b>Safe Exploration:</b><span style="font-weight: 400;"> ensuring that AI systems don’t make exploratory moves (i.e. try new things) with bad repercussions.</span></li>
<li style="font-weight: 400;"><b>Robustness to Distributional Shift:</b><span style="font-weight: 400;"> ensuring that AI systems recognise, and behave robustly, when in an environment different from its training environment.</span></li>
</ol>
<span style="font-weight: 400;">In addition to these safety problems from </span><i><span style="font-weight: 400;">unintentional accidents</span></i><span style="font-weight: 400;"> are the safety problems with AI systems designed to inflict </span><i><span style="font-weight: 400;">intentional </span></i><span style="font-weight: 400;">harm. We saw tastes of this during the </span><a href="http://www.npr.org/sections/alltechconsidered/2017/04/03/522503844/how-russian-twitter-bots-pumped-out-fake-news-during-the-2016-election"><span style="font-weight: 400;">2016 US Presidential Election</span></a><span style="font-weight: 400;">. Russian actors used Twitter and Facebook bots to create and proliferate derogatory claims and ‘Fake News’ about the Clinton campaign. While it appears most of these bots are considered ‘dumb AI’ (for example, programmed only to robotically retweet specific accounts), it’s a firm step towards </span><a href="https://medium.com/join-scout/the-rise-of-the-weaponized-ai-propaganda-machine-86dac61668b"><span style="font-weight: 400;">AI political propaganda</span></a><span style="font-weight: 400;">. There’s an immediate risk that machine learning techniques will be applied at scale in political campaigns to manipulate public engagement. This automated political mobilisation won’t be concerned with what’s ‘real’ or ‘true’. Its goals are to build followings, change minds, and mobilise votes.</span>
<span style="font-weight: 400;">Therefore, the risks of </span><i><span style="font-weight: 400;">unintentional</span></i><span style="font-weight: 400;"> and </span><i><span style="font-weight: 400;">intentional</span></i><span style="font-weight: 400;"> AI harm to the public are significant. As representatives of the public, it’s incumbent upon governments to: (a) develop institutional competencies and expertise in AI; and (b) institute measured and appropriate policies, guided by these competencies and expertise, that maximise the public benefits of AI, while minimising the public risks. </span>
<b>Skilling-Up Governments</b>
<span style="font-weight: 400;">Improving internal AI competencies within governments is a recommendation that regularly arises.[note]For example, see: Calo, Ryan (2014) “<a href="https://www.brookings.edu/research/the-case-for-a-federal-robotics-commission/">The Case for a Federal Robotics Commission</a>.” <i>Brookings</i>; Scherer, Matthew U. (2015) "Regulating artificial intelligence systems: Risks, challenges, competencies, and strategies." <i>Harvard Journal of Technology and Law</i>; UK House of Commons (2016) “<a href="https://www.publications.parliament.uk/pa/cm201617/cmselect/cmsctech/145/145.pdf">Robotics and Artificial Intelligence - United Kingdom Parliament.</a>”[/note]</span><span style="font-weight: 400;"> For reasons aforementioned, it’s broadly agreed to be an important step. A key challenge is how to attract and retain talent.</span>
<span style="font-weight: 400;">If governments (particularly Western governments) are to develop internal AI expertise, they’ll inevitably compete for talent with the likes of Google and Facebook. These companies build their technical outfit by offering enormous remuneration, lots of autonomy, great working conditions, and the social status of working for a company ‘changing the world’.</span>
<span style="font-weight: 400;">Working for the Department of Social Services for salary doesn’t quite have the same ring to it. </span>
<span style="font-weight: 400;">It’s also a function of social mood. Nationalism is on the rise,[note]Onder, Harun (2016) <a href="https://www.brookings.edu/blog/future-development/2016/07/18/the-age-factor-and-rising-nationalism/"><i>The age factor and rising nationalism</i></a>, Brookings Institution.[/note]</span><span style="font-weight: 400;"> trust in institutions has collapsed,[note]“<a href="https://www.edelman.com/trust2017/">2017 Edelman TRUST BAROMETER</a>.” 2017. <i>Edelman</i>.[/note]</span><span style="font-weight: 400;"> and the internet has afforded more opportunities for flexible and independent labour than any point in history. These cultural dynamics affect how governments operate and what people demand from them. </span>
<span style="font-weight: 400;">Of course, it’s possible for governments to stir inspiration and coordinate talent towards hard, technical goals. After all, we did ‘</span><a href="https://er.jsc.nasa.gov/seh/ricetalk.htm"><span style="font-weight: 400;">choose to go to the moon</span></a><span style="font-weight: 400;">’. But the audacious and inspirational plans of the 1960’s, encouraged by a culture of </span><a href="http://blakemasters.com/post/23435743973/peter-thiels-cs183-startup-class-13-notes"><span style="font-weight: 400;">definite optimism</span></a><span style="font-weight: 400;">, feels a far cry from the incremental plans and rising fog of pessimism that weighs on many governments today. </span>
<span style="font-weight: 400;">For problems as hard as AI policy, attracting the best and brightest is a crucial, yet formidable task. This is </span><a href="https://www.vox.com/2017/3/14/14924524/denis-mcdonough-podcast"><span style="font-weight: 400;">especially difficult</span></a><span style="font-weight: 400;"> for the roles at, or below, middle-management.</span>
<span style="font-weight: 400;">Governments are experimenting with ways to build technical talent. And there are some interesting initiatives happening in the periphery. For instance, the Obama Administration introduced the </span><a href="https://presidentialinnovationfellows.gov/"><span style="font-weight: 400;">Presidential Innovation Fellows</span></a><span style="font-weight: 400;"> program. Fellows serve for 12 months as embedded entrepreneurs-in-residence to help build the technological capacities of government agencies. The program attracts talent from the most prominent tech firms for ‘tours of duty’. They’re given resources and support to help work on important technical projects under the auspices of the Federal Government. </span>
<span style="font-weight: 400;">While positive, secondments won’t suffice for building AI competencies within governments. As the applications of AI are so broad, affecting so many safety-critical areas, governments have a responsibility to be prepared. At this stage, preparation involves understanding the unique opportunities and implications of AI. For many governments, it’s not clear who is responsible for this task and where expertise should reside.</span>
<span style="font-weight: 400;">This issue has led researchers, such as Ryan Calo[note]Calo, Ryan (2014) “<a href="https://www.brookings.edu/research/the-case-for-a-federal-robotics-commission/">The Case for a Federal Robotics Commission</a>.” <i>Brookings.</i>[/note]</span><span style="font-weight: 400;"> and Matt Scherer[note]Supra note 3.[/note],</span><span style="font-weight: 400;"> to recommend the establishment of government agencies specifically for AI & Robotics.</span>
<b>The case for AI & Robotics Agencies</b>
<span style="font-weight: 400;">Efforts to address AI & Robotics policy decisions have been piecemeal, at best. However, as AI is increasingly scaled across sectors, Calo argues that the diffusion of expertise across existing agencies and departments makes less sense.[note]Supra note 12.[/note]</span><span style="font-weight: 400;"> A more centralised agency would provide a repository of expertise to advise and formulate policies recommended for governments.</span>
<i><span style="font-weight: 400;">Why an Agency?</span></i>
<span style="font-weight: 400;">In the context of AI, Scherer outlines the benefits and appropriateness of administrative agencies, stating:[note]Supra note 3, pg. 381.[/note]</span>
<ul>
<li><b>Flexibility<span style="font-weight: 400;"> - Agencies can be ‘tailor-made’ for a specific industry or particular social problem;</span></b></li>
</ul>
<ul>
<li><strong>Specialisation & Expertise</strong><span style="font-weight: 400;"> - Policymakers in agencies can be experts in their field rather than the more generalist roles required by courts and legislatures; </span></li>
</ul>
<ul>
<li><strong>Independence & Autonomy</strong><span style="font-weight: 400;"> - Agencies have more latitude to conduct independent factual investigations that serve as a basis for their policy decisions; and</span></li>
<li><strong>Ex Ante Action</strong><span style="font-weight: 400;"> - Similar to legislatures, agencies have the ability to formulate policy before harmful conduct occurs.</span></li>
</ul>
<i><span style="font-weight: 400;">What would an AI Agency do?</span></i>
<span style="font-weight: 400;">Views on the potential roles of AI & Robotics agencies vary. They’re also necessarily country-specific. The key point of conjecture surrounds the degree of enforceability that an agency would assume.</span>
<span style="font-weight: 400;">Calo’s view is that a Federal Agency should act, at least initially, as an internal repository of expertise. A standalone entity tasked ‘</span><i><span style="font-weight: 400;">with the purpose of fostering, learning about, and advising upon</span></i><span style="font-weight: 400;">’ the impacts of AI & Robotics on society.[note]Supra note 12. NB: While Calo refers only to Robotics in this publication, his position has been expanded to AI more generally. For example, see this <a href="https://www.wired.com/story/elon-forget-killer-robots-focus-on-the-real-ai-problems/">article</a>.[/note]</span><span style="font-weight: 400;"> This would help cultivate a deep appreciation of the technologies underlying AI. And governments will have unconstrained access to independent advice on the development trends, deployment progress, and inevitable risks that AI actually presents. If the risks develop such that stronger safety regulations are necessary, then an agency is in place to formulate these policies.</span>
<span style="font-weight: 400;">Both Scherer and Calo agree that an agency would provide an increasingly important resource for governments. However, Scherer proposes that a US AI Federal Agency should assume a more proactive role, sooner rather than later. The regulatory framework put forth would be based on a voluntary AI certification process, which would be managed by the agency.[note]Supra note 3.[/note]</span><span style="font-weight: 400;"> AI systems that are certified enjoy limited liability, whereas uncertified AI systems would assume full liability. (To learn more about it, read my blog </span><a href="https://bitsandatoms.co/ai-governance-11-key-issues/"><span style="font-weight: 400;">article</span></a><span style="font-weight: 400;"> summarising the research or read Scherer’s </span><a href="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=2609777"><span style="font-weight: 400;">paper</span></a><span style="font-weight: 400;">) </span>
<span style="font-weight: 400;">Despite these differing views on the scope of roles, the advent of new agencies in response to new technologies is not new. Radio, aviation, and space travel all resulted in the creation of new agencies across many nations. As these safety-critical technologies, and others, grew in capability and prominence, governments foresaw their opportunities and impacts, opting for dedicated agencies of expertise. Of course, the scope and responsibilities of different agencies vary, and some have been more effective than others. Therefore, an important research task for AI policy development is to assess the design, scope, and implementation of previous safety-critical government agencies, pulling out key lessons that might apply to AI & Robotics. (I intend to write more on this subject in forthcoming blog posts)</span>
<b>What makes ‘effective AI Policy’?</b>
<span style="font-weight: 400;">This is perhaps the central question of AI policy. It speaks to the criteria that should be used to determine the merits of safety-critical AI policies. Specifically, </span><i><span style="font-weight: 400;">what constitutes effective policies in safety-critical AI? How will we know they’re effective?</span></i>
<span style="font-weight: 400;">The development of such criteria is a necessarily interdisciplinary task. It requires thoughtful input from a diversity of stakeholders and careful consideration of any recommendations. For the recommendation of any authoritative criteria assessing the effectiveness of AI policies could influence government actions. As </span><a href="http://www.drucker.institute/about-peter-f-drucker/"><span style="font-weight: 400;">Peter Drucker</span></a><span style="font-weight: 400;"> said: “<em>What gets measured gets managed</em>”.</span>
<span style="font-weight: 400;">The development of robust criteria needs to both sufficiently assess near-term policies but also provide insight into the projected longer-term impacts. It also needs to cater for the host of policies currently in place that directly, or indirectly, affect AI. These include </span><i><span style="font-weight: 400;">de facto</span></i><span style="font-weight: 400;"> policies such as privacy laws, intellectual property, and government research & development investment.[note]Supra note 6.[/note]</span><span style="font-weight: 400;"> While the scholarship in AI policy is thin, there have been some ideas put forth to advance research discussions.</span>
<i><span style="font-weight: 400;">Near-term Criteria</span></i>
<span style="font-weight: 400;">In their seminal research survey, the </span><a href="https://futureoflife.org/"><span style="font-weight: 400;">Future of Life Institute</span></a><span style="font-weight: 400;"> published ‘</span><a href="https://futureoflife.org/2016/01/25/a-survey-of-research-questions-for-robust-and-beneficial-ai/"><span style="font-weight: 400;">A survey of research questions for robust and beneficial AI</span></a><span style="font-weight: 400;">’. In this survey, the authors proposed the following points for consideration:[note]Future of Life Institute (2016) “<a href="https://futureoflife.org/data/documents/research_survey.pdf?x56934">A Survey of Research Questions for Robust and Beneficial AI</a>.”[/note]</span>
<ol>
<li style="font-weight: 400;"><b>Verifiability of compliance</b><span style="font-weight: 400;"> - How governments will know that any rules or policies are being adequately followed</span></li>
<li style="font-weight: 400;"><b>Enforceability</b><span style="font-weight: 400;"> - Ability of governments to institute rules or policies, and maintain accountability</span></li>
<li style="font-weight: 400;"><b>Ability to reduce AI risk</b></li>
<li style="font-weight: 400;"><b>Ability to avoid stifling desirable technology development and have other negative consequences</b> <b> </b>
<ol>
<li style="font-weight: 400;"><span style="font-weight: 400;">What happens when governments ban or restrict certain kinds of technological development? </span></li>
<li style="font-weight: 400;"><span style="font-weight: 400;">What happens when a certain kind of technological development is banned or restricted in one country but not in other countries where technological development sees heavy investment?</span></li>
</ol>
</li>
<li style="font-weight: 400;"><b>Adoptability</b><span style="font-weight: 400;"> - How well received the policies are from key stakeholders (the prospects of adoption increase when policy benefits those whose support is needed for implementation and when its merits can be effectively explained to decision-makers and opinion leaders)</span></li>
<li style="font-weight: 400;"><b>Ability to adapt over time to changing circumstances </b></li>
</ol>
<span style="font-weight: 400;">These policy criteria points share similarities with other safety-critical technologies, such as nuclear weapons, nanotechnology, and aviation. So, there are definitely lessons from the design and management of previous Science & Technology policies. A core challenge, however, is for policymakers to apply these lessons appropriately, recognise the unique challenges of AI, and develop policy responses accordingly. To do so effectively, will require a considered eye on the longer-term implications of any policy decisions.</span>
<i><span style="font-weight: 400;">Long-term Criteria</span></i>
<span style="font-weight: 400;">Professor Nick Bostrom et al. from the Future of Humanity Institute provides an overview of some key long-term policy considerations in their working paper: ‘</span><a href="https://www.fhi.ox.ac.uk/new-working-paper-policy-desiderata-in-the-development-of-machine-superintelligence/"><span style="font-weight: 400;">Policy Desiderata in the Development of Machine Superintelligence</span></a><span style="font-weight: 400;">’. This paper focuses on policy considerations for Artificial General Intelligence (AGI), which are considered ‘unique’ to AGI or ‘unusual’. </span>
<span style="font-weight: 400;">The paper distils the following desiderata:[note]Bostrom, Nick, Allan Dafoe, and Carrick Flynn. (2016) “<a href="https://www.fhi.ox.ac.uk/new-working-paper-policy-desiderata-in-the-development-of-machine-superintelligence/">Policy Desiderata in the Development of Machine Superintelligence</a>.” <i>Future of Humanity Institute</i>.[/note]</span>
<ul>
<li><b>Expeditious progress<span style="font-weight: 400;"> - Ensure that AI development and the path with a high probability to superintelligence is speedy. Socially beneficial products and applications are made widely available in a timely fashion.</span></b></li>
</ul>
<ul>
<li><strong>AI safety</strong><span style="font-weight: 400;"> - Techniques are developed to make it possible to ensure that advanced AIs act as intended.</span></li>
</ul>
<ul>
<li><strong>Conditional stablisation</strong><span style="font-weight: 400;"> (kill switch) - The ability to temporarily or permanently stablise the AI to avert catastrophe.</span></li>
</ul>
<ul>
<li><strong>Non-turbulence</strong><span style="font-weight: 400;"> (social stability) - The path avoids excessive efficiency losses from chaos and conflict. Political systems maintain stability and order, adapt successfully to change, and mitigate socially disruptive impacts.</span></li>
</ul>
<ul>
<li><strong>Universal benefit</strong><span style="font-weight: 400;"> - All humans who are alive at the transition of AGI get some share of the benefit, in compensation for the risk externality to which they were exposed.</span></li>
</ul>
<ul>
<li><strong>Magnanimity</strong><span style="font-weight: 400;"> (altruism and compassion) - A wide range of resource-satiable values (ones to which there is little objection aside from cost-based considerations), are realized if and when it becomes possible to do so using a minute fraction of total resources. This may encompass basic welfare provisions and income guarantees to all human individuals.</span></li>
<li><strong>Continuity</strong><span style="font-weight: 400;"> (fair resource allocation) - (i) maintain order and provide the institutional stability needed for actors to benefit from opportunities for trade behind the current veil of ignorance, including social safety nets; and (ii) prevent concentration and permutation from radically exceeding the levels implicit in the current social contract (basically, gross and growing resource inequality).</span></li>
<li style="font-weight: 400;"><b>Mind crime prevention</b><span style="font-weight: 400;"> - AI is governed in such a way that maltreatment of sentient digital minds is avoided or minimised.</span></li>
<li style="font-weight: 400;"><b>Population policy</b><span style="font-weight: 400;"> - Procreative choices, concerning what new beings bring into existence, are made in a coordinated manner and with sufficient foresight to avoid unwanted Malthusian dynamics and political erosion. (For e.g. what happens to population policy if humans become economically unproductive beings and governments are no longer incentivised to support them?)</span></li>
<li style="font-weight: 400;"><b>Responsibility and wisdom</b><span style="font-weight: 400;"> - The seminal applications of advanced AI are shaped by an agency (individual or distributed) that has an expansive sense of responsibility and the practical wisdom to see what needs to be done in radically unfamiliar circumstances.</span></li>
</ul>
<span style="font-weight: 400;">As the authors stress, these summarised criteria points aren’t ‘the answer’. Rather, they’re ideas to be built upon. What is clear, however, is that given the development speeds of AI, any near-term policies will need to closely consider its longer-term implications. For as the capacities of intelligent systems continue to compound, so too will their impacts. Therefore, policy decisions, whether deliberately or not, will affect the development of AI and its deployment throughout society. Establishing criteria to assess these policy decisions will help ensure that AI is safe and beneficial to humanity.</span>
<b>Conclusion</b>
<span style="font-weight: 400;">As AI becomes more pervasive, questions of policy will intensify. While shrouded in complexity, policymakers can help ensure the safe passage of AI that’s beneficial to humanity. As the representatives of the public, they have a responsibility to be informed and involved.</span>
<i><span style="font-weight: 400;">I hope this essay on some of the key issues and arguments in AI policy was helpful. I’d love any feedback. Feel free to get in touch either by commenting below or send me an email via: </span></i><a href="mailto:[email protected]"><i><span style="font-weight: 400;">[email protected]</span></i></a><i><span style="font-weight: 400;">.</span></i>]]></content:encoded>
<excerpt:encoded><![CDATA[]]></excerpt:encoded>
<wp:post_id>152</wp:post_id>
<wp:post_date><![CDATA[2017-09-22 05:02:39]]></wp:post_date>
<wp:post_date_gmt><![CDATA[2017-09-22 05:02:39]]></wp:post_date_gmt>
<wp:comment_status><![CDATA[open]]></wp:comment_status>
<wp:ping_status><![CDATA[open]]></wp:ping_status>
<wp:post_name><![CDATA[effective-policies-for-safety-critical-ai]]></wp:post_name>
<wp:status><![CDATA[publish]]></wp:status>
<wp:post_parent>0</wp:post_parent>
<wp:menu_order>0</wp:menu_order>
<wp:post_type><![CDATA[post]]></wp:post_type>
<wp:post_password><![CDATA[]]></wp:post_password>
<wp:is_sticky>0</wp:is_sticky>
<category domain="post_tag" nicename="ai-policy"><![CDATA[AI Policy]]></category>
<category domain="category" nicename="ai-policy"><![CDATA[AI Policy]]></category>
<category domain="post_tag" nicename="ai-safety"><![CDATA[AI Safety]]></category>
<category domain="category" nicename="artificial-intelligence"><![CDATA[Artificial Intelligence]]></category>
<category domain="post_tag" nicename="governments"><![CDATA[Governments]]></category>
<wp:postmeta>
<wp:meta_key><![CDATA[_themify_builder_settings_json]]></wp:meta_key>
<wp:meta_value><![CDATA[[{"row_order":"0","gutter":"gutter-default","equal_column_height":"","column_alignment":"col_align_top","cols":[{"column_order":"0","grid_class":"col-full first last","grid_width":"","modules":[],"styling":[]}],"styling":[]}]]]></wp:meta_value>
</wp:postmeta>
<wp:postmeta>
<wp:meta_key><![CDATA[_edit_last]]></wp:meta_key>
<wp:meta_value><![CDATA[1]]></wp:meta_value>
</wp:postmeta>
<wp:postmeta>
<wp:meta_key><![CDATA[builder_switch_frontend]]></wp:meta_key>
<wp:meta_value><![CDATA[0]]></wp:meta_value>
</wp:postmeta>
<wp:postmeta>
<wp:meta_key><![CDATA[_thumbnail_id]]></wp:meta_key>
<wp:meta_value><![CDATA[158]]></wp:meta_value>
</wp:postmeta>
</item>
<item>
<title>PhD Tools & Tactics</title>
<link>https://bitsandatoms.co/phd-tools-tactics/</link>
<pubDate>Sun, 15 Oct 2017 20:50:59 +0000</pubDate>
<dc:creator><![CDATA[[email protected]]]></dc:creator>
<guid isPermaLink="false">https://bitsandatoms.co/?p=165</guid>
<description></description>
<content:encoded><![CDATA[<span style="font-weight: 400;">It’s daunting to start a PhD. Looking out to three years of prolonged focus and intense concentration is intimidating. </span>
<span style="font-weight: 400;">It’s an exercise in truth and independent thought. Where thinking for yourself is the hardest task of all.</span>
<span style="font-weight: 400;">Thankfully, there’s a sea of resources to help researchers manage their dissertations. In fact, there’s so much available that it’s difficult to pick the best tools for the job. </span>
<span style="font-weight: 400;">When I started in August 2017, I spent a good amount of time researching products, trialing software systems, and speaking with other researchers about the tools they use to manage their research. After a few months in, here are some resources that I recommend:</span>
<h3><b>Document Management</b></h3>
<a href="https://paperpile.com/"><span style="font-weight: 400;">Paperpile</span></a>
<span style="font-weight: 400;">Having a solid and logical reference management system is essential for research. You’ll quickly accumulate hundreds (and eventually thousands) of books, papers, and other resources. Paperpile is great because you can save all the bibliographic information in one click, sort all your resources into folders and tags, and save documents to your Google Drive. There’s also a handy Google Chrome plugin and Google Search & Scholar integrations.</span>
<span style="font-weight: 400;">Whether you choose to use Paperpile or something like </span><a href="http://www.endnote.com"><span style="font-weight: 400;">EndNote</span></a><span style="font-weight: 400;"> is a matter of personal preference. I chose Paperpile because of how well it integrates with GSuite. The most important thing is to have some reference management system in place right at the beginning, ideally with folders and tags organised by keywords.</span>
<a href="https://evernote.com/"><span style="font-weight: 400;">Evernote</span></a>
<span style="font-weight: 400;">I’ve been using Evernote for a while and love it. It’s a great tool to organise all your documents and notes into folders and tags. The Evernote Web Clipper is also terrific. The real draw cards, however, are in the Premium Version. Here you can mark-up PDF docs, take photos of written notes to transform them into text, and there’s 10GB of storage per month. Students can also get 50% off the annual subscription.</span>
<h3><b>Writing and Collaboration</b></h3>
<a href="https://www.google.com.au/docs/about/"><span style="font-weight: 400;">Google Docs</span></a>
<span style="font-weight: 400;">Collaborating with your Supervisor or other researchers by Google Docs is an invaluable way to get feedback. They can add/edit/delete and provide comment on sections, all in real-time. Again, other software applications have caught-up on these features, so it’s a matter of personal preference. Just make sure you’re writing in a word processor with live update functionality. You don’t want to keep track of the 106th version of a doc!</span>
<h3><b>Productivity</b></h3>
<a href="https://www.rescuetime.com/"><span style="font-weight: 400;">RescueTime</span></a>
<span style="font-weight: 400;">There are so many things competing for our attention. But high-quality research demands extended periods of intense focus. RescueTime tracks how you’ve spent your time online. It breaks down this time into application categories and the proportion of time spent in focused applications (for e.g. </span><i><span style="font-weight: 400;">Design and Composition</span></i><span style="font-weight: 400;"> for Google Docs is considered ‘focused’ and ‘productive’, whereas </span><i><span style="font-weight: 400;">Social Media</span></i><span style="font-weight: 400;"> is considered ‘distracting’). While it doesn’t measure output, it does provide an insight into how you spend your time, which directly affects productivity.</span>
<h3><b>Project Management</b></h3>
<a href="https://basecamp.com/"><span style="font-weight: 400;">Basecamp</span></a>
<span style="font-weight: 400;">Like with any big project, there’s always the nagging question: ‘</span><i><span style="font-weight: 400;">Is this the best thing that I can be doing, right now?</span></i><span style="font-weight: 400;">’. I’ve found the project management software Basecamp to be useful for planning and managing my PhD. You’re able to set goals, schedule To-dos, link to documents, collaborate with peers, generate activity reports, and Automatic check-ins for personal accountability and reflection. The best news: Basecamp for Education is FREE for students! That’s enterprise project management software for NOTHING. </span>
<h3><b>Automatic Alerts</b></h3>
<a href="https://www.google.com.au/alerts"><span style="font-weight: 400;">Google Alerts</span></a>
<span style="font-weight: 400;">Once you’ve established your research keywords, setup Google Alerts and </span><a href="https://scholar.google.com/intl/en/scholar/help.html"><span style="font-weight: 400;">Google Scholar Alerts</span></a><span style="font-weight: 400;"> for the most recent published news and research. It saves a ton of time from trawling online, trying to keep up with the latest literature and industry developments.</span>
<h3><b>Clarity and Focus</b></h3>
<a href="https://www.calm.com/"><span style="font-weight: 400;">Calm</span></a>
<span style="font-weight: 400;">It’s easy to feel scattered and overwhelmed by the enormity of a PhD. But clarity of thought is essential to quality research. Daily meditation has been the most important practice I’ve established in the past year. Given the </span><a href="https://www.researchgate.net/publication/273774412_The_neuroscience_of_mindfulness_meditation"><span style="font-weight: 400;">neuroscientific evidence</span></a><span style="font-weight: 400;"> behind the practice, it’s unsurprising that it’s had such a significant effect.</span>
<span style="font-weight: 400;">I started with the free guided meditations from </span><a href="https://www.calm.com/"><span style="font-weight: 400;">Calm</span></a><span style="font-weight: 400;">, </span><a href="https://www.headspace.com/headspace-meditation-app"><span style="font-weight: 400;">Headspace</span></a><span style="font-weight: 400;">, and </span><a href="https://smilingmind.com.au/"><span style="font-weight: 400;">Smiling Mind</span></a><span style="font-weight: 400;">. I even paid for an annual subscription to Headspace, which was ok but it didn’t hook me in. I eventually found consistency in the free version of Calm using the ‘Timed Meditation’. There’s no guide, just the background noise of nature on loop and you select how long you want to meditate for. I’ve found 20 minutes first thing in the morning and 20 minutes before bed are great bookends to the day. I feel more present during the day, clearer in my thinking, and sleep better at night.</span>
<span style="font-weight: 400;">For those starting out, I’d recommend doing at least a few of the guided meditations to begin with. Following your breath for even 10 seconds is hard and requires training. So the guided beginner sessions definitely help to bring you back to the breath. </span>
<h3><b>Books</b></h3>
<a href="https://www.goodreads.com/book/show/25744928-deep-work"><span style="font-weight: 400;">Deep Work</span></a><span style="font-weight: 400;"> by </span><i><span style="font-weight: 400;">Cal Newport</span></i>
<span style="font-weight: 400;">This book changed the way I think about work and is the most important book I’ve read on productivity. Its central thesis is that to create things of value in society requires consistent and intense periods of focus. This is a far cry from the reactionary and distraction riddled practices of how most people work. Newport lays out strategies for how to cultivate a deep work ethic and discusses how these trained behaviours are becoming more needed and valued.</span>
<a href="https://www.goodreads.com/book/show/15999568-a-manual-for-writers-of-research-papers-theses-and-dissertations-eigh?rating=2"><span style="font-weight: 400;">A Manual for Writers of Research Papers, Theses, and Dissertations, Eighth Edition</span></a><span style="font-weight: 400;"> by </span><i><span style="font-weight: 400;">Kate L. Turabian</span></i>
<span style="font-weight: 400;">This is the best dissertation writer’s manual I’ve come across. It includes the full life-cycle of a research thesis; from establishing your topic to presenting and revising your final draft. The book is well structured and provides useful conceptual frameworks to organise and improve your research process and output.</span>
<i><span style="font-weight: 400;">I hope you found this useful. If anyone has any questions or suggestions for other resources, I’d love to hear them! Please comment below or email me at </span></i><a href="mailto:[email protected]"><i><span style="font-weight: 400;">[email protected]</span></i></a>]]></content:encoded>
<excerpt:encoded><![CDATA[]]></excerpt:encoded>
<wp:post_id>165</wp:post_id>
<wp:post_date><![CDATA[2017-10-15 20:50:59]]></wp:post_date>
<wp:post_date_gmt><![CDATA[2017-10-15 20:50:59]]></wp:post_date_gmt>
<wp:comment_status><![CDATA[open]]></wp:comment_status>
<wp:ping_status><![CDATA[open]]></wp:ping_status>
<wp:post_name><![CDATA[phd-tools-tactics]]></wp:post_name>
<wp:status><![CDATA[publish]]></wp:status>
<wp:post_parent>0</wp:post_parent>
<wp:menu_order>0</wp:menu_order>
<wp:post_type><![CDATA[post]]></wp:post_type>
<wp:post_password><![CDATA[]]></wp:post_password>
<wp:is_sticky>0</wp:is_sticky>
<category domain="post_tag" nicename="dissertation"><![CDATA[Dissertation]]></category>
<category domain="post_tag" nicename="phd"><![CDATA[PhD]]></category>
<category domain="post_tag" nicename="research-resources"><![CDATA[Research Resources]]></category>
<category domain="category" nicename="research-tools"><![CDATA[Research Tools]]></category>
<wp:postmeta>
<wp:meta_key><![CDATA[_edit_last]]></wp:meta_key>
<wp:meta_value><![CDATA[1]]></wp:meta_value>
</wp:postmeta>
<wp:postmeta>
<wp:meta_key><![CDATA[_thumbnail_id]]></wp:meta_key>
<wp:meta_value><![CDATA[166]]></wp:meta_value>
</wp:postmeta>
<wp:postmeta>
<wp:meta_key><![CDATA[builder_switch_frontend]]></wp:meta_key>
<wp:meta_value><![CDATA[0]]></wp:meta_value>
</wp:postmeta>
<wp:postmeta>
<wp:meta_key><![CDATA[_themify_builder_settings_json]]></wp:meta_key>
<wp:meta_value><![CDATA[[{"row_order":"0","gutter":"gutter-default","equal_column_height":"","column_alignment":"col_align_top","cols":[{"column_order":"0","grid_class":"col-full first last","grid_width":"","modules":[],"styling":[]}],"styling":[]}]]]></wp:meta_value>
</wp:postmeta>
</item>
<item>
<title>Recommendations: books, podcasts, videos, and other great stuff</title>
<link>https://bitsandatoms.co/october-recommendations/</link>
<pubDate>Wed, 18 Oct 2017 03:05:32 +0000</pubDate>
<dc:creator><![CDATA[[email protected]]]></dc:creator>
<guid isPermaLink="false">https://bitsandatoms.co/?p=172</guid>
<description></description>
<content:encoded><![CDATA[<h3><b>Books</b></h3>
<strong><a href="https://www.goodreads.com/book/show/34272565-life-3-0?ac=1&from_search=true">Life 3.0: Being Human in the Age of Artificial Intelligence</a> by<i> Max Tegmark</i></strong>
<span style="font-weight: 400;">This is one of the few books on AI that I’ve read that does justice to the core issues of Artificial General Intelligence (AGI), while still being accessible to a mainstream audience. Working as a Professor in MIT, Tegmark is one of the leading minds in AGI research. If you’re new to AI and interested in its implications, this is a great place to start. Once you’ve read this, then read Nick Bostrom’s ‘</span><a href="https://www.goodreads.com/book/show/20527133-superintelligence?ac=1&from_search=true"><span style="font-weight: 400;">Superintelligence: Paths, Dangers, Strategies</span></a><span style="font-weight: 400;">’.</span>
<strong><a href="https://www.goodreads.com/book/show/34536488-principles?ac=1&from_search=true">Principles: Life and Work</a> by <i>Ray Dalio</i></strong>
<span style="font-weight: 400;">Dalio is a giant in the investing world. His company, </span><a href="https://www.bridgewater.com/"><span style="font-weight: 400;">Bridgewater Associates</span></a><span style="font-weight: 400;">, is among the most successful Hedge Funds in history and they’re known for their unique management practices. Central to Bridgewater is the concept of an ‘Idea Meritocracy’; a practical philosophy where the best ideas rise to the top and are openly debated. To implement this concept, Bridgewater demands a culture of ‘radical truth’ and ‘radical transparency’, where almost all meetings are filmed, individual weaknesses are consistently identified, and truth and honesty are prioritised.</span>
<span style="font-weight: 400;">The book is a compilation and explanation of Dalio’s life and work principles. They’re the underpinnings that have guided the successes of both Dalio and Bridgewater. This is both the most practical and confronting self-improvement book I’ve read. While I found that certain principles evoked visceral responses, they were hard to argue with and difficult to dismiss. It’s changed my perspective and my actions. And it’s a book I plan to regularly revisit. </span>
<span style="font-weight: 400;">For those interested in Economics, also check out his research and video on ‘</span><a href="https://www.bridgewater.com/research-library/how-the-economic-machine-works/"><span style="font-weight: 400;">How the Economic Machine Works</span></a><span style="font-weight: 400;">’. As someone trained in Economics, it’s the most accessible and comprehensive description I’ve found. </span>
<strong><a href="https://www.goodreads.com/book/show/174713.The_Lessons_of_History?from_search=true">The Lessons of History</a> by <i>Will Durant & Ariel Durant</i></strong>
<span style="font-weight: 400;">This book was recommended by Ray Dalio as essential reading in a podcast. And it didn’t disappoint. Will & Ariel Durant are Pulitzer Prize historians, and they’re best known for their 11 volume series on </span><a href="https://www.goodreads.com/book/show/78159.The_Story_of_Civilization?from_search=true"><span style="font-weight: 400;">The Story of Civilization</span></a><span style="font-weight: 400;">. The husband and wife duo synthesised this series into 120 pages for </span><a href="https://www.goodreads.com/book/show/174713.The_Lessons_of_History?ac=1&from_search=true"><span style="font-weight: 400;">The Lessons of History</span></a><span style="font-weight: 400;">. Beautifully written and masterfully organised; it’s the best book I’ve read this year.</span>
<strong><a href="https://www.goodreads.com/book/show/20342617-just-mercy?ac=1&from_search=true">Just Mercy: A Story of Justice and Redemption</a> by <i>Bryan Stevenson</i></strong>
<span style="font-weight: 400;">Bryan Stevenson is an Attorney working on Death Row cases in Montgomery, Alabama. Stevenson has dedicated his professional life to overcoming the injustices of the US legal system and racial discrimination in America’s South. This important read is a coming of age of a talented and idealistic lawyer, leading him to the realisation that ‘</span><i><span style="font-weight: 400;">the opposite of poverty isn’t wealth, but justice</span></i><span style="font-weight: 400;">’.</span>
<span style="font-weight: 400;">Before reading the book, I highly recommend listening to the Ezra Klein Show podcast ‘Bryan Stevenson on why the opposite of poverty isn’t wealth, but justice’ on the 16th of May 2017 (</span><a href="https://soundcloud.com/ezra-klein-show/bryan-stevenson-on-why-the"><span style="font-weight: 400;">SoundCloud</span></a><span style="font-weight: 400;"> | </span><a href="https://itunes.apple.com/au/podcast/the-ezra-klein-show/id1081584611?mt=2"><span style="font-weight: 400;">Apple Podcasts</span></a><span style="font-weight: 400;">).</span>
<h3><b>Blogs & Research</b></h3>
<strong><a href="https://unenumerated.blogspot.com">Unenumerated</a> by <i>Nick Szabo</i></strong>
<span style="font-weight: 400;">I’ve been trying to wrap my head around cryptocurrencies of late. Like with anything complex, the more you learn, the more you realise how little you know. Cryptocurrencies, like Bitcoin and Ethereum, are prime examples. Nick Szabo’s blog has the most comprehensive selection of long-form articles on cryptocurrency and blockchain technologies I’ve come across.</span>
<strong><a href="https://www.bridgewater.com/resources/bwam032217.pdf">Populism: The Phenomenon</a> by <i>Bridgewater Associates</i></strong>
<span style="font-weight: 400;">Populist politics has surged across major nations. Trump, Brexit, Germany’s AfD, and many others have either taken office or are growing in popularity. Bridgewater shows that Populism is at its highest levels since the 1930’s. It also draws parallels to today with the economic conditions of the 30’s. When projecting economic futures under these conditions, the most important thing to monitor is how conflict is handled. The research is a unique and independent thought-piece in a time of rising political populism.</span>
<h3><b>Videos</b></h3>
<strong><a href="https://www.youtube.com/playlist?list=PLfc2WtGuVPdmhYaQjd449k-YeY71fiaFp">A Brief History of Humankind</a> by <i>Dr. Yuval Noah Harari</i></strong>
<span style="font-weight: 400;">Harari is one of my favourite authors, and I believe </span><a href="https://www.goodreads.com/book/show/23692271-sapiens?ac=1&from_search=true"><span style="font-weight: 400;">Sapiens</span></a><span style="font-weight: 400;"> and </span><a href="https://www.goodreads.com/book/show/31138556-homo-deus"><span style="font-weight: 400;">Homo Deus: A Brief History of Tomorrow</span></a><span style="font-weight: 400;"> are two of the most important books written in the 21st century. I wanted more, so I’ve been watching his online course, </span><a href="https://www.youtube.com/playlist?list=PLfc2WtGuVPdmhYaQjd449k-YeY71fiaFp"><span style="font-weight: 400;">A Brief History of Humankind</span></a><span style="font-weight: 400;">, on YouTube. I’ve found it to be a great way of reinforcing what I’ve already read in his books. </span>
<strong><a href="https://www.netflix.com/au/title/80057883">Abstract: The Art of Design</a> on Netflix</strong>
<span style="font-weight: 400;">This series follows leading designers across disciplines to provide insight into their methods and routines. From architecture to footwear, we’re given a glimpse into the creative genius of the world’s most prominent designers.</span>
<strong><a href="https://vimeo.com/223928856">The Endurance Test: The 1000 Days</a> by <i>Ivan Olita</i></strong>
<span style="font-weight: 400;">Kaihōgyō is a pilgrimage lasting 1,000 days performed by the Tendai Buddhists monks of Mount Hiei, Japan to achieve enlightenment. This beautifully shot 6 minute short film shows the training and endurance behind the 7 year journey, of which only 46 men have completed since 1885.</span>
<h3><b>Podcasts</b></h3>
<a href="https://tim.blog/2017/10/14/walter-isaacson/"><span style="font-weight: 400;">Lessons from Steve Jobs, Leonardo da Vinci, and Ben Franklin</span></a><span style="font-weight: 400;"> by </span><i><span style="font-weight: 400;">The Tim Ferriss Show</span></i><span style="font-weight: 400;">, 13th October 2017 (</span><a href="https://www.stitcher.com/podcast/the-tim-ferriss-show"><span style="font-weight: 400;">Stitcher</span></a><span style="font-weight: 400;"> | </span><a href="https://itunes.apple.com/us/podcast/the-tim-ferriss-show/id863897795?mt=2"><span style="font-weight: 400;">Apple Podcasts</span></a><span style="font-weight: 400;">)</span>
<span style="font-weight: 400;">In the interview, Walter Isaacson states that ‘</span><i><span style="font-weight: 400;">The best way to teach human history is to tell the stories of the people that have made it</span></i><span style="font-weight: 400;">.’ He’s dedicated the best part of his career to this. And he’s now done the same for Leonardo da Vinci. This wide-ranging interview highlights the common threads and key difference behind some of history’s greatest minds.</span>
<h3><b>Some Recent AI Resources</b></h3>
<ul>
<li style="font-weight: 400;"><a href="https://www.gov.uk/government/uploads/system/uploads/attachment_data/file/652097/Growing_the_artificial_intelligence_industry_in_the_UK.pdf"><span style="font-weight: 400;">GROWING THE ARTIFICIAL INTELLIGENCE INDUSTRY IN THE UK</span></a><span style="font-weight: 400;"> [Government Paper]</span></li>
<li style="font-weight: 400;"><a href="https://www.youtube.com/watch?v=bsuvM1jO-4w&feature=youtu.be"><span style="font-weight: 400;">Deep RL Bootcamp Frontiers Lecture I: Recent Advances, Frontiers and Future of Deep RL</span></a><span style="font-weight: 400;"> [YouTube video]</span></li>
<li style="font-weight: 400;"><a href="http://www.parliamentlive.tv/Event/Index/073717ca-484b-4015-bd10-f847cea3f249"><span style="font-weight: 400;">Professor Nick Bostrom presents evidence to the House of Lords Select Committee on Artificial intelligence in Westminster</span></a><span style="font-weight: 400;"> [Video]</span></li>
<li style="font-weight: 400;"><a href="https://intelligence.org/2017/10/13/fire-alarm/">There’s No Fire Alarm for Artificial General Intelligence</a> [Blog by MIRI]</li>
</ul>]]></content:encoded>
<excerpt:encoded><![CDATA[]]></excerpt:encoded>
<wp:post_id>172</wp:post_id>
<wp:post_date><![CDATA[2017-10-18 03:05:32]]></wp:post_date>
<wp:post_date_gmt><![CDATA[2017-10-18 03:05:32]]></wp:post_date_gmt>
<wp:comment_status><![CDATA[open]]></wp:comment_status>
<wp:ping_status><![CDATA[open]]></wp:ping_status>
<wp:post_name><![CDATA[october-recommendations]]></wp:post_name>
<wp:status><![CDATA[publish]]></wp:status>
<wp:post_parent>0</wp:post_parent>
<wp:menu_order>0</wp:menu_order>
<wp:post_type><![CDATA[post]]></wp:post_type>
<wp:post_password><![CDATA[]]></wp:post_password>
<wp:is_sticky>0</wp:is_sticky>
<category domain="post_tag" nicename="ai-resources"><![CDATA[AI Resources]]></category>
<category domain="post_tag" nicename="blogs"><![CDATA[Blogs]]></category>
<category domain="post_tag" nicename="books"><![CDATA[Books]]></category>
<category domain="category" nicename="monthly-recommendations"><![CDATA[Monthly Recommendations]]></category>
<category domain="post_tag" nicename="podcasts"><![CDATA[Podcasts]]></category>
<category domain="post_tag" nicename="recommendations"><![CDATA[Recommendations]]></category>
<category domain="post_tag" nicename="research"><![CDATA[Research]]></category>
<category domain="post_tag" nicename="videos"><![CDATA[Videos]]></category>
<wp:postmeta>
<wp:meta_key><![CDATA[_edit_last]]></wp:meta_key>
<wp:meta_value><![CDATA[1]]></wp:meta_value>
</wp:postmeta>
<wp:postmeta>
<wp:meta_key><![CDATA[_thumbnail_id]]></wp:meta_key>
<wp:meta_value><![CDATA[173]]></wp:meta_value>
</wp:postmeta>
<wp:postmeta>
<wp:meta_key><![CDATA[_yoast_wpseo_content_score]]></wp:meta_key>
<wp:meta_value><![CDATA[30]]></wp:meta_value>
</wp:postmeta>
<wp:postmeta>
<wp:meta_key><![CDATA[_yoast_wpseo_primary_category]]></wp:meta_key>
<wp:meta_value><![CDATA[16]]></wp:meta_value>
</wp:postmeta>
<wp:postmeta>
<wp:meta_key><![CDATA[_themify_builder_settings_json]]></wp:meta_key>
<wp:meta_value><![CDATA[[{"row_order":"0","gutter":"gutter-default","equal_column_height":"","column_alignment":"col_align_top","cols":[{"column_order":"0","grid_class":"col-full first last","grid_width":"","modules":[],"styling":[]}],"styling":[]}]]]></wp:meta_value>
</wp:postmeta>
<wp:postmeta>
<wp:meta_key><![CDATA[builder_switch_frontend]]></wp:meta_key>
<wp:meta_value><![CDATA[0]]></wp:meta_value>
</wp:postmeta>
</item>
<item>
<title>A Primer on Reinforcement Learning</title>
<link>https://bitsandatoms.co/primer-reinforcement-learning/</link>
<pubDate>Fri, 03 Nov 2017 04:47:03 +0000</pubDate>
<dc:creator><![CDATA[[email protected]]]></dc:creator>
<guid isPermaLink="false">https://bitsandatoms.co/?p=179</guid>
<description></description>
<content:encoded><![CDATA[<span style="font-weight: 400;">The rate of development in AI continues at a rapid pace. And a couple of weeks ago we saw another important milestone. DeepMind </span><a href="http://nature.com/articles/doi:10.1038/nature24270"><span style="font-weight: 400;">published</span></a><span style="font-weight: 400;"> their latest developments of </span><a href="https://deepmind.com/research/alphago/"><span style="font-weight: 400;">AlphaGo</span></a><span style="font-weight: 400;">, a computer program designed to play the ancient Chinese game of </span><a href="https://en.wikipedia.org/wiki/Go_(game)"><span style="font-weight: 400;">Go</span></a><span style="font-weight: 400;"> at superhuman levels. Go is incredibly complex. It has a possible 10 to the power of 170 configurations. That’s more than the number of atoms in the known universe!</span>
<h3><span style="font-weight: 400;">AlphaGo Zero</span></h3>
<span style="font-weight: 400;">DeepMind’s latest version, </span><a href="https://deepmind.com/blog/alphago-zero-learning-scratch/"><span style="font-weight: 400;">AlphaGo Zero</span></a><span style="font-weight: 400;">, exceeded the performance all previous versions. It did so by using a novel form of self-play Reinforcement Learning (a subset of Machine Learning), which I’ll explain in more detail later.</span>
<span style="font-weight: 400;">AlphaGo Zero represents an important advancement not only because of its performance but also because of its method. All </span><span style="font-weight: 400;">previous versions</span><span style="font-weight: 400;">, like </span><i><span style="font-weight: 400;">AlphaGo Fan</span></i><span style="font-weight: 400;"> in 2015 and </span><i><span style="font-weight: 400;">AlphaGo Lee</span></i><span style="font-weight: 400;"> in 2016 that beat the European and World Champions respectively, were trained on the data of thousands of human games. They used two Deep Neural Networks to output move probabilities (policy network) and position evaluations (value network). Once trained, these networks were combined with a lookahead search (Monte-Carlo Tree Search) to evaluate move positions in the tree.[note]Silver, David, Julian Schrittwieser, Karen Simonyan, Ioannis Antonoglou, Aja Huang, Arthur Guez, Thomas Hubert, et al. 2017. “Mastering the Game of Go without Human Knowledge.” <i>Nature</i> 550 (7676): 354–59.[/note]</span>
<span style="font-weight: 400;">The latest AlphaGo Zero version differs in four important ways:</span>
<ul>
<li><b>Trained solely by self-play reinforcement learning: <span style="font-weight: 400;">Basically, AlphaGo Zero became its own teacher with no training data provided. It started from random self-play and trained itself from first principles to become the world’s most advanced Go player.</span></b></li>
<li><strong>It only used the black and white stones on the Board as input features:</strong><span style="font-weight: 400;"> Previous versions had included human-engineered input features to help guide the program.</span></li>
<li><strong>Combines policy and value networks into a single neural network:</strong><span style="font-weight: 400;"> AlphaGo Zero simplifies the network architecture from two neural networks to one, improving the efficiency of the structure.</span></li>
<li><strong>Simpler tree search:</strong><span style="font-weight: 400;"> It relies upon the single neural network to evaluate moves and positions, and doesn’t perform Monte-Carlo ‘rollouts’ to predict which player will win based on the current board configuration.</span></li>
</ul>
<span style="font-weight: 400;">The most important developments, however, are the algorithmic improvements. As discussed by David Silver, this shows that ‘</span><i><span style="font-weight: 400;">Algorithms matter much more than either compute or data availability… we used an order of magnitude less computation than we did with previous versions of Alpha Go and yet it was able to perform at a much higher level due to using much more principled algorithms.</span></i><span style="font-weight: 400;">’[note]Hassabis, Demis, and David Silver. 2017. “AlphaGo Zero: Learning from Scratch.” <i>DeepMind</i>. October 18. <a href="https://deepmind.com/blog/alphago-zero-learning-scratch/">https://deepmind.com/blog/alphago-zero-learning-scratch/</a>.[/note]</span>
<span style="font-weight: 400;">The image below[note]Supra note 2.[/note]</span><span style="font-weight: 400;"> shows that after 3 days AlphaGo Zero surpassed the version that beat the best human player in the world. And after 40 days, AlphaGo Zero was able to beat all previous versions of AlphaGo to become the strongest Go program in the world. (NB: </span><i><span style="font-weight: 400;">the Elo Rating on the vertical axis is a widely used measure of player performance in games such as Go and Chess</span></i><span style="font-weight: 400;">)</span>
<img class="aligncenter " src="http://blogs.discovermagazine.com/d-brief/files/2017/10/AlphaGo-Zero-Training-Time.gif" width="731" height="389" />
<span style="font-weight: 400;">This impressive feat was achieved through a novel form of Reinforcement Learning (RL). AlphaGo Zero was able to teach itself by always having an opponent at just the right level, which was calibrated exactly to its current level of performance. </span>
<span style="font-weight: 400;">Before we dive into an overview of how RL works, it’s important first to understand the key concepts of how machines learn and their infrastructure.</span>
<h3><span style="font-weight: 400;">Brief Overview of Machine Learning</span></h3>
<span style="font-weight: 400;">Machine Learning (ML) is a subset of AI and it’s principally concerned with teaching machines to learn on their own. ML systems learn from data to autonomously make predictions and/or decisions.[note]Li, Yuxi. 2017. “Deep Reinforcement Learning: An Overview.” arXiv [cs.LG]. <a href="http://arxiv.org/abs/1701.07274">http://arxiv.org/abs/1701.07274</a>.[/note]</span><span style="font-weight: 400;"> The three main categories of ML are Supervised, Unsupervised, and Reinforcement Learning.[note]</span><span style="font-weight: 400;">Jordan, M. I. and Mitchell, T. 2015. “Machine learning: Trends, perspectives, and prospects”. Science, 349(6245):255–260.</span>
<span style="font-weight: 400;">While these three subsets of ML place some structure around the discipline, lots of the current research explores the intersection of these subsets. For example, semi-supervised learning makes use of unlabeled data to augment labelled data in a supervised learning context.[/note]</span>
<img class="aligncenter wp-image-186" src="https://bitsandatoms.co/wp-content/uploads/2017/11/ML_branches.jpg" alt="" width="366" height="347" />
<p style="text-align: center;"><span style="font-weight: 400;">Branches of Machine Learning</span></p>
<p style="text-align: center;"><span style="font-weight: 400;">Source: </span><a href="https://www.youtube.com/watch?v=2pWv7GOvuf0&t=2767s"><span style="font-weight: 400;">RL Course by David Silver - Lecture 1: Introduction to Reinforcement Learning</span></a></p>
<span style="font-weight: 400;">I’ll give a brief overview of Supervised, Unsupervised, and also Deep Learning before I dig into RL.</span>
<h3><span style="font-weight: 400;">Supervised Learning</span></h3>
<span style="font-weight: 400;">Supervised Learning uses a set of correctly labelled training data that are provided by a knowledgeable external supervisor (i.e. a human).[note]Richard S. Sutton and Andrew G. Barto. 2016. <i>Reinforcement Learning: An Introduction</i>. The MIT Press.[/note]</span><span style="font-weight: 400;"> It essentially trains networks by practising on training data to predict and/or make decisions where the correct answer is already known. The purpose of Supervised Learning is to learn by analysing these vast reams of labelled data to extrapolate representations or generalise trends, so it can correctly categorise situations not present in the training set.</span>
<span style="font-weight: 400;"> The two tasks of Supervised Learning are:</span>
<ul>
<li style="font-weight: 400;"><strong>Classification:</strong><span style="font-weight: 400;"> Correctly assign class labels to unseen instances. For e.g. correctly identify cats in YouTube videos</span></li>
<li style="font-weight: 400;"><strong>Regression:</strong><span style="font-weight: 400;"> Predict continuous values based on inputs. For e.g. predicting house prices based on location, square footage, number of bedrooms etc.</span></li>
</ul>
<span style="font-weight: 400;">Training the network to learn these representations is achieved by splitting a labelled dataset into two piles. The first pile is the </span><i><span style="font-weight: 400;">training data</span></i><span style="font-weight: 400;"> (usually ~80% of the dataset) where the output (i.e the answer) is known. During training, the model makes predictions for each instance and receives feedback based on the outputs in the training data. This feedback is quantified by the algorithm to determine ‘how close’ the prediction was to the known output, which is referred to as the </span><i><span style="font-weight: 400;">cost function</span></i><span style="font-weight: 400;"> or </span><i><span style="font-weight: 400;">utility function.</span></i><span style="font-weight: 400;"> The changes to the function are fed back to the network to modify the strength of connections between the nodes to optimise predictability of the output.</span>
<span style="font-weight: 400;">The second pile is the </span><i><span style="font-weight: 400;">validation data</span></i><span style="font-weight: 400;"> where the output is taken away to test the accuracy of the function. This is the final stage of verification before you apply the model to unseen data.</span>
<span style="font-weight: 400;">Once the learned function gets to a point of acceptable accuracy (for e.g. 90%+), then the model can start to be applied to unseen data where the output is unknown but new input values are provided.</span>
<h3><span style="font-weight: 400;">Unsupervised Learning</span></h3>
<span style="font-weight: 400;">Unsupervised Learning attempts to find structure hidden in collections of unlabelled data.[note]Supra note 6.[/note]</span><span style="font-weight: 400;"> It’s basically trying to find patterns from the input set. Unsupervised Learning achieves this in two main ways:</span>
<ol>
<li style="font-weight: 400;"><strong>Clustering:</strong><span style="font-weight: 400;"> organising data into groups based on similarities and relationships.</span></li>
<li style="font-weight: 400;"><strong>Anomaly Detection:</strong><span style="font-weight: 400;"> identifying the outliers to reduce the complexity of data, while keeping a relevant structure as much as possible.</span></li>
</ol>
<span style="font-weight: 400;">A cool recent development in unsupervised learning is </span><a href="https://en.wikipedia.org/wiki/Generative_adversarial_network"><span style="font-weight: 400;">Generative Adversarial Networks</span></a><span style="font-weight: 400;"> (GANs). Introduced by </span><a href="https://en.wikipedia.org/wiki/Ian_Goodfellow"><span style="font-weight: 400;">Ian Goodfellow</span></a><span style="font-weight: 400;"> in 2014, GANs use a system of two neural networks to compete against each other to produce an output. It works by having one network, called the ‘Generator’, which is tasked with generating data that's designed to try and trick the other network, called the ‘Discriminator’. For example, GANs have been tasked with creating computer-generated images that are indistinguishable from human-generated images. The Generator network (think ‘Artist) creates an image and tries to trick the Discriminator network (think ‘Art Critic’) into identifying it as real. This is a hard problem because there aren’t solid measures of success or universal metrics in art. So GANs pit two neural networks against each other to make sense of this unstructured data.</span>
<h3><span style="font-weight: 400;">Deep Learning & Neural Networks</span></h3>
<span style="font-weight: 400;">Deep Learning (DL) is concerned with learning data representations, as opposed to task-specific algorithms. DL works on a wide variety of AI problems, such as Natural Language Processing and Computer Vision, and has exploded in popularity over the past five or so years. It’s a great tool because it helps to make sense of complex information. For example, with natural language problems, there’s huge variance in vocabulary within the same language; DL can be an effective means to learn function approximators to make sense of this messy information, like correctly interpreting accents. </span>
<span style="font-weight: 400;">The most important thing about DL is that it uses deep Neural Networks to automatically find low-dimensional representations (features) of high-dimensional data (e.g. voice, images etc.).[note]Kai Arulkumaran, Marc Peter Deisenroth, Miles Brundage, and Anil Anthony Bharath. 2017. “A Brief Survey of Deep Reinforcement Learning.” <i>arXiv</i>. <a href="http://arxiv.org/abs/1708.05866v2">http://arxiv.org/abs/1708.05866v2</a>.[/note]</span><span style="font-weight: 400;"> It does this using neural network architectures to make hierarchical representations, which are loosely inspired by the ways that neurons in the brain work. Unlike many other Machine Learning algorithms, which just have an input and output layers with manual feature engineering, DL has one or more ‘hidden layers’ in its architecture. This enables DL to be more precise in its representations.</span>
<span style="font-weight: 400;">While there are many variants of neural networks, here’s a simple example of a multi-layer perceptron neural network (otherwise known as Feedforward Neural Networks):</span>
<img class="aligncenter size-full wp-image-187" src="https://bitsandatoms.co/wp-content/uploads/2017/11/Feedforward-Neural-Network.png" alt="" width="531" height="305" />
<p style="text-align: center;"><span style="font-weight: 400;">Source: </span><a href="https://jsalatas.ictpro.gr/wp-content/uploads/2011/09/3_3.png"><span style="font-weight: 400;">John Salatas</span></a></p>
<span style="font-weight: 400;">A neural network is made up of different layers of nodes (kind of like neurons). These nodes are connected to the next layer of nodes (kind of like how neurons are connected by axons). In the above example, raw inputs are fed into the Input layer, where the nodes receive data. For example, this data could be something like a greyscale image of pixel brightness that’s represented by a number between 0 and 1. The data then progressively flows through the layers from left to right, activating the nodes and refining feature representations along the way. Each neuron assigns a weighting to its input, which is basically saying how ‘correct’ it is relative to the task it’s performing (or how strongly those nodes are connected). The final output is then determined by the total of these weightings, which is the probability of each class being correctly labelled. </span>
<span style="font-weight: 400;">For a more detailed intro to DL, check out this great visual representation of DL and Neural Networks by </span><a href="https://www.youtube.com/channel/UCYO_jab_esuFRV4b17AJtAw"><span style="font-weight: 400;">3Blue1Brown</span></a><span style="font-weight: 400;">:</span>
[embed]https://www.youtube.com/watch?v=aircAruvnKk[/embed]
<h3><span style="font-weight: 400;">Reinforcement Learning</span></h3>
<span style="font-weight: 400;">Reinforcement Learning (RL) is about learning what decisions to make in an environment to maximise a reward function.[note]Supra note 6.[/note]</span><span style="font-weight: 400;"> The learning agent does this through ‘trial and error’, receiving feedback on the amount of reward that a particular action yields. Think of this like the simple ‘hotter and colder’ game. The game involves one person searching for an object and another person instructing them how ‘hot’ (close) or ‘cold’ (far) they are from attaining the object. </span>
<span style="font-weight: 400;">Unlike in Supervised Learning, a RL agent isn’t trained on labelled examples around the correct actions to take. RL is also different from Unsupervised Learning because it isn’t trying to find a hidden structure in unlabelled data. While uncovering patterns and relationships in data might be helpful to a learning agent, and there are examples of </span><a href="https://deepmind.com/blog/reinforcement-learning-unsupervised-auxiliary-tasks/"><span style="font-weight: 400;">combining these two approaches</span></a><span style="font-weight: 400;">, RL is principally concerned with maximising its reward function. A RL agent learns from direct interaction with an environment, without relying on complete models of an environment or strong supervision.</span>
<span style="font-weight: 400;">Of all the forms of Machine Learning, RL represents the closest form to how humans learn, and RL is predicted to play a crucial role in the quest for General Artificial Intelligence.[note]Li, Yuxi. 2017. “Deep Reinforcement Learning: An Overview.” <i>arXiv [cs.LG]</i>. <a href="http://arxiv.org/abs/1701.07274">http://arxiv.org/abs/1701.07274</a>; Silver, D. (2016). Deep reinforcement learning, a tutorial at ICML 2016. http://icml.cc/ 2016/tutorials/deep_rl_tutorial.pdf; Supra note 6.[/note]</span>
<h3><span style="font-weight: 400;">Elements of RL</span></h3>
<span style="font-weight: 400;">As described by Sutton and Barto,[note]Supra note 6, pg. 6.[/note]</span><span style="font-weight: 400;"> there are four main elements of a RL system, which are not always required, but may or may not be used. I’ll unpack some of the key concepts behind these elements further, but for now, here’s a summary:</span>
<ol>
<li style="font-weight: 400;"><strong>Policy:</strong><span style="font-weight: 400;"> A policy defines the behaviour of an agent and how the agent picks its actions. It does so by mapping from perceived </span><i><span style="font-weight: 400;">states</span></i><span style="font-weight: 400;"> of an </span><i><span style="font-weight: 400;">environment</span></i><span style="font-weight: 400;"> to </span><i><span style="font-weight: 400;">actions</span></i><span style="font-weight: 400;"> to be taken when in those states. The policy element is central to RL as it essentially develops the ‘rules’ for an agent to determine what actions it should take. In general, the environment is stochastic, which means that the next environment has a degree of randomness.</span></li>
<li style="font-weight: 400;"><strong>Reward signal</strong><span style="font-weight: 400;"><strong> (or function):</strong> The reward signal defines the goal of the RL problem. A reward signal is sent from the environment at every action that a RL agent takes (referred to as </span><i><span style="font-weight: 400;">time step</span></i><span style="font-weight: 400;">). This reward signal determines what's considered ‘good’ and ‘bad’ events, and is the primary basis for altering a policy. For example, if the policy of a RL agent selects an action that yields a low reward, then the policy may be altered to improve the reward signal in the future. The RL agent can influence the reward signal directly through its actions and indirectly through altering the environment’s state. But it can’t change the problem it’s been tasked with.</span></li>
<li style="font-weight: 400;"><strong>Value function:</strong><span style="font-weight: 400;"> The value function is a prediction of expected future reward. As opposed to the more ‘immediate gratification’ of the reward signal, the value function is concerned with the accumulation of rewards over the longer-term. It’s principally concerned with ‘how good’ it is for an agent to be in a particular state and/or perform a particular action. It does this by estimating how much reward the agent can expect to get if it takes an action in a corresponding state. Unsurprisingly, this is extremely difficult to do. While the reward signal occurs in a tight feedback loop directly from the environment, values must be consistently re-estimated to optimise the long-run reward. Therefore, having efficient and effective methods of value estimation is arguably the most important part of RL algorithms.</span></li>
<li style="font-weight: 400;"><strong>Model of the environment:</strong><span style="font-weight: 400;"> Models help determine how the agent ‘thinks’ the environment works and predicting what it will do next. </span><i><span style="font-weight: 400;">Model-based </span></i><span style="font-weight: 400;">methods of RL are used for </span><i><span style="font-weight: 400;">planning</span></i><span style="font-weight: 400;"> to help the agent to understand the environment and to predict the best actions to take. Based on a given state and action, the model predicts the resultant next state and next reward. </span></li>
</ol>
<h3><span style="font-weight: 400;">Markov Decision Processes - </span><i><span style="font-weight: 400;">Formalising the RL Problem</span></i></h3>
<span style="font-weight: 400;">Markov Decision Processes (MDPs) are basically ways to frame the most important aspects of the problem faced by a learning agent that’s interacting with an environment to achieve a goal. In its simplest form, the MDP formulation includes (1) sensing the environment, (2) performing an action, and (3) trying to achieve a specified goal. So, any method that’s well suited to solving this combination of problems is considered to be a reinforcement learning method.[note]Supra note 6, pg. 47.[/note]</span>
<span style="font-weight: 400;">Let’s use the example of a robot fitted with a camera that’s learning to pick-up randomly shaped objects from one bucket with the goal of placing them into another defined bucket. The reward might be a value of +1 for every object successfully placed in the goal bucket. Negative rewards could also be allocated for actions that incorrectly place objects.</span>
<img class="aligncenter wp-image-188" src="https://bitsandatoms.co/wp-content/uploads/2017/11/Deep-Learning-for-Robots.jpg" alt="" width="585" height="378" />
<p style="text-align: center;">Source: <a href="http://robohub.org/wp-content/uploads/2017/06/Deep-Learning-for-Robots.jpg">RoboHub</a></p>
<span style="font-weight: 400;">The robot begins at its </span><i><span style="font-weight: 400;">agent state</span></i><span style="font-weight: 400;"> within its surrounding environment </span><b><i>S</i></b><span style="font-weight: 400;">. And each interaction represents a sequence of </span><i><span style="font-weight: 400;">time steps</span></i><span style="font-weight: 400;">, </span><b><i>t = 0, 1, 2… n</i></b><span style="font-weight: 400;">.</span>