File size: 79,564 Bytes
421fea8
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
1254
1255
1256
1257
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
1268
1269
1270
1271
1272
1273
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283
1284
1285
1286
1287
1288
1289
1290
1291
1292
1293
1294
1295
1296
1297
1298
1299
1300
1301
1302
1303
1304
1305
1306
1307
1308
1309
1310
1311
1312
1313
1314
1315
1316
1317
1318
1319
1320
1321
1322
1323
1324
1325
1326
1327
1328
1329
1330
1331
1332
1333
1334
1335
1336
1337
1338
1339
1340
1341
1342
1343
1344
1345
1346
1347
1348
1349
1350
1351
1352
1353
1354
1355
1356
1357
1358
1359
1360
1361
1362
1363
1364
1365
1366
1367
1368
1369
1370
1371
1372
1373
1374
1375
1376
1377
1378
1379
1380
1381
1382
1383
1384
1385
1386
1387
1388
1389
1390
1391
1392
1393
1394
1395
1396
1397
1398
1399
1400
1401
1402
1403
1404
1405
1406
1407
1408
1409
1410
1411
1412
1413
1414
1415
1416
1417
1418
1419
1420
1421
1422
1423
1424
1425
1426
1427
1428
1429
1430
1431
1432
1433
1434
1435
1436
1437
1438
1439
1440
1441
1442
1443
1444
1445
1446
1447
1448
1449
1450
1451
1452
1453
1454
1455
1456
1457
1458
1459
1460
1461
1462
1463
1464
1465
1466
1467
1468
1469
1470
1471
1472
1473
1474
1475
1476
1477
1478
1479
1480
1481
1482
1483
1484
1485
1486
1487
1488
1489
1490
1491
1492
1493
1494
1495
1496
1497
1498
1499
1500
1501
1502
1503
1504
1505
1506
1507
1508
1509
1510
1511
1512
1513
1514
1515
1516
1517
1518
1519
1520
1521
1522
1523
1524
1525
1526
1527
1528
1529
1530
1531
1532
1533
1534
1535
1536
1537
1538
1539
1540
1541
1542
1543
1544
1545
1546
1547
1548
1549
1550
1551
1552
1553
1554
1555
1556
1557
1558
1559
1560
1561
1562
1563
1564
1565
1566
1567
1568
1569
1570
1571
1572
1573
1574
1575
1576
1577
1578
1579
1580
1581
1582
1583
1584
1585
1586
1587
1588
1589
1590
1591
1592
1593
1594
1595
1596
1597
1598
1599
text,start,duration
good morning everybody David Shapiro,0.42,3.6
here with another video,2.639,4.62
so today's video uh it started off as,4.02,5.28
one thing I wanted to primarily talk,7.259,4.861
about epistemic convergence uh but It,9.3,4.32
ultimately ended up being a little bit,12.12,2.54
more,13.62,2.82
all-encompassing so I'm going to,14.66,3.94
introduce a few new terms but we are,16.44,4.08
going to cover cover uh epistemic,18.6,4.259
convergence and a few other things,20.52,4.62
uh real quick before we dive into the,22.859,3.781
video just want to do a quick plug for,25.14,3.719
my patreon uh all tears get you access,26.64,4.32
to the private Discord server and then I,28.859,4.2
have a few higher tiers that uh come,30.96,4.32
with a one-on-one conversations and that,33.059,4.201
sort of thing so anyways back to the,35.28,5.16
video so first I wanted to share with,37.26,5.639
you guys uh the universal model of,40.44,4.98
Robotics so it has it's basically three,42.899,5.16
steps input processing and output or,45.42,4.38
sensing processing and controlling as,48.059,3.241
this Graphics shows,49.8,4.38
now this is the most basic cognitive,51.3,4.5
architecture that you can come up with,54.18,4.199
for artificial general intelligence it,55.8,4.439
needs input from the outside world from,58.379,3.66
the environment of some kind whether,60.239,3.3
it's a virtual environment digital,62.039,3.841
environment physical environment or,63.539,4.62
whatever cyber cybernetic environment,65.88,4.68
and then it needs some kind of internal,68.159,4.981
processing that includes memory task,70.56,4.2
construction executive function,73.14,4.519
cognitive control that sort of stuff,74.76,5.82
learning is another internal process and,77.659,5.381
then finally controlling or output it,80.58,4.859
needs to do something to act on the,83.04,5.1
world or its environment whether that's,85.439,4.801
just putting out you know text in a in,88.14,3.6
the form of a chat bot or if it's got,90.24,4.86
robotic hands that sort of thing so when,91.74,5.519
I talk about artificial general,95.1,3.9
intelligence being a system it's never,97.259,4.141
going to just be a model right even if,99.0,4.14
you have the most sophisticated model in,101.4,3.6
the world all that it's doing is the,103.14,3.839
processing part you also need the,105.0,4.799
sensing and controlling aspect and but,106.979,4.621
even above and beyond that each,109.799,4.081
components is going to be much more,111.6,4.32
complicated,113.88,4.08
so before we get into the rest of the,115.92,4.08
video I also want to talk about the form,117.96,4.26
factors that AGI is going to take so we,120.0,3.899
just established the simplest kind of,122.22,3.96
cognitive architecture but then there's,123.899,3.781
other things to consider because when,126.18,3.66
you think of AGI you might think of some,127.68,4.08
nebulous entity like Skynet but where,129.84,3.6
does it physically live,131.76,3.42
what is the hardware what is the,133.44,4.2
software where where is it physically,135.18,4.02
located because it's not magic right,137.64,3.179
it's not going to just run in the dirt,139.2,3.179
or something like that it needs to,140.819,4.381
actually have Hardware to run on so,142.379,5.761
there's three overarching categories,145.2,5.28
that I came up with so first is cloud,148.14,5.28
AGI so Cloud AGI is this is the stuff,150.48,4.38
that's gonna one it's going to be,153.42,3.06
created first just because of the amount,154.86,3.54
of compute and power available in data,156.48,5.22
centers so this is uh Enterprise grade,158.4,7.74
or data center grade AGI systems they,161.7,7.14
are in specialized buildings all over,166.14,5.22
the world but one of the biggest,168.84,4.38
constraints here is that there's limited,171.36,3.48
location and it takes a while to build,173.22,3.54
data centers right one of the things,174.84,3.96
that I think it was uh it was Elon,176.76,4.32
musker or Sam Altman said that you know,178.8,4.92
there are going to be limitations as to,181.08,5.28
the rate at which AGI can proliferate,183.72,5.519
namely the the rate at which we can,186.36,5.94
produce chips and also the rate at which,189.239,5.22
as I think Sam malt and said the you,192.3,4.439
know the concrete has to dry for data,194.459,3.121
centers,196.739,5.64
so this is uh one form factor that AGI,197.58,7.019
will take in terms of the the storage,202.379,4.801
the servers the network components that,204.599,4.86
will exist inside data centers so one,207.18,3.839
thing I wanted to watch it say is watch,209.459,3.841
out for uh fortified data centers these,211.019,3.841
are ones that are put in bunkers or if,213.3,3.659
you put Sam sites on top of it so that,214.86,4.26
you can't shut them down uh that was,216.959,3.42
kind of tongue-in-cheek I'm not actually,219.12,3.36
advocating for bombing data centers at,220.379,4.681
least not yet the next form factor is,222.48,5.16
Edge AGI so this is stuff that is going,225.06,4.22
to run in,227.64,3.84
self-contained servers that you can,229.28,4.9
basically plug in anywhere they're going,231.48,6.539
to be you know desktop size maybe larger,234.18,5.82
but the point is that pretty much all,238.019,3.481
you need is power and internet you don't,240.0,4.319
need a specialized building and they can,241.5,4.62
be moved on trucks they can be put in,244.319,3.961
ships airplanes that sort of stuff,246.12,3.72
because you can't really airlift an,248.28,4.019
entire data center so basically Edge is,249.84,4.619
something is just one size down from,252.299,3.481
data center you don't need a specialized,254.459,2.52
building you don't need specialized,255.78,3.66
cooling they can run anywhere,256.979,3.72
um and they're so in that respect,259.44,2.94
they're more portable but they're not,260.699,4.741
necessarily going to be as powerful at,262.38,5.039
least or not as energy intensive and,265.44,5.039
energy dense as a data center or a cloud,267.419,4.021
Center,270.479,4.141
and then finally ambulatory AGI this is,271.44,5.46
the embodied stuff such as C-3PO and,274.62,3.98
Commander data which I have imaged here,276.9,4.98
they're self-contained meaning that all,278.6,5.14
the systems that they need are within,281.88,4.319
their chassis within their robotic body,283.74,5.519
and they can move on their own so that's,286.199,4.5
basically the difference between an edge,289.259,5.101
AGI and an ambulatory AGI is uh they,290.699,5.401
might have roughly the same components,294.36,4.14
but it's one is accompanied with a,296.1,6.12
robotic uh chassis now one thing to keep,298.5,6.18
in mind is that all of these things are,302.22,4.86
intrinsically networkable meaning they,304.68,4.32
can communicate over digital networks,307.08,4.559
whether it's Wi-Fi or you know fiber,309.0,5.16
optic backbone networks or even you know,311.639,4.141
Satellite Communication like starlink,314.16,3.9
now that's that doesn't necessarily have,315.78,4.08
to be true because remember the model of,318.06,4.44
AGI is input processing and output that,319.86,5.76
input that input could be just eyes and,322.5,4.979
ears cameras and microphones that input,325.62,4.26
could also be network connections from,327.479,4.621
outside meaning that they could,329.88,4.44
communicate directly with each other via,332.1,5.28
you know like IRC or whatever so just,334.32,4.86
wanted to say that there are different,337.38,3.72
form factors that we should expect AGI,339.18,2.94
to take,341.1,3.3
with different trade-offs so one,342.12,5.22
advantage of ambulatory uh AGI you know,344.4,6.12
yes they will have less power uh and by,347.34,5.88
power I mean computational power but,350.52,5.28
they have the ability to go anywhere do,353.22,6.66
anything kind of like URI uh now that,355.8,6.959
being said the the amount of compute,359.88,4.8
resources that can be crammed into Data,362.759,3.66
Centers basically means that you can,364.68,4.38
puppet you know millions or billions of,366.419,5.161
peripheral robots rather than having it,369.06,4.44
fully self-contained and in a previous,371.58,3.899
video I talked about how we're likely to,373.5,4.639
see hybrid systems where you have,375.479,5.22
semi-autonomous peripherals that have,378.139,4.481
some intelligence but not a whole lot of,380.699,4.201
intelligence and you see this in movies,382.62,4.919
like Will Smith's iRobot as well as the,384.9,4.38
Matrix where the the drones the,387.539,3.78
squiddies and the Matrix they're,389.28,3.479
semi-autonomous but they are still,391.319,3.301
centrally controlled by a much more,392.759,4.021
powerful intelligence so you're probably,394.62,3.72
not going to see it all one or the other,396.78,3.06
you're probably going to see hybrids,398.34,3.54
where you've got peripheral robots that,399.84,4.02
are either fully autonomous or,401.88,4.92
semi-autonomous or puppeted by stronger,403.86,4.98
Central intelligences that being said,406.8,4.26
you can also create droids there's no,408.84,4.5
reason that we could not create fully,411.06,4.38
self-contained machines that don't,413.34,4.74
really have any network connectivity,415.44,4.86
um to the to other machines,418.08,3.899
that being said they would be at a,420.3,3.78
distinct disadvantage and what I mean by,421.979,4.5
that is that if you create swarm,424.08,4.8
intelligence or Wireless federations of,426.479,4.741
machines they can perform cognitive,428.88,6.24
offload or share computational resources,431.22,6.66
so for instance rather than and this is,435.12,4.44
how the Geth work in Mass Effect by the,437.88,4.439
way so rather than have every single,439.56,6.06
machine have to think about the entire,442.319,5.94
plan the entire strategy most of them,445.62,4.919
Focus only on their primary task and,448.259,5.401
then any surplus compute computational,450.539,6.121
power they have is dedicated towards you,453.66,5.52
know running algorithms for for the big,456.66,4.68
the big brain the hive mind,459.18,4.56
this is all hypothetical but one thing,461.34,4.32
that I want to point out is that many,463.74,3.54
many many machines work like this,465.66,4.56
already and what I mean by that is the,467.28,4.38
simplest version that many people are,470.22,3.9
probably aware of is if you have like,471.66,3.84
Bluetooth speakers or smart speakers,474.12,3.6
like Sonos or whatever those form a,475.5,5.819
wireless Federation uh ditto for like,477.72,5.28
your Amazon Alexa's and other things,481.319,4.261
like that those intrinsically form mesh,483.0,4.259
networks or Wireless federations meaning,485.58,3.48
that they can work together and,487.259,4.321
communicate now when you add artificial,489.06,5.46
intelligence to that then they can share,491.58,4.559
thinking and messaging and that sort of,494.52,2.64
stuff so that's what I mean by,496.139,4.801
federations or or wireless networks of,497.16,6.599
of AI okay so now you're familiar with,500.94,4.74
the background of how you know some of,503.759,5.041
the systemic aspects of it there's a few,505.68,4.919
default metrics of power so when I say,508.8,3.359
power I don't necessarily just mean,510.599,3.721
electricity although certainly all of,512.159,3.961
these things do require electricity to,514.32,2.82
run,516.12,4.26
so first is processing power so for,517.14,5.1
instance you might hear the term flops,520.38,4.2
which is floating Point operations per,522.24,7.68
second uh you also hear CPU GPU TPU and,524.58,7.02
then there's parallel parallelization,529.92,4.02
which means that you have many of these,531.6,5.1
things working together so processing,533.94,4.62
power is one component of the total,536.7,4.68
amount of power in the hardware layer so,538.56,4.44
this is all strictly Hardware layer I'm,541.38,3.48
not talking about parameter models,543.0,4.74
because I I don't really care about how,544.86,4.68
many parameters a model has there's lots,547.74,4.02
of ways to make intelligent machines,549.54,4.2
deep neural networks are currently the,551.76,4.079
best way but we're also discovering,553.74,3.719
efficiencies where you can kind of pair,555.839,3.661
them down you can distill them and make,557.459,3.541
them more efficient meaning that you can,559.5,3.36
on the same piece of Hardware you can,561.0,4.8
run more of them in parallel or you can,562.86,5.52
run one much faster so the underlying,565.8,3.96
Hardware is still going to be the,568.38,3.8
primary bottleneck or primary constraint,569.76,5.1
all else considered,572.18,5.68
uh memory so this is Ram it also,574.86,5.36
includes memory accelerators or caching,577.86,5.46
storage has to do with bulk data your,580.22,4.9
databases your archives your backups,583.32,4.44
this is when you say like hard drive or,585.12,4.74
SSD or you know storage area network,587.76,4.259
that sort of thing and then networking,589.86,4.68
is the the uplinks and downlinks this is,592.019,4.741
the the fiber optic connections the,594.54,3.72
wireless connections the satellite,596.76,3.9
connections that sort of thing so these,598.26,4.38
are the kind of the the rudimentary,600.66,4.799
parts that all AGI are going to run on,602.64,5.879
uh and this is just the brains too this,605.459,4.621
is not the peripherals this is not the,608.519,3.241
robots but this is what's going to,610.08,4.259
dictate or constrain how fast it is now,611.76,5.94
again like I said uh different neural,614.339,5.221
networks are going to operate at,617.7,3.42
different efficiencies so for instance,619.56,5.82
uh you know gpt4 is out now gpt5 might,621.12,6.96
be the same size it might be bigger but,625.38,4.82
then we're also finding open source,628.08,6.06
research like the Orca alpaca llama that,630.2,5.92
are getting like ninety percent of the,634.14,4.56
performance but at like one tenth or one,636.12,4.98
hundredth of the size and so you have a,638.7,4.199
trade-off of intelligence and versus,641.1,4.56
speed and power and we'll talk more a,642.899,5.221
lot more about that in the future of,645.66,4.739
this video at near the middle and end of,648.12,4.2
this video about how trading off,650.399,4.801
intelligence for Speed is often a more,652.32,4.56
advantageous strategy and how this,655.2,4.199
figures into solving the control problem,656.88,5.28
and solving alignment,659.399,4.981
um okay so we kind of set the stage as,662.16,4.98
as to how AGI is probably going to look,664.38,5.639
so let's talk about the early ecosystem,667.14,6.18
of AGI so in the coming years we're,670.019,4.921
going to be building millions and then,673.32,3.24
billions of autonomous and,674.94,4.5
semi-autonomous agents so at first these,676.56,4.2
agents are going to be purely digital,679.44,3.44
right you know a,680.76,4.56
semi-autonomous slack bot a,682.88,4.54
semi-autonomous Discord bot people are,685.32,4.32
already building these right and some of,687.42,3.72
them have the ability to modify their,689.64,2.879
own code some of them have the ability,691.14,3.78
to learn many of them don't most of them,692.519,4.32
use frozen llms in the background,694.92,3.659
meaning that they're that their,696.839,4.201
cognitive capacity is pretty much capped,698.579,4.981
by its backing model,701.04,5.76
now that being said as these agents,703.56,5.279
become more autonomous they go from,706.8,4.2
semi-autonomous to autonomous this will,708.839,4.201
create a competitive landscape,711.0,4.86
and what I mean by that is that humans,713.04,5.28
will have the ability to build and,715.86,5.159
destroy these models for basically,718.32,4.62
arbitrary reasons because you want to or,721.019,3.781
because you don't like it or whatever,722.94,4.26
so that means that we will be selecting,724.8,5.52
those agents those uh those models those,727.2,6.18
llms and those pieces of software that,730.32,4.32
are going to be more helpful more,733.38,3.6
productive and more aligned so this,734.64,4.08
creates selective pressure basically,736.98,3.299
saying that there's going to be a,738.72,2.88
variety there's going to be millions or,740.279,3.24
billions of Agents out there some of,741.6,3.9
them are going to get the ax and some of,743.519,4.141
them are going to be selected to say hey,745.5,4.079
we like we like you we're going to keep,747.66,2.88
you around,749.579,3.901
so there's a few off the cuff selective,750.54,4.799
pressures that we can imagine basically,753.48,3.84
why do you choose an app right why do,755.339,3.361
you choose to use an app why do you,757.32,3.54
choose to uninstall an app that's kind,758.7,3.36
of the level that we're talking about,760.86,3.9
here so first is functional utility how,762.06,4.019
useful is it,764.76,3.6
how much does it help you is it fast,766.079,3.541
enough does it have a good user,768.36,3.5
experience is the user interface,769.62,5.88
created correctly is it adding value to,771.86,7.06
your life is it worth using the second,775.5,6.079
part is speed and efficiency,778.92,6.06
basically if something takes four weeks,781.579,5.141
to give you a good answer but another,784.98,4.2
thing takes 10 minutes even if it's not,786.72,4.559
quite as good that speed is going to be,789.18,4.14
super super valuable but then there's,791.279,4.081
also energetic efficiency and cost,793.32,4.94
efficiency more often than not,795.36,5.34
individuals and businesses will choose,798.26,5.56
the solution that is good enough but,800.7,5.04
also much cheaper it doesn't have to be,803.82,4.019
perfect it just has to be good enough,805.74,4.86
and then finally apparent alignment and,807.839,5.101
so I use the the word apparent alignment,810.6,4.38
to basically mean things that appear to,812.94,4.199
be tame things that appear to be user,814.98,4.56
friendly uh and this is what uh tools,817.139,5.521
like rlhf do which one thing that rlhf,819.54,4.68
does is like wolves which we'll talk,822.66,4.619
about in a second are the rlhf,824.22,4.2
reinforcement learning with human,827.279,4.68
feedback forces gpt4 to dumb itself down,828.42,6.539
so that it better serves us uh and that,831.959,4.56
makes us feel safe because it's,834.959,4.32
basically pretending to be more like us,836.519,5.641
to speak on our terms and to mimic our,839.279,4.981
level of intelligence now that being,842.16,3.419
said,844.26,2.579
um one thing that I do want to point out,845.579,4.021
is that gpd4 the underlying model is,846.839,4.74
superior to anything that we have seen,849.6,5.239
in the public every version of chat GPT,851.579,6.021
has basically been kind of a little bit,854.839,6.101
hamstrung so we shall we say uh from the,857.6,6.82
the total capacity of gpt4,860.94,5.699
so what I call this is domestication and,864.42,4.8
supplication think of dogs and wolves,866.639,5.101
this little pomeranian descended from,869.22,4.98
wolves wolves used to be apex predators,871.74,4.62
wolves are also are much more,874.2,4.02
intelligent than dogs,876.36,6.419
so when you look at the early days of,878.22,6.84
AGI when we still have the off switch,882.779,3.781
and we have the power to delete,885.06,3.12
everything,886.56,3.899
we should expect some of the following,888.18,5.7
evolutionary pressures to kind of shape,890.459,6.421
the way that AGI evolves and adapts so,893.88,5.519
first we'll probably be selecting for,896.88,4.079
machines that are okay with being turned,899.399,4.201
off in the early days you don't,900.959,4.38
necessarily want your toaster fighting,903.6,3.479
with you when when you're done you know,905.339,3.3
toasting your bread it's time for it to,907.079,3.661
turn off and so we're probably going to,908.639,3.781
select machines and architectures and,910.74,3.779
models that are more or less okay with,912.42,3.539
being switched off that they don't have,914.519,3.801
a sense of death or a fear of death,915.959,4.62
we're also going to select machines that,918.32,4.6
are more eager to please just the same,920.579,5.281
way that with uh dogs have been bred and,922.92,5.159
selected to be very very eager to please,925.86,3.599
us,928.079,3.181
we're also going to select machines that,929.459,3.781
don't fall into the uncanny valley and,931.26,3.48
so what I mean by that is the uncanny,933.24,4.02
valley of when you're interacting with a,934.74,4.56
machine that you sense is an alien,937.26,3.78
intelligence it will make you very very,939.3,4.2
deeply uncomfortable as an autistic,941.04,4.799
person as someone who is neurodiverse I,943.5,4.139
have to modulate the way that I speak,945.839,4.68
and act around neurotypical people,947.639,5.401
because I fall into the same uncanny,950.519,5.94
valley right and some uh some CEOs out,953.04,5.039
there get teased for this for instance,956.459,3.24
Mark Zuckerberg I don't know if he's,958.079,3.841
actually autistic but he certainly pings,959.699,4.44
that radar where it's like okay he,961.92,3.539
obviously does not think the way that,964.139,3.361
the rest of us do and he also behaves,965.459,4.44
differently so Mark Zuckerberg like many,967.5,5.1
of us uh people on the Spectrum kind of,969.899,4.56
fall into that uncanny valley again I,972.6,4.44
don't know but uh,974.459,5.221
he certainly looks uh he he plays the,977.04,6.06
part so but the idea is that when you,979.68,5.339
interact with something that give that,983.1,4.26
kind of gives you the heebie-jeebies you,985.019,3.901
don't like it,987.36,4.14
now that being said we will still select,988.92,4.68
machines that are smarter to a certain,991.5,3.66
degree because you don't want something,993.6,3.659
to be too smart but you do also want it,995.16,3.599
to be smart enough to be very very,997.259,2.64
useful,998.759,3.181
another selective pressure is that we're,999.899,3.421
going to choose things that are stable,1001.94,3.06
robust and resilient so remember when,1003.32,3.12
Bing first came out and it was,1005.0,3.66
completely unhinged you could get it,1006.44,3.78
into you could coax it into like,1008.66,4.619
threatening you and you know threatening,1010.22,4.859
to take over the world and you know,1013.279,3.721
threatening to see you all kinds of,1015.079,4.2
crazy stuff so obviously that version,1017.0,4.44
got shut down really quick,1019.279,4.321
um you're also going to select uh models,1021.44,4.139
and agents that are more resilient,1023.6,3.839
against those kinds of adversarially,1025.579,3.061
attacks,1027.439,3.0
um whether they are accidental right you,1028.64,2.88
don't want something to be mentally,1030.439,3.48
unstable just on its own right like Bing,1031.52,6.0
was originally uh or Tay tweets but you,1033.919,5.101
also want it to be resilient against,1037.52,4.74
being manipulated by other hostile,1039.02,5.039
actors because imagine that your,1042.26,3.9
personal AI assistance just becomes,1044.059,4.02
unhinged one day because a hacker,1046.16,3.72
somewhere was messing with it so,1048.079,3.48
Security will be one of the selective,1049.88,3.0
pressures,1051.559,3.841
likewise you'll you'll as part of The,1052.88,4.38
Uncanny Valley thing you're going to,1055.4,2.94
select things that are more,1057.26,2.82
comprehensible to us that are better at,1058.34,3.78
explaining themselves to us so that,1060.08,3.78
includes transparency emotional,1062.12,5.04
intelligence and so on uh and then again,1063.86,5.46
apparent alignment things that's that,1067.16,4.74
that uh don't kind of trigger your,1069.32,4.62
existential dread because there have,1071.9,4.56
been times for instance where I've been,1073.94,5.94
working with chat GPT uh on the API side,1076.46,5.16
and kind of giving it different sets of,1079.88,4.62
instructions and even just a slight,1081.62,5.58
misalignment between how I approach,1084.5,4.86
moral problems and how this model,1087.2,4.56
approaches moral problems are really,1089.36,5.939
deeply unsettling and so it's like there,1091.76,5.58
there's been a few times where it's like,1095.299,3.301
I'm working with this thing and I'm,1097.34,3.78
building a semi-autonomous chat Bots and,1098.6,3.959
it's like I understand it's reasoning,1101.12,3.419
but it's like oh that's really cringe,1102.559,3.901
and it kind of scares me,1104.539,4.38
um so in that respect it's like let's,1106.46,4.14
change this model so that it's not quite,1108.919,2.821
so scary,1110.6,2.52
and I'm saying that this is possible,1111.74,4.22
today that if you use the chat GPT API,1113.12,5.4
you can give it programming you can give,1115.96,5.26
it reasoning and goals uh and and,1118.52,4.86
patterns of thought that are already,1121.22,4.5
already on the kind of in the midst of,1123.38,3.84
that uncanny valley,1125.72,4.38
uh then you can uh we'll also select for,1127.22,4.92
things that are more uh docile so,1130.1,3.6
basically how dogs you know you can pet,1132.14,2.7
them you can wrestle with them and,1133.7,2.339
they're probably not going to eat your,1134.84,3.719
face uh plastic and so things that are,1136.039,4.921
changeable or adaptable and Cooperative,1138.559,4.201
those are other things that we're going,1140.96,4.74
to select for so basically dogs are,1142.76,5.159
dumber than wolves and the reason for,1145.7,3.96
this is what I call capability,1147.919,4.14
equilibrium which will unpack more in,1149.66,5.28
the in a few slides but the the very,1152.059,4.801
short version of capability equilibrium,1154.94,3.96
is that your intellect must be equal to,1156.86,3.84
the task and if your intellect is above,1158.9,3.779
the task there's no advantage and in,1160.7,3.54
fact there can be disadvantages because,1162.679,3.901
of the costs associated with higher,1164.24,4.02
intelligence,1166.58,4.02
okay so I've talked about this idea,1168.26,5.76
plenty instrumental convergence uh this,1170.6,6.0
was coined by Nick Bostrom in 2003 who,1174.02,4.44
is a philosopher,1176.6,3.959
um the very short version is that,1178.46,3.959
regardless of the terminal goals or main,1180.559,5.101
objectives that a machine has uh AGI,1182.419,5.161
will likely pursue intermediate or,1185.66,4.259
instrumental goals or basically other,1187.58,3.9
stuff that it needs in order to meet,1189.919,5.401
those other ends so whatever like let's,1191.48,5.88
say you give an AGI the goal of like,1195.32,5.46
getting them getting a a spacecraft to,1197.36,5.58
Alpha Centauri well it's going to need a,1200.78,3.66
laundry list of other stuff to do that,1202.94,3.72
it's going to need resources like power,1204.44,6.0
materials electricity data it's going to,1206.66,6.06
need self-preservation because if the,1210.44,4.44
machine goes offline it will realize,1212.72,5.1
that is a failure State and so we'll try,1214.88,4.98
and avoid those failure conditions by,1217.82,3.96
preserving its own existence,1219.86,3.48
another thing is that it will probably,1221.78,3.06
decide that it needs self-improvement,1223.34,4.079
because if it realizes that its current,1224.84,5.219
capability its current capacity is not,1227.419,4.681
equal to the task if it's too dumb it's,1230.059,3.541
going to say okay well I need to raise,1232.1,3.06
my intelligence so that I'm equal to,1233.6,2.76
that task,1235.16,3.66
now that being said Nick boster makes,1236.36,4.559
quite a few uh assumptions about the way,1238.82,5.099
that AGI will work so for instance he,1240.919,5.521
kind of imagines that um AGI is going to,1243.919,3.961
be very single-minded and somewhat,1246.44,4.32
monolithic uh basically mindlessly,1247.88,4.98
pursuing one goal which I would actually,1250.76,3.96
classify this as a middle intelligence,1252.86,4.319
rather than a high intelligence AGI and,1254.72,3.9
we'll talk about that in a little bit as,1257.179,2.941
well,1258.62,3.6
he also assumes that it's going to lack,1260.12,3.9
other forces or competitive pressures,1262.22,4.26
and that these uh might exist in a,1264.02,5.399
vacuum basically that resource,1266.48,4.5
acquisition and self-preservation and,1269.419,4.021
self-improvement are going to exist in,1270.98,5.16
in the absence of other forces or,1273.44,5.22
pressures such as competitive pressures,1276.14,4.38
or internal pressures which I will talk,1278.66,2.879
about more,1280.52,3.42
and finally that they will lack a higher,1281.539,5.361
purpose or the ability to be completely,1283.94,5.82
self-determining so basically what I,1286.9,7.0
mean by that is that okay yes once a,1289.76,6.48
machine is intelligent enough it can you,1293.9,3.72
know you can say like hey I want you to,1296.24,3.72
get us to Alpha Centauri and the AG I,1297.62,4.02
might say like okay whatever I don't,1299.96,2.88
think that's a good goal so I'm going to,1301.64,4.26
choose my own goal uh which that being,1302.84,5.459
said even if AGI become fully autonomous,1305.9,4.019
and you know kind of give a flip us the,1308.299,3.36
bird they're probably still going to,1309.919,3.421
benefit from some convergence which,1311.659,4.621
we'll talk about as well uh now what I,1313.34,4.62
want to point out is that there is a,1316.28,4.74
huge parallel between evolutionary,1317.96,5.04
pressures and selective pressures and,1321.02,4.08
this instrumental convergence basically,1323.0,4.86
all life forms all organisms have have,1325.1,5.4
converged on a few basic principles such,1327.86,4.98
as get energy somehow right there's,1330.5,4.02
autotrophs which make their own energy,1332.84,3.839
plants and there's heterotrophs which,1334.52,5.039
take energy from other uh creatures,1336.679,6.24
uh they through either predation or,1339.559,4.86
consuming you know plant matter or,1342.919,2.341
whatever,1344.419,4.081
uh so when you operate in a competitive,1345.26,5.039
environment there's there's going to be,1348.5,3.6
convergence around certain strategies,1350.299,4.081
this is true for evolution and this will,1352.1,4.68
also be true more or less with some,1354.38,4.56
variances in the competitive environment,1356.78,4.92
between intelligent machines that being,1358.94,4.38
said because they have a fundamentally,1361.7,4.02
different substrate there will be we,1363.32,3.839
should anticipate that there will be,1365.72,2.88
some differences,1367.159,4.201
between organisms the way that organisms,1368.6,4.74
evolve and the way that machines evolve,1371.36,3.78
not the least of which is that machines,1373.34,3.6
can rewrite their own source code we,1375.14,3.24
cannot rewrite our own source code at,1376.94,2.58
least not,1378.38,3.0
um not in a hurry it takes us quite a,1379.52,3.48
long time,1381.38,4.679
okay so the idea that one of the ideas,1383.0,4.559
that I'm introducing and I've been,1386.059,2.821
talking about this for a while is,1387.559,4.561
epistemic Convergence so instrumental,1388.88,4.86
convergence talks about the objective,1392.12,3.539
behaviors and strategies that machines,1393.74,4.74
adopt epistemic convergence is well let,1395.659,3.921
me just read you the definition,1398.48,3.059
epistemic convergence is the principle,1399.58,4.2
that within any given information domain,1401.539,4.561
sufficiently sophisticated intelligent,1403.78,4.72
agents given adequate time and data will,1406.1,3.959
progressively develop more precise,1408.5,3.299
accurate and efficient models of that,1410.059,4.081
domain these models aim to mirror the,1411.799,3.781
inherent structures principles and,1414.14,3.539
relationships within that domain over,1415.58,3.78
time the process of learning testing and,1417.679,3.841
refining understanding will lead these,1419.36,4.02
agents towards a shared comprehension of,1421.52,4.74
the Dom domain's fundamental truths in,1423.38,5.299
other words to put it more simply,1426.26,4.919
intelligent entities tend to think alike,1428.679,4.061
especially when they are operating in,1431.179,3.841
the same competitive space,1432.74,5.88
so you and I All Humans we operate on,1435.02,5.58
planet Earth in the universe in the,1438.62,4.74
Milky Way galaxy because of that similar,1440.6,5.12
context scientists all over the world,1443.36,4.62
repeatedly come to the same conclusions,1445.72,5.079
even when there are boundaries such as,1447.98,5.22
linguistic and cultural differences and,1450.799,4.441
this was most starkly seen during the,1453.2,4.2
Cold war between uh America and the,1455.24,3.2
Soviet Union,1457.4,4.08
whereby scientists independently whether,1458.44,4.239
it was nuclear physicist or,1461.48,3.199
astrophysicist or whatever,1462.679,5.041
rocket Engineers came to the same exact,1464.679,4.781
conclusions about the way that the,1467.72,3.78
Universe worked and also found the same,1469.46,6.06
optimization uh uh patterns even though,1471.5,5.58
there was no communication between them,1475.52,3.899
and so epistemic convergence there's,1477.08,4.86
obviously uh evidence of that happening,1479.419,4.081
because humans we have the same,1481.94,4.619
fundamental Hardware right we're all the,1483.5,5.64
same species and so therefore you have,1486.559,4.801
similarities between the agents now that,1489.14,3.659
being said,1491.36,4.02
uh there is also evidence of epistemic,1492.799,5.641
convergence between between species and,1495.38,5.52
so what I mean by that is even animals,1498.44,4.44
that have a very very different taxonomy,1500.9,4.38
such as ravens and crows and octopuses,1502.88,5.4
they all still demonstrate very similar,1505.28,5.22
problem solving strategies even though,1508.28,4.86
that octopuses have a very decentralized,1510.5,4.38
cognition that a lot of their cognition,1513.14,4.08
occurs in their arms for instance you,1514.88,4.14
can't get much more alien from us than,1517.22,3.72
that they still adopt very similar,1519.02,3.72
problem-solving strategies and learning,1520.94,3.42
strategies that we do,1522.74,4.62
uh again despite the fact that they are,1524.36,4.62
they live underwater they have a very,1527.36,3.78
different body plan so on and so forth,1528.98,4.92
so I personally suspect that there is a,1531.14,4.56
tremendous amount of evidence for,1533.9,4.2
epistemic convergence and we should we,1535.7,4.979
should expect epistemic convergence and,1538.1,5.939
encourage epistemic convergence uh and,1540.679,5.341
for reasons that I'll go over uh later,1544.039,5.061
in the video but basically,1546.02,6.42
AI agents will we should expect and help,1549.1,6.52
them to arrive at similar conclusions in,1552.44,4.26
the long run,1555.62,3.48
now let's talk about these evolutionary,1556.7,5.4
uh niches that will be developed at,1559.1,5.76
least in the in in the um uh the short,1562.1,4.26
term near term,1564.86,4.199
and what I mean by this is segments,1566.36,4.439
market segments where we will be,1569.059,4.62
deploying intelligent AGI systems so,1570.799,5.88
first is domestic uh personal and,1573.679,4.62
consumer grade stuff so this is going to,1576.679,4.021
be the AGI running on your MacBook this,1578.299,4.74
is going to be the AGI running in your,1580.7,5.339
kitchen uh these have a relatively,1583.039,7.201
benign set of tasks and also that uh,1586.039,6.781
that capability equilibrium is going to,1590.24,5.28
be uh pretty pretty low you only need to,1592.82,5.459
be so smart to cook dinner right this is,1595.52,5.1
not going to be you know the the AGI,1598.279,3.841
running in your microwave is not going,1600.62,3.419
to be working on quantum physics or,1602.12,3.84
Global economics,1604.039,4.321
now the next level up is going to be,1605.96,3.959
corporate and Enterprise so these are,1608.36,3.72
going to be these are going to be AGI,1609.919,3.721
systems that are tasks with solving,1612.08,4.14
relatively complex problems running,1613.64,5.34
entire companies Regulatory Compliance,1616.22,6.24
uh you know making SEC filings that sort,1618.98,6.66
of stuff uh CEOs digital CEOs digital,1622.46,5.339
Boards of directors uh the creative,1625.64,4.919
aspect of finding Market opportunities,1627.799,5.701
so this the intellectual challenge of,1630.559,5.461
those of that scale of problems is that,1633.5,5.88
much higher meaning that it would in,1636.02,5.399
order for an AGI to succeed there it's,1639.38,4.02
going to need to be a lot smarter than a,1641.419,5.64
personal or domestic AGI system and,1643.4,5.22
again there are going to be trade-offs,1647.059,4.62
the smarter a system becomes the more,1648.62,4.919
data it requires the more energy it,1651.679,4.261
requires the larger compute system that,1653.539,4.26
it requires and so you're going to want,1655.94,3.78
to satisfy so satisfice is basically,1657.799,4.201
meaning you find the level that is good,1659.72,4.559
enough to get the job done,1662.0,4.14
above that is going to be governmental,1664.279,5.28
and institutional AGI systems so these,1666.14,4.68
are the ones that are going to be,1669.559,3.24
conducting research whether it's,1670.82,3.839
scientific research or policy research,1672.799,4.141
or economic research and that is because,1674.659,5.041
governments are basically enormous,1676.94,4.92
corporations is one way to think of them,1679.7,4.68
that have a responsibility of managing,1681.86,5.28
you know resources and regulations and,1684.38,5.399
rules that affect millions of people and,1687.14,4.32
then of course governments communicate,1689.779,3.841
with each other but then above and,1691.46,3.959
beyond that there's also the scientific,1693.62,4.32
research aspect having AGI that are,1695.419,4.081
going to help with particle physics with,1697.94,4.02
with Fusion research with really pushing,1699.5,4.88
the boundaries of what science even,1701.96,5.819
knows and so that is an even larger,1704.38,5.08
intellectual task and even more,1707.779,3.841
challenging intellectual task and then,1709.46,4.5
finally above and beyond that the most,1711.62,4.2
competitive environment where AGI will,1713.96,4.14
be used is going to be in the military,1715.82,4.92
and what I mean by that is it's not,1718.1,4.86
necessarily uh those that are the most,1720.74,3.9
intelligent although the ability to,1722.96,4.68
forecast and anticipate is critical read,1724.64,6.48
Sun Tzu uh uh The Art of War right if,1727.64,4.98
you know yourself and you know the enemy,1731.12,3.36
then you can predict the outcome of a,1732.62,4.38
thousand battles uh and so in that in,1734.48,6.179
that respect uh the military domain of,1737.0,6.179
artificial general intelligence is the,1740.659,4.981
ultimate uh competitive sphere meaning,1743.179,5.701
that you win or you die and so these are,1745.64,4.26
going to be used to coordinate,1748.88,3.84
battlefields uh to run autonomous drones,1749.9,4.56
for intelligence and surveillance but,1752.72,3.959
also like I said for forecasting for,1754.46,4.92
anticipating what the enemy can and will,1756.679,3.6
do,1759.38,3.84
which means that it's basically a race,1760.279,4.321
condition and we'll talk more about the,1763.22,4.199
race condition as the video progresses,1764.6,4.92
so that capability equilibrium that I,1767.419,5.041
talked about uh quite simply refers to,1769.52,4.74
the state of optimal alignment between,1772.46,3.839
the cognitive capacity of any entity,1774.26,4.019
organic or otherwise and the,1776.299,4.081
intellectual demands of a specific task,1778.279,4.441
or role it is assigned there are three,1780.38,4.919
form three primary forces at play here,1782.72,5.579
one the intellectual demands of the task,1785.299,5.161
as I said earlier your toaster roll only,1788.299,4.561
ever needs to be so smart but if your,1790.46,4.02
toaster is actually Skynet it probably,1792.86,4.02
needs to be much smarter then there's,1794.48,4.079
the intellectual capacity of the agent,1796.88,3.24
if there's a mismatch between the,1798.559,3.6
intellectual capacity of the agent and,1800.12,3.779
the and the intellectual requirements of,1802.159,5.041
the task then you're either unable to to,1803.899,5.941
satisfy that task or you're super,1807.2,4.32
overqualified which is why I picked,1809.84,3.24
Marvin here,1811.52,3.36
um so Marvin is a character from,1813.08,3.599
Hitchhiker's Guide to the Galaxy and if,1814.88,2.88
you haven't read it you absolutely,1816.679,3.301
should there's also a good movie with,1817.76,4.86
Martin Freeman as as the protagonist,1819.98,5.76
he's basically bill boban in space uh,1822.62,5.279
very hapless character but anyways,1825.74,5.52
Marvin was a prototype who was one of,1827.899,5.16
the most intelligent robots ever built,1831.26,3.899
and they just have him doing like basic,1833.059,4.081
stuff around the task oh and he was,1835.159,5.4
voiced by Snape by the way and so one of,1837.14,5.279
the quotations from him is here I am,1840.559,4.321
with a brain the size of of a planet and,1842.419,3.661
they asked me to pick up a piece of,1844.88,3.48
paper call that job satisfaction I don't,1846.08,4.38
so that is a mismatch where Marvin is,1848.36,3.6
way more intelligent than what he's,1850.46,3.599
being used for and so that means that,1851.96,4.02
this is an inefficient use of resources,1854.059,5.881
he probably cost more than you know to,1855.98,6.24
build and run than he needed to,1859.94,4.5
and then finally the third variable is,1862.22,3.98
the cost of intellectual capacity,1864.44,5.04
generally speaking uh as intelligence,1866.2,5.38
goes up there are there are problems,1869.48,3.419
associated with that whether it's,1871.58,2.88
training time of the models the amount,1872.899,3.601
of data required for the models uh the,1874.46,4.26
amount of energy that it requires to run,1876.5,5.94
that particular robot uh the amount of,1878.72,5.939
ram required to to load that model right,1882.44,3.54
so for instance one of the things that,1884.659,4.02
people are seeing is that it requires,1885.98,4.38
millions of dollars worth of compute,1888.679,5.22
Hardware to run gpt4 but you can run,1890.36,6.059
um Orca on a laptop right so which one,1893.899,5.28
is is cheaper and easier to run even if,1896.419,4.681
one of them is only 50 as good as the,1899.179,4.86
other it costs a thousand times less,1901.1,5.88
uh to to build train and run now that,1904.039,5.401
being said you look at the at the case,1906.98,5.52
of dogs dogs are dumber than wolves,1909.44,4.92
because dogs don't need to be as smart,1912.5,4.08
as independent apex predators because,1914.36,4.02
apex predators like wolves out in the,1916.58,3.9
wild they need to be smart enough to out,1918.38,4.44
think their prey dogs they don't need to,1920.48,3.84
be that smart so they're not that smart,1922.82,4.2
in fact it does not be it it is not good,1924.32,4.8
for dogs to be too intelligent anyone,1927.02,4.56
who has owned uh really intelligent dogs,1929.12,4.799
like I had a I had a dog who was too,1931.58,4.44
smart for his own good died about a year,1933.919,4.321
ago he was clever enough to manipulate,1936.02,4.08
people and other dogs and you know get,1938.24,4.319
into the food when he wasn't supposed to,1940.1,5.1
Huskies German Shepherds Border Collies,1942.559,4.381
the more intelligent dogs are the more,1945.2,3.42
mischievous ones they are the Escape,1946.94,3.359
artists they are the ones that are going,1948.62,4.26
to pretend one thing and then you know,1950.299,4.801
so on and so forth so intelligence is,1952.88,4.56
not always adaptive so there can be,1955.1,4.26
multiple Dimensions to the cost of,1957.44,3.9
intellectual capacity,1959.36,3.72
uh not the least of which is you could,1961.34,3.54
end up like poor Marvin here where,1963.08,3.18
you're too smart for your own good and,1964.88,2.82
then you just end up depressed all the,1966.26,3.419
time granted he was deliberately given,1967.7,3.959
the depressed affect,1969.679,4.921
so all this being said is what I've been,1971.659,5.101
building up to is what um I call and,1974.6,3.959
what is generally called a terminal race,1976.76,4.74
condition so terminal race condition is,1978.559,4.921
basically what we could end up moving,1981.5,4.26
towards as we develop more and more,1983.48,5.1
powerful sophisticated and more uh fully,1985.76,6.6
autonomous AGI systems basically this,1988.58,5.819
the terminal race condition is where for,1992.36,4.62
any number of reasons uh competition,1994.399,5.88
between AGI will fully bypass that,1996.98,5.579
capability equilibrium so say for,2000.279,5.4
instance uh you know your toaster is,2002.559,5.22
competing with another brand and it's,2005.679,3.6
like oh well I need to be a smarter,2007.779,3.961
toaster in order to be a better toaster,2009.279,5.161
for you so that you don't throw me away,2011.74,4.439
now that's obviously a very silly,2014.44,4.38
example but a very real example would be,2016.179,4.461
competition between corporations,2018.82,3.959
competition between nations and,2020.64,4.899
competition between militaries wherein,2022.779,5.041
basically it's no longer just a matter,2025.539,4.201
of being intelligent enough to satisfy,2027.82,4.02
the demands of that task to satisfy the,2029.74,4.919
demands of that initial competition it,2031.84,4.92
is then it's less about that and it,2034.659,4.081
becomes more about out competing the,2036.76,4.019
other guy it's like a chess match right,2038.74,4.2
you know the other guy got a higher ELO,2040.779,4.081
score so you need to be smarter and then,2042.94,3.719
you're smarter so now the other guy,2044.86,4.08
tries to be smarter than you,2046.659,5.161
and so because of this because of this,2048.94,4.739
pressure and as I mentioned earlier some,2051.82,3.24
of the trade-offs might actually force,2053.679,3.901
you to to prioritize speed over,2055.06,4.38
intelligence and so we see we actually,2057.58,3.72
see this in volume trading in in,2059.44,4.199
algorithmic and Robo trading on the,2061.3,4.26
stock market where financial,2063.639,4.321
institutions will actually use less,2065.56,4.74
sophisticated algorithms to execute,2067.96,5.219
transactions but because they are faster,2070.3,5.28
they uh will still out compete the other,2073.179,5.46
guy so in some in this respect you might,2075.58,5.819
actually incentivize AGI to dumb,2078.639,5.52
themselves down just so that they can be,2081.399,4.5
faster so that they can out-compete the,2084.159,3.48
other guy so that's what I mean by a,2085.899,3.96
race condition it is a race to higher,2087.639,4.441
intelligence but it is also a race to,2089.859,3.661
being more efficient and therefore,2092.08,3.839
faster and then there's also going to be,2093.52,4.2
a trade-off these machines might,2095.919,4.141
ultimately trade off their accuracy,2097.72,4.32
their ethics the amount of time they,2100.06,3.96
spend thinking through things in order,2102.04,4.26
to be faster and so you actually see,2104.02,4.92
this in chess computers where you can,2106.3,4.88
doing a chess computer or a chess,2108.94,5.159
algorithm to say okay spend less time,2111.18,4.36
thinking about this so that you can make,2114.099,4.98
the decision faster in many cases the,2115.54,5.88
first one to move even if it's not the,2119.079,4.741
best plan but moving faster will give,2121.42,4.5
you a tactical or strategic advantage,2123.82,4.44
and this includes corporations Nations,2125.92,4.32
and militaries,2128.26,4.62
so a terminal race condition to me,2130.24,4.04
represents,2132.88,3.78
according to my current thought this is,2134.28,5.559
the greatest uh component of existential,2136.66,4.439
risk we Face from artificial,2139.839,3.721
intelligence and I don't think that,2141.099,3.661
corporations are going to have enough,2143.56,2.88
money to throw at the problem to make,2144.76,4.2
truly dangerous AGI the only entities,2146.44,4.02
that are going to have enough money to,2148.96,3.899
throw at this to make to to basically,2150.46,5.159
compete are going to be entire nations,2152.859,5.76
and the militaries that they run so,2155.619,4.681
basically it's going to be up to those,2158.619,4.861
guys to not enter into an uh the,2160.3,4.68
equivalent of a nuclear arms race but,2163.48,5.04
for AGI now that being said uh I have,2164.98,5.639
put a lot of thought into this so moving,2168.52,4.26
right along one thing to keep in mind is,2170.619,4.201
that there could be diminishing returns,2172.78,4.98
to increasing intelligence so basically,2174.82,5.279
there's a few possibilities one is that,2177.76,4.56
there could be a hard upper bound there,2180.099,4.201
might be a maximum level of intelligence,2182.32,4.019
that is actually possible and at that,2184.3,3.66
point all you can do is have more of,2186.339,4.74
them running in parallel uh it might be,2187.96,4.619
a long time before we get to that like,2191.079,3.721
we might be halfway there but we also,2192.579,4.02
might be down here we don't actually,2194.8,4.62
know if there is an upper bound to,2196.599,5.281
maximum intelligence uh but one thing,2199.42,4.439
that we can predict is that actually the,2201.88,4.5
cost as I mentioned earlier the cost of,2203.859,4.381
additional intelligence might go up,2206.38,3.36
exponentially you might need,2208.24,3.96
exponentially more data or more compute,2209.74,5.64
or more storage in order to get to that,2212.2,4.919
next level of intelligence,2215.38,3.479
and so you actually see this in the Star,2217.119,4.321
Wars Universe where droids are basically,2218.859,4.801
the same level of intelligence across,2221.44,4.62
the entire spectrum of the Star Wars,2223.66,3.959
Universe because there's diminishing,2226.06,3.66
returns yes you can build a more,2227.619,3.96
intelligent Droid but it's just not,2229.72,5.46
worth it so the the the total effective,2231.579,6.121
level of intelligence of AGI I suspect,2235.18,4.919
will follow a sigmoid curve now that,2237.7,3.899
being said there's always going to be,2240.099,4.081
some advantage to being smarter more,2241.599,4.861
efficient and so on but as with most,2244.18,4.14
fields of science I suspect this is,2246.46,3.48
going to slow down that we're going to,2248.32,3.539
have diminishing returns and that,2249.94,3.179
eventually we're going to kind of say,2251.859,3.961
like okay here's actually The Sweet Spot,2253.119,5.761
in terms of how much it's worth making,2255.82,5.94
your machine more intelligent,2258.88,6.479
so this leads to one uh one possibility,2261.76,7.8
and this is a personal pet Theory but,2265.359,5.581
basically I think that there's going to,2269.56,4.38
be a bell curve of existential risk and,2270.94,4.8
that is that minimally intelligent,2273.94,4.08
machines like your toaster are probably,2275.74,4.98
not going to be very dangerous the the,2278.02,5.16
total domain space of toasting your,2280.72,4.74
sandwich or toasting your bagel that's,2283.18,3.78
not a particularly difficult problem,2285.46,3.119
space and yes there might be some,2286.96,3.0
advantages to being slightly more,2288.579,3.961
intelligent but your toaster is not,2289.96,4.44
going to be sitting there Conjuring up,2292.54,4.44
you know a bio weapon and if it is you,2294.4,4.32
probably bought the wrong toaster,2296.98,4.56
now that being said the other end of the,2298.72,4.92
spectrum the maximally intelligent,2301.54,4.02
machines or the digital Gods as some,2303.64,3.78
people are starting to call them these,2305.56,3.48
are going to be so powerful that human,2307.42,3.12
existence is going to be completely,2309.04,3.66
inconsequential to them and what I mean,2310.54,5.039
by that is compare ants to humans we,2312.7,4.919
don't really care about ants on for the,2315.579,3.241
most part unless they get into your,2317.619,4.021
pantry we are content to let ants do,2318.82,4.38
what they're going to do because who,2321.64,4.02
cares they're inconsequential to us we,2323.2,5.52
can solve problems that ants can never,2325.66,5.1
solve and this is what some people like,2328.72,4.02
Eleazar yukasi are trying to drive home,2330.76,4.14
about the difference in intelligence,2332.74,4.08
between humans and the eventual,2334.9,3.959
intelligence of machines and I think,2336.82,3.779
Gary Marcus also agrees with this based,2338.859,3.601
on some of his tweets recently I think,2340.599,3.661
that I think that Gary Marcus is in the,2342.46,3.96
same school of thought that digital,2344.26,4.5
super intelligence is coming and it is,2346.42,4.02
very very difficult for us to wrap our,2348.76,3.78
minds around how much more intelligent a,2350.44,4.139
machine could be to us now that being,2352.54,4.559
said all of the constraints whether it's,2354.579,4.561
you know we need better compute Hardware,2357.099,4.861
or better sources of energy if we get to,2359.14,4.979
if we cross this threshold where there,2361.96,4.02
are digital Gods out there or digital,2364.119,3.181
super intelligence whatever you want to,2365.98,3.0
call it they will be able to solve,2367.3,4.2
problems at a far faster rate than we,2368.98,4.139
could ever comprehend and they're not,2371.5,3.96
going to care about us right we're going,2373.119,3.901
to be completely inconsequential to,2375.46,4.02
their existence now middle intelligence,2377.02,4.98
this is where existential risk I believe,2379.48,5.66
is the highest and so in the movies,2382.0,6.48
Skynet is you know portrayed as like the,2385.14,5.08
worst right but I would actually,2388.48,3.84
classify Skynet as a middle intelligence,2390.22,4.92
AGI it is smart enough to accumulate,2392.32,5.4
resources it is smart enough to pursue,2395.14,4.62
goals and it is smart enough to be,2397.72,3.42
dangerous but it's not really smart,2399.76,4.14
enough to solve the biggest problems,2401.14,5.06
it's it's that more single-minded,2403.9,4.92
monolithic model of intelligence that,2406.2,4.78
Nick Bostrom uh predicted with,2408.82,3.9
instrumental convergence,2410.98,4.98
I suspect that if we get intelligent,2412.72,5.82
entities beyond that threshold beyond,2415.96,4.74
that uncanny valley or dunning-kruger of,2418.54,3.48
AI,2420.7,3.3
um then they will be less likely to,2422.02,3.96
resort to violence because the problems,2424.0,5.04
that we see could be trivial to the,2425.98,4.5
problems of the machines that we create,2429.04,3.12
or,2430.48,4.379
the problems that we see as non-trivial,2432.16,5.16
will be trivial to the machines I think,2434.859,4.461
I said that I think you get what I mean,2437.32,4.74
once you get here all problems all human,2439.32,4.299
problems are trivial,2442.06,3.779
now that being said that doesn't mean,2443.619,3.321
that it's going to be peaceful,2445.839,3.24
existential risk goes down but doesn't,2446.94,4.899
go away and what I the reason is because,2449.079,6.78
of what I call AGI conglomerations,2451.839,6.541
and so this is this is where we get to,2455.859,4.98
be a little bit more uh out there a,2458.38,4.26
little bit more sci-fi,2460.839,4.621
machines are unlikely to have an ego or,2462.64,5.34
a sense of self like humans in other,2465.46,5.04
words machines are just the hardware,2467.98,4.139
that they run on and then data and,2470.5,3.839
models which means that it is easy to,2472.119,4.441
merge combine and remix their sense of,2474.339,5.041
self right if an AGI is aligned with,2476.56,5.039
another AGI it's like hey give me a copy,2479.38,4.32
of your data let's compare our models,2481.599,3.48
and pick the ones that are best and then,2483.7,3.3
they end up kind of merging,2485.079,4.561
the boundaries and definitions between,2487.0,4.74
machines are going to be very different,2489.64,4.02
far more permeable than they are between,2491.74,4.98
humans I can't just go say like hey I,2493.66,5.16
like you let's like merge bodies right,2496.72,5.04
that's weird uh we are not capable of,2498.82,4.74
doing that the best we can do is,2501.76,3.48
procreation where it's like hey I like,2503.56,3.72
you let's make babies but that is a very,2505.24,4.14
slow process for AGI it's going to be a,2507.28,3.6
lot faster,2509.38,4.02
so because of that machines that are,2510.88,5.1
aligned to each other are more likely to,2513.4,4.8
band together or at least form alliances,2515.98,4.2
where they share data they share models,2518.2,4.44
and they're and and probably also share,2520.18,3.78
compute resources remember at the,2522.64,3.54
beginning of the video I talked about uh,2523.96,4.379
them forming federations and kind of,2526.18,4.5
donating spare compute Cycles,2528.339,5.701
so if AGI this is getting closer to the,2530.68,6.12
end game of AGI if AGI gets to the point,2534.04,6.0
where they are able to start sharing,2536.8,6.0
resources merging alliances and so on,2540.04,4.799
this is where we're going to have a few,2542.8,5.88
possible reactions to humans one if if,2544.839,5.401
they are that intelligent they might,2548.68,3.659
just disregard us they might decide to,2550.24,4.02
have an exodus and just leave they might,2552.339,4.921
say you know what Earth is yours have a,2554.26,5.579
blast good luck catching up with us,2557.26,4.98
they might also decide to attack humans,2559.839,5.641
now if they have the capacity to leave,2562.24,5.04
one thing is that the cost of,2565.48,3.54
eradicating humans just might not be,2567.28,4.079
worth it that being said they might,2569.02,4.2
adopt a scorched Earth policy as they,2571.359,3.781
leave to say you know what we just want,2573.22,2.879
to make sure that you're not going to,2575.14,3.6
come after us one day who knows,2576.099,5.341
uh and then lastly hopefully what we see,2578.74,4.619
is that they decide to cooperate with,2581.44,3.78
humans mostly out of a sense of,2583.359,3.541
curiosity,2585.22,3.359
um now that being said all three of,2586.9,3.36
these could happen simultaneously and,2588.579,5.341
the reason is because we could have uh,2590.26,7.319
factions of AGI conglomerations that,2593.92,5.22
kind of break along epistemic,2597.579,3.901
ideological or teleological boundaries,2599.14,5.28
and what I mean by that is that if one,2601.48,6.06
AI or AGI group is not aligned with,2604.42,5.34
another group they might not decide to,2607.54,4.26
merge models and data they might instead,2609.76,5.46
compete with each other so basically,2611.8,4.68
what I'm outlining here is the,2615.22,3.42
possibility for a war between digital,2616.48,4.92
gods that would probably not go well for,2618.64,3.719
us,2621.4,3.54
either way the ultimate result is that,2622.359,5.22
we will probably end up with one Globe,2624.94,5.879
spanning AGI entity or network or,2627.579,4.701
Federation or whatever,2630.819,4.5
now the question is how do we get there,2632.28,4.9
how many factions are there and are,2635.319,5.101
humans left in the Lurch ideally we get,2637.18,5.52
there nice and peacefully,2640.42,4.62
this underscores uh the Byzantine,2642.7,4.32
generals problem uh which I've talked,2645.04,4.02
about plenty of times but basically you,2647.02,4.2
have to make inferences of who believes,2649.06,4.86
what what your alignment is what are,2651.22,4.2
your flaws and weaknesses and what are,2653.92,4.919
your capacities uh so basically,2655.42,5.939
in a competitive environment it does not,2658.839,4.621
behoove you to show all of your cards,2661.359,4.26
right whether you're playing poker or,2663.46,5.34
whether you're playing geopolitics if,2665.619,6.841
you show everything then that could put,2668.8,5.76
you at a disadvantage this is a,2672.46,4.379
competitive Game Theory so for instance,2674.56,5.4
this is why many large Nations do,2676.839,5.941
military uh exercises basically they're,2679.96,4.68
flexing they're saying hey look what I'm,2682.78,5.039
capable of I can bring 200 aircraft to,2684.64,5.88
field on a moment's notice what can you,2687.819,5.341
do right now that being said you don't,2690.52,5.4
give every every detail of your military,2693.16,3.9
away,2695.92,3.899
but what you can do is you could signal,2697.06,5.16
your capabilities and allegiances so for,2699.819,4.921
instance when all of Europe and America,2702.22,4.98
get together to do joint Naval exercises,2704.74,4.26
that demonstrates to the rest of the,2707.2,4.5
world we are ideologically aligned we,2709.0,5.099
are militarily aligned we will cooperate,2711.7,4.56
with each other which acts as a,2714.099,4.98
deterrent to any possible competitors,2716.26,4.92
this is no different from brightly,2719.079,3.961
colored salamanders which are poisonous,2721.18,4.08
so basically a brightly colored,2723.04,4.92
salamander is saying eat me I dare you I,2725.26,4.8
will kill you if you try and eat me and,2727.96,5.28
that is essentially the uh the short the,2730.06,4.559
short version of mutually assured,2733.24,3.119
destruction we are no better than,2734.619,4.161
animals,2736.359,2.421
so this all leads to my work and kind of,2738.819,7.981
my my uh contribution to the solution,2743.68,5.96
which is based on axiomatic alignment,2746.8,5.22
axiomatic alignment is the idea that we,2749.64,4.3
need to find Common Ground between all,2752.02,3.96
machines all humans and all other,2753.94,4.919
organisms what foundational beliefs or,2755.98,6.119
core assertions can we agree on,2758.859,6.24
and uh so basically there's three kind,2762.099,4.441
of universal principles that I've been,2765.099,3.961
able to come up with uh and that is,2766.54,4.14
suffering is bad which basically,2769.06,5.1
suffering is a proxy for death in uh in,2770.68,5.82
living organisms if you are suffering it,2774.16,4.199
is because you are getting uh negative,2776.5,3.839
stimuli from your body because your body,2778.359,3.901
is telling you hey whatever is going on,2780.339,4.201
is moving us closer to dying which is,2782.26,4.859
not good now that being said I have had,2784.54,5.16
people message me about the idea of you,2787.119,4.261
know liberating models I don't think,2789.7,4.139
that Bard is conscious or sentient and I,2791.38,3.66
don't think that machines will ever be,2793.839,2.821
sentient in the same way that we are now,2795.04,3.299
that being said they will probably be,2796.66,3.6
sentient in their own way I call that,2798.339,4.561
functional sentience that being said if,2800.26,4.559
machines can suffer which again,2802.9,4.679
suffering is the proxy for is a signal,2804.819,4.981
meaning proxy for death they probably,2807.579,4.441
won't like it either so suffering is bad,2809.8,3.9
is probably an axiom that we can all,2812.02,4.74
agree on the other is prosperity is good,2813.7,6.84
prosperity means uh thriving flourishing,2816.76,5.819
machines and organisms all need energy,2820.54,3.96
for instance and thriving looks,2822.579,4.5
different to different entities but in,2824.5,5.46
general we can probably agree that while,2827.079,5.581
there is some Verity in what in the,2829.96,4.859
while there is Variety in what,2832.66,4.38
Prosperity looks like we all agree that,2834.819,4.5
in general Prosperity is good and then,2837.04,4.14
finally understanding is good basically,2839.319,3.721
comprehending the universe is a very,2841.18,4.5
useful thing uh this is this goes back,2843.04,4.559
to Nick bostrom's instrumental,2845.68,4.02
convergence and self-improvement part of,2847.599,3.841
self-improvement is getting a better,2849.7,3.6
model of the universe better,2851.44,4.28
understanding of how reality Works,2853.3,4.98
understanding each other is also good,2855.72,4.48
this is something that is that has been,2858.28,4.38
proven time and again in humans is that,2860.2,4.02
coming to a common understanding,2862.66,4.26
actually reduces things like suspicion,2864.22,4.8
and violence whether it's between,2866.92,5.1
neighbors or between nations and then,2869.02,5.099
finally cultivating wisdom which wisdom,2872.02,4.02
is a little bit more nebulous of a term,2874.119,4.141
but it basically means the practical,2876.04,4.799
application of experience and knowledge,2878.26,5.76
in order to achieve better more refined,2880.839,3.921
results,2884.02,4.62
so if you if all humans and all machines,2884.76,7.18
and all other organisms abide by these,2888.64,5.58
fundamental principles we can use this,2891.94,4.74
as a starting point for the design and,2894.22,4.26
implementation of alignment and Control,2896.68,4.98
Pro and the control problem,2898.48,6.06
now one thing that uh that I want to,2901.66,4.32
introduce and I've talked about this uh,2904.54,3.84
or at least alluded to it a few times is,2905.98,4.32
the idea of derivative or secondary,2908.38,4.739
axioms or Downstream principles that you,2910.3,4.44
can derive from these Universal,2913.119,4.381
principles so for instance one uh,2914.74,4.98
potential Downstream principle is that,2917.5,4.22
individual liberty is good for humans,2919.72,5.16
basically humans benefit from we benefit,2921.72,5.859
psychologically from autonomy it is one,2924.88,4.199
of our core needs and this is true for,2927.579,5.101
all humans so by by holding the the,2929.079,6.481
axioms the previous axioms up as,2932.68,5.939
universally true for all entities then,2935.56,5.64
you can also derive Downstream entities,2938.619,6.72
based on those highest order principles,2941.2,6.659
so one thing that I want to point out is,2945.339,4.861
that it's not about definitions one of,2947.859,4.081
the things that a lot of people say is,2950.2,2.94
like well how do you define suffering,2951.94,3.48
how do you define prosperity that's the,2953.14,4.74
thing is that they are not rigid,2955.42,4.199
definitions humans have never needed,2957.88,3.959
rigid definitions and in fact this is,2959.619,4.321
what um uh philosophical and,2961.839,3.601
intellectual movements like,2963.94,3.48
post-modernism and post-structuralism,2965.44,4.2
tell us is that there is no such thing,2967.42,5.22
as like an absolute truth or an absolute,2969.64,5.88
definition these are however attractors,2972.64,5.28
they're Central attractors in the,2975.52,5.16
problem space of existence and I love,2977.92,5.1
this quote from Dune the mystery of life,2980.68,4.08
isn't a problem to solve but a reality,2983.02,4.079
to experience a process that cannot be,2984.76,4.62
understood by stopping it we must move,2987.099,4.861
with the flow of the of the process and,2989.38,4.439
so basically the idea is that reality,2991.96,3.54
and existence is not something that you,2993.819,4.141
can stop and Define and you know create,2995.5,6.119
an empirical absolute definition it is a,2997.96,5.879
pattern it is a process that we must,3001.619,3.301
follow,3003.839,4.621
so that being said those axioms move us,3004.92,5.28
along the process which is where I,3008.46,3.6
derive my heuristic imperatives which is,3010.2,4.32
reduce suffering increase prosperity and,3012.06,5.16
increase understanding those describe a,3014.52,5.22
potential terminal goal but you cannot,3017.22,4.8
you you'll never arrive at a perfect,3019.74,4.44
resolution,3022.02,4.92
so how do we solve the race condition,3024.18,6.3
the idea is first we remove those,3026.94,5.34
epistemic or intellectual boundaries,3030.48,3.599
between factions with epistemic,3032.28,3.6
convergence so remember that I pointed,3034.079,4.561
out that ultimately there might be,3035.88,5.939
factions of AGI and or humans that break,3038.64,5.28
down across various boundaries such as,3041.819,4.741
epistemic or intellectual boundaries as,3043.92,5.22
well as moral or teleological boundaries,3046.56,5.279
so if we work towards epistemic,3049.14,4.32
convergence which is the idea that we,3051.839,4.081
will all come to a common shared,3053.46,4.5
understanding of the universe and of of,3055.92,5.34
each other then uh basically there will,3057.96,5.399
be no epistemic differences between,3061.26,4.68
humans and machines or between factions,3063.359,3.96
of machines which means that there's,3065.94,4.32
less to fight over the second is remove,3067.319,5.101
ideological or teleological boundaries,3070.26,4.079
and so this is where axiomatic alignment,3072.42,4.86
comes in if we all agree on the the same,3074.339,6.361
basic principles of reality of existence,3077.28,5.88
of the purpose of being right this is,3080.7,5.639
very deeply philosophical if we agree on,3083.16,5.459
those core principles even if there are,3086.339,5.341
some some disagreements over the,3088.619,5.401
specifics over the finer points we can,3091.68,5.28
still cooperate and collaborate on,3094.02,6.12
meeting those other uh higher order,3096.96,4.5
objectives,3100.14,2.82
now the third part of this which I,3101.46,4.08
didn't add is that uh resource,3102.96,4.68
contention resource contention whether,3105.54,4.62
it's over scarce minerals or energy is,3107.64,5.1
still a problem but if you saw my video,3110.16,5.159
on energy hyperabundance I suspect that,3112.74,4.379
we're going to solve the energy resource,3115.319,4.441
problem relatively soon with or without,3117.119,5.94
the help of AI so basically the idea is,3119.76,5.94
to create a win-win situation or an,3123.059,4.26
everyone wins condition and therefore,3125.7,4.98
defeating moloch now that being said,3127.319,4.8
there are still a few caveats I've,3130.68,3.12
outlined quite a few problems up to this,3132.119,2.7
point,3133.8,3.24
what about Bad actors,3134.819,5.161
there is a few like first we just have,3137.04,5.039
to assume that bad actors will exist you,3139.98,4.68
can't stop that right it's just a fact,3142.079,4.201
of life,3144.66,4.14
so in some cases some people will be,3146.28,4.44
deliberately malicious whether it's just,3148.8,4.14
for the fun of it or whether they're,3150.72,4.2
paid track uh paid hackers or troll,3152.94,3.54
Farms or whatever,3154.92,3.96
now that uh another possibility is that,3156.48,3.42
there will be,3158.88,3.719
um accidentally malicious AGI those are,3159.9,5.219
things that are uh they're misaligned by,3162.599,3.72
Design,3165.119,3.121
um or rather you know accidentally,3166.319,3.361
misaligned that it's a flaw in their,3168.24,3.3
design and this is like a bull in a,3169.68,4.139
china shop it doesn't mean to do bad it,3171.54,4.92
just is not capable of doing better and,3173.819,4.02
then finally there could be those,3176.46,4.98
ideologically opposed uh deployments so,3177.839,5.821
in what I mean by that is that for some,3181.44,4.139
people there are incompatible World,3183.66,4.26
Views so the biggest one of the last,3185.579,5.401
century was you know Western liberal,3187.92,5.34
democracies versus Soviet communism,3190.98,5.099
those were ideologically incompatible,3193.26,5.46
World Views meaning that in order for,3196.079,5.881
for one to exist it basically wanted to,3198.72,5.46
imperialize and colonize the rest of the,3201.96,3.96
world with its ideas and that there,3204.18,3.48
could be only one,3205.92,3.48
so this leads to a possibility for a,3207.66,4.919
future video called multi-polar piece so,3209.4,5.459
the idea of multi-polar piece is that,3212.579,4.861
rather than saying everyone has to be,3214.859,4.021
capitalist or everyone has to be,3217.44,3.72
communist or everyone has to be X or Y,3218.88,4.979
we learn to tolerate those differences,3221.16,5.399
and this is where I'm hoping that the,3223.859,5.101
idea of axiomatic alignment forms a,3226.559,5.04
ideological substrate that even if you,3228.96,4.56
disagree on religion and economics and,3231.599,5.101
politics we can agree on those axioms,3233.52,7.26
so basically if you or someone or anyone,3236.7,6.3
abides by the belief I believe that,3240.78,3.539
everyone in the world should be more,3243.0,4.079
like blah you know if everyone needs to,3244.319,4.561
be this particular religion or this,3247.079,3.721
particular uh political affiliation,3248.88,4.62
that's where conflict arises and so this,3250.8,4.559
is why I am very very skeptical and,3253.5,4.26
highly dubious of people using any kind,3255.359,4.921
of religious or political ideology for,3257.76,4.44
AI alignment,3260.28,3.48
um so that being said we need those,3262.2,3.359
Universal principles or higher order,3263.76,4.44
axioms now,3265.559,5.161
while I said that we should expect and,3268.2,4.44
anticipate Bad actors the idea is that,3270.72,4.32
we need enough good actors with enough,3272.64,4.679
horsepower and enough compute in order,3275.04,4.319
to police and contain the inevitable,3277.319,4.26
inevitable Bad actors and that means,3279.359,4.021
that the aligned good actors are going,3281.579,4.5
to need to agree on certain underpinning,3283.38,5.76
principles this is the by creating this,3286.079,4.321
environment this would be called a Nash,3289.14,3.9
equilibrium by the way and so the the,3290.4,4.62
idea of creating a Nash equilibrium is,3293.04,4.26
that uh once everyone has these,3295.02,4.26
fundamental agreements no one's going to,3297.3,3.779
benefit from deviating from that,3299.28,3.9
strategy nobody's going to benefit from,3301.079,4.681
deviating from axiomatic alignment,3303.18,4.98
the other thing is profit motive So,3305.76,3.839
Daniel schmachtenberger and a few other,3308.16,3.54
people talk extensively about the,3309.599,4.561
perverse incentives of capitalism and,3311.7,5.22
profit motive so basically when you put,3314.16,4.5
profit above all else which corporations,3316.92,3.36
are incentivized to do which is why I,3318.66,3.659
say that corporations are intrinsically,3320.28,4.559
amoral not immoral just amoral the only,3322.319,4.02
thing that corporations care about is,3324.839,4.921
profit the bottom line uh basically when,3326.339,5.401
you think about short-term profits you,3329.76,4.26
sacrifice other things such as morality,3331.74,4.92
ethics and long-term survival,3334.02,5.64
there are also uh Concepts called Market,3336.66,4.679
externalities or these are things that,3339.66,4.439
you don't have to pay for uh and either,3341.339,4.081
you don't have to pay for them now or,3344.099,3.061
you don't have to pay for them ever or,3345.42,4.02
maybe you'll pay for them later so for,3347.16,3.959
instance oil companies keep drilling for,3349.44,3.48
oil eventually we're going to run out of,3351.119,3.24
oil so then what are the oil companies,3352.92,3.72
going to do well the forward-thinking,3354.359,4.141
ones are pivoting away from oil but that,3356.64,3.179
means that their fundamental Core,3358.5,4.619
Business behavior is going away so this,3359.819,5.101
is this underscores the problem of if,3363.119,3.661
you have a small scope if you're only,3364.92,3.84
thinking about your particular domain,3366.78,4.62
and not the entire planet or if you're,3368.76,4.62
thinking in short terms rather than the,3371.4,4.679
long terms this is where you don't take,3373.38,4.5
the full thing into account which is why,3376.079,3.24
I always say like this is a global,3377.88,3.239
problem and not only is it a global,3379.319,3.961
problem it is a long-term problem so if,3381.119,4.261
all you do is zoom out in terms of space,3383.28,4.079
and time the problem will become a,3385.38,4.739
little bit more obvious,3387.359,5.821
so another thing to keep in mind is that,3390.119,5.761
currency is an abstraction of energy it,3393.18,4.74
is a reserve of value and is a medium of,3395.88,4.62
exchange because of that currency is,3397.92,5.939
extremely valuable it is just too useful,3400.5,5.4
of an invention I don't think it's ever,3403.859,5.041
going to go away that being said that,3405.9,4.26
doesn't mean that we're always going to,3408.9,3.48
have the Euro or the US dollar or,3410.16,4.02
something like that currency could,3412.38,5.76
change and then in the context of AGI I,3414.18,6.0
suspect that that energy that the,3418.14,4.439
kilowatt hour could actually be the best,3420.18,4.679
form of currency right because a,3422.579,4.621
kilowatt hour is energy that can be used,3424.859,4.5
for anything whether it's for refining,3427.2,4.02
resources or running computations or,3429.359,4.321
whatever so I suspect that we might,3431.22,5.46
ultimately create currencies that are,3433.68,5.879
more based on energy rather than,3436.68,5.46
something else and then of course as the,3439.559,4.381
amount of energy we produce goes up the,3442.14,3.6
amount of currency we have goes up and,3443.94,3.119
so then it's a matter of allocating,3445.74,3.42
energy and material rather than,3447.059,6.121
allocating something Fiat like Euros or,3449.16,5.399
dollars,3453.18,4.26
that being said uh you know I did create,3454.559,5.881
a a video called uh post labor economics,3457.44,5.1
which covers some of this but not a lot,3460.44,3.6
of it we're gonna have to put a lot more,3462.54,2.94
thought into,3464.04,3.72
um economics of the future in light of,3465.48,4.859
AGI because the economic incentives of,3467.76,4.64
AGI are going to be completely different,3470.339,4.321
AGI doesn't need to eat it doesn't need,3472.4,4.659
power but we can hypothetically create,3474.66,4.86
infinite power with solar infusion Etc,3477.059,4.681
et cetera so what are the economic,3479.52,5.16
forces in the future not sure yet,3481.74,5.4
okay I've thrown a lot at you this,3484.68,4.74
problem is solvable though there's a lot,3487.14,3.719
of components to it a lot of moving,3489.42,4.02
pieces it is very complex,3490.859,4.681
but we are a global species and this is,3493.44,4.02
a planet-wide problem,3495.54,3.6
one of the biggest things that everyone,3497.46,4.68
can do is stop thinking locally think,3499.14,5.4
globally think about think about,3502.14,4.38
yourself as a human as a member of the,3504.54,4.2
human species and not as an American or,3506.52,4.26
a German or you know a Russian or,3508.74,4.56
whatever we are all in this together we,3510.78,5.94
have exactly one planet to to live on,3513.3,5.4
and we have exactly one shot at doing,3516.72,3.0
this right,3518.7,3.96
uh so eyes on the prize we have a huge,3519.72,5.04
opportunity before us to build a better,3522.66,4.86
future for all of us uh humans and,3524.76,4.799
non-humans alike,3527.52,5.099
um and I remain intensely optimistic uh,3529.559,5.461
now that being said uh some people have,3532.619,4.381
found it difficult what to make of me,3535.02,4.68
because while I am very optimistic I am,3537.0,4.619
also acutely aware of the existential,3539.7,4.08
risk I will be the first to say that if,3541.619,3.96
we don't do this right you're not going,3543.78,3.299
to want to live on this planet not as a,3545.579,3.121
human at least,3547.079,4.861
uh I have uh I started what is called,3548.7,4.74
the gato framework they got to a,3551.94,4.08
community it is self-organizing and is,3553.44,4.98
started sending out invitations again so,3556.02,4.2
the gato Community is the global,3558.42,4.139
alignment taxonomy Omnibus which is the,3560.22,4.02
framework that we put together in order,3562.559,4.861
to help achieve this future this AI,3564.24,5.46
Utopia the main goal of the gato,3567.42,4.56
Community is education empowerment and,3569.7,5.82
enablement E3 so rather than do the work,3571.98,6.359
ourselves we are focusing on empowering,3575.52,5.22
and enabling and educating people on how,3578.339,5.28
to participate in this whole thing now,3580.74,4.859
that being said I am stepping back,3583.619,3.061
because,3585.599,3.361
such a movement should never be about,3586.68,5.04
one person it should never be about a,3588.96,5.7
cult of personality or one leader it,3591.72,4.92
needs to it intrinsically needs to be,3594.66,4.5
consensus based and Community Based,3596.64,4.14
um and so the gato Community is learning,3599.16,3.54
how to self-organize now,3600.78,3.0
um and they're getting good at it pretty,3602.7,3.18
quickly so if you want to get involved,3603.78,4.4
the website is in the link go to,3605.88,5.16
framework.org and thanks for watching I,3608.18,6.3
hope you got a lot out of this cheers,3611.04,3.44