davidshapiro_youtube_transcripts / Terminal Race Condition The greatest danger we face from AGI and how to prevent it_transcript.csv
Stevross's picture
Upload 50 files
421fea8
text,start,duration
good morning everybody David Shapiro,0.42,3.6
here with another video,2.639,4.62
so today's video uh it started off as,4.02,5.28
one thing I wanted to primarily talk,7.259,4.861
about epistemic convergence uh but It,9.3,4.32
ultimately ended up being a little bit,12.12,2.54
more,13.62,2.82
all-encompassing so I'm going to,14.66,3.94
introduce a few new terms but we are,16.44,4.08
going to cover cover uh epistemic,18.6,4.259
convergence and a few other things,20.52,4.62
uh real quick before we dive into the,22.859,3.781
video just want to do a quick plug for,25.14,3.719
my patreon uh all tears get you access,26.64,4.32
to the private Discord server and then I,28.859,4.2
have a few higher tiers that uh come,30.96,4.32
with a one-on-one conversations and that,33.059,4.201
sort of thing so anyways back to the,35.28,5.16
video so first I wanted to share with,37.26,5.639
you guys uh the universal model of,40.44,4.98
Robotics so it has it's basically three,42.899,5.16
steps input processing and output or,45.42,4.38
sensing processing and controlling as,48.059,3.241
this Graphics shows,49.8,4.38
now this is the most basic cognitive,51.3,4.5
architecture that you can come up with,54.18,4.199
for artificial general intelligence it,55.8,4.439
needs input from the outside world from,58.379,3.66
the environment of some kind whether,60.239,3.3
it's a virtual environment digital,62.039,3.841
environment physical environment or,63.539,4.62
whatever cyber cybernetic environment,65.88,4.68
and then it needs some kind of internal,68.159,4.981
processing that includes memory task,70.56,4.2
construction executive function,73.14,4.519
cognitive control that sort of stuff,74.76,5.82
learning is another internal process and,77.659,5.381
then finally controlling or output it,80.58,4.859
needs to do something to act on the,83.04,5.1
world or its environment whether that's,85.439,4.801
just putting out you know text in a in,88.14,3.6
the form of a chat bot or if it's got,90.24,4.86
robotic hands that sort of thing so when,91.74,5.519
I talk about artificial general,95.1,3.9
intelligence being a system it's never,97.259,4.141
going to just be a model right even if,99.0,4.14
you have the most sophisticated model in,101.4,3.6
the world all that it's doing is the,103.14,3.839
processing part you also need the,105.0,4.799
sensing and controlling aspect and but,106.979,4.621
even above and beyond that each,109.799,4.081
components is going to be much more,111.6,4.32
complicated,113.88,4.08
so before we get into the rest of the,115.92,4.08
video I also want to talk about the form,117.96,4.26
factors that AGI is going to take so we,120.0,3.899
just established the simplest kind of,122.22,3.96
cognitive architecture but then there's,123.899,3.781
other things to consider because when,126.18,3.66
you think of AGI you might think of some,127.68,4.08
nebulous entity like Skynet but where,129.84,3.6
does it physically live,131.76,3.42
what is the hardware what is the,133.44,4.2
software where where is it physically,135.18,4.02
located because it's not magic right,137.64,3.179
it's not going to just run in the dirt,139.2,3.179
or something like that it needs to,140.819,4.381
actually have Hardware to run on so,142.379,5.761
there's three overarching categories,145.2,5.28
that I came up with so first is cloud,148.14,5.28
AGI so Cloud AGI is this is the stuff,150.48,4.38
that's gonna one it's going to be,153.42,3.06
created first just because of the amount,154.86,3.54
of compute and power available in data,156.48,5.22
centers so this is uh Enterprise grade,158.4,7.74
or data center grade AGI systems they,161.7,7.14
are in specialized buildings all over,166.14,5.22
the world but one of the biggest,168.84,4.38
constraints here is that there's limited,171.36,3.48
location and it takes a while to build,173.22,3.54
data centers right one of the things,174.84,3.96
that I think it was uh it was Elon,176.76,4.32
musker or Sam Altman said that you know,178.8,4.92
there are going to be limitations as to,181.08,5.28
the rate at which AGI can proliferate,183.72,5.519
namely the the rate at which we can,186.36,5.94
produce chips and also the rate at which,189.239,5.22
as I think Sam malt and said the you,192.3,4.439
know the concrete has to dry for data,194.459,3.121
centers,196.739,5.64
so this is uh one form factor that AGI,197.58,7.019
will take in terms of the the storage,202.379,4.801
the servers the network components that,204.599,4.86
will exist inside data centers so one,207.18,3.839
thing I wanted to watch it say is watch,209.459,3.841
out for uh fortified data centers these,211.019,3.841
are ones that are put in bunkers or if,213.3,3.659
you put Sam sites on top of it so that,214.86,4.26
you can't shut them down uh that was,216.959,3.42
kind of tongue-in-cheek I'm not actually,219.12,3.36
advocating for bombing data centers at,220.379,4.681
least not yet the next form factor is,222.48,5.16
Edge AGI so this is stuff that is going,225.06,4.22
to run in,227.64,3.84
self-contained servers that you can,229.28,4.9
basically plug in anywhere they're going,231.48,6.539
to be you know desktop size maybe larger,234.18,5.82
but the point is that pretty much all,238.019,3.481
you need is power and internet you don't,240.0,4.319
need a specialized building and they can,241.5,4.62
be moved on trucks they can be put in,244.319,3.961
ships airplanes that sort of stuff,246.12,3.72
because you can't really airlift an,248.28,4.019
entire data center so basically Edge is,249.84,4.619
something is just one size down from,252.299,3.481
data center you don't need a specialized,254.459,2.52
building you don't need specialized,255.78,3.66
cooling they can run anywhere,256.979,3.72
um and they're so in that respect,259.44,2.94
they're more portable but they're not,260.699,4.741
necessarily going to be as powerful at,262.38,5.039
least or not as energy intensive and,265.44,5.039
energy dense as a data center or a cloud,267.419,4.021
Center,270.479,4.141
and then finally ambulatory AGI this is,271.44,5.46
the embodied stuff such as C-3PO and,274.62,3.98
Commander data which I have imaged here,276.9,4.98
they're self-contained meaning that all,278.6,5.14
the systems that they need are within,281.88,4.319
their chassis within their robotic body,283.74,5.519
and they can move on their own so that's,286.199,4.5
basically the difference between an edge,289.259,5.101
AGI and an ambulatory AGI is uh they,290.699,5.401
might have roughly the same components,294.36,4.14
but it's one is accompanied with a,296.1,6.12
robotic uh chassis now one thing to keep,298.5,6.18
in mind is that all of these things are,302.22,4.86
intrinsically networkable meaning they,304.68,4.32
can communicate over digital networks,307.08,4.559
whether it's Wi-Fi or you know fiber,309.0,5.16
optic backbone networks or even you know,311.639,4.141
Satellite Communication like starlink,314.16,3.9
now that's that doesn't necessarily have,315.78,4.08
to be true because remember the model of,318.06,4.44
AGI is input processing and output that,319.86,5.76
input that input could be just eyes and,322.5,4.979
ears cameras and microphones that input,325.62,4.26
could also be network connections from,327.479,4.621
outside meaning that they could,329.88,4.44
communicate directly with each other via,332.1,5.28
you know like IRC or whatever so just,334.32,4.86
wanted to say that there are different,337.38,3.72
form factors that we should expect AGI,339.18,2.94
to take,341.1,3.3
with different trade-offs so one,342.12,5.22
advantage of ambulatory uh AGI you know,344.4,6.12
yes they will have less power uh and by,347.34,5.88
power I mean computational power but,350.52,5.28
they have the ability to go anywhere do,353.22,6.66
anything kind of like URI uh now that,355.8,6.959
being said the the amount of compute,359.88,4.8
resources that can be crammed into Data,362.759,3.66
Centers basically means that you can,364.68,4.38
puppet you know millions or billions of,366.419,5.161
peripheral robots rather than having it,369.06,4.44
fully self-contained and in a previous,371.58,3.899
video I talked about how we're likely to,373.5,4.639
see hybrid systems where you have,375.479,5.22
semi-autonomous peripherals that have,378.139,4.481
some intelligence but not a whole lot of,380.699,4.201
intelligence and you see this in movies,382.62,4.919
like Will Smith's iRobot as well as the,384.9,4.38
Matrix where the the drones the,387.539,3.78
squiddies and the Matrix they're,389.28,3.479
semi-autonomous but they are still,391.319,3.301
centrally controlled by a much more,392.759,4.021
powerful intelligence so you're probably,394.62,3.72
not going to see it all one or the other,396.78,3.06
you're probably going to see hybrids,398.34,3.54
where you've got peripheral robots that,399.84,4.02
are either fully autonomous or,401.88,4.92
semi-autonomous or puppeted by stronger,403.86,4.98
Central intelligences that being said,406.8,4.26
you can also create droids there's no,408.84,4.5
reason that we could not create fully,411.06,4.38
self-contained machines that don't,413.34,4.74
really have any network connectivity,415.44,4.86
um to the to other machines,418.08,3.899
that being said they would be at a,420.3,3.78
distinct disadvantage and what I mean by,421.979,4.5
that is that if you create swarm,424.08,4.8
intelligence or Wireless federations of,426.479,4.741
machines they can perform cognitive,428.88,6.24
offload or share computational resources,431.22,6.66
so for instance rather than and this is,435.12,4.44
how the Geth work in Mass Effect by the,437.88,4.439
way so rather than have every single,439.56,6.06
machine have to think about the entire,442.319,5.94
plan the entire strategy most of them,445.62,4.919
Focus only on their primary task and,448.259,5.401
then any surplus compute computational,450.539,6.121
power they have is dedicated towards you,453.66,5.52
know running algorithms for for the big,456.66,4.68
the big brain the hive mind,459.18,4.56
this is all hypothetical but one thing,461.34,4.32
that I want to point out is that many,463.74,3.54
many many machines work like this,465.66,4.56
already and what I mean by that is the,467.28,4.38
simplest version that many people are,470.22,3.9
probably aware of is if you have like,471.66,3.84
Bluetooth speakers or smart speakers,474.12,3.6
like Sonos or whatever those form a,475.5,5.819
wireless Federation uh ditto for like,477.72,5.28
your Amazon Alexa's and other things,481.319,4.261
like that those intrinsically form mesh,483.0,4.259
networks or Wireless federations meaning,485.58,3.48
that they can work together and,487.259,4.321
communicate now when you add artificial,489.06,5.46
intelligence to that then they can share,491.58,4.559
thinking and messaging and that sort of,494.52,2.64
stuff so that's what I mean by,496.139,4.801
federations or or wireless networks of,497.16,6.599
of AI okay so now you're familiar with,500.94,4.74
the background of how you know some of,503.759,5.041
the systemic aspects of it there's a few,505.68,4.919
default metrics of power so when I say,508.8,3.359
power I don't necessarily just mean,510.599,3.721
electricity although certainly all of,512.159,3.961
these things do require electricity to,514.32,2.82
run,516.12,4.26
so first is processing power so for,517.14,5.1
instance you might hear the term flops,520.38,4.2
which is floating Point operations per,522.24,7.68
second uh you also hear CPU GPU TPU and,524.58,7.02
then there's parallel parallelization,529.92,4.02
which means that you have many of these,531.6,5.1
things working together so processing,533.94,4.62
power is one component of the total,536.7,4.68
amount of power in the hardware layer so,538.56,4.44
this is all strictly Hardware layer I'm,541.38,3.48
not talking about parameter models,543.0,4.74
because I I don't really care about how,544.86,4.68
many parameters a model has there's lots,547.74,4.02
of ways to make intelligent machines,549.54,4.2
deep neural networks are currently the,551.76,4.079
best way but we're also discovering,553.74,3.719
efficiencies where you can kind of pair,555.839,3.661
them down you can distill them and make,557.459,3.541
them more efficient meaning that you can,559.5,3.36
on the same piece of Hardware you can,561.0,4.8
run more of them in parallel or you can,562.86,5.52
run one much faster so the underlying,565.8,3.96
Hardware is still going to be the,568.38,3.8
primary bottleneck or primary constraint,569.76,5.1
all else considered,572.18,5.68
uh memory so this is Ram it also,574.86,5.36
includes memory accelerators or caching,577.86,5.46
storage has to do with bulk data your,580.22,4.9
databases your archives your backups,583.32,4.44
this is when you say like hard drive or,585.12,4.74
SSD or you know storage area network,587.76,4.259
that sort of thing and then networking,589.86,4.68
is the the uplinks and downlinks this is,592.019,4.741
the the fiber optic connections the,594.54,3.72
wireless connections the satellite,596.76,3.9
connections that sort of thing so these,598.26,4.38
are the kind of the the rudimentary,600.66,4.799
parts that all AGI are going to run on,602.64,5.879
uh and this is just the brains too this,605.459,4.621
is not the peripherals this is not the,608.519,3.241
robots but this is what's going to,610.08,4.259
dictate or constrain how fast it is now,611.76,5.94
again like I said uh different neural,614.339,5.221
networks are going to operate at,617.7,3.42
different efficiencies so for instance,619.56,5.82
uh you know gpt4 is out now gpt5 might,621.12,6.96
be the same size it might be bigger but,625.38,4.82
then we're also finding open source,628.08,6.06
research like the Orca alpaca llama that,630.2,5.92
are getting like ninety percent of the,634.14,4.56
performance but at like one tenth or one,636.12,4.98
hundredth of the size and so you have a,638.7,4.199
trade-off of intelligence and versus,641.1,4.56
speed and power and we'll talk more a,642.899,5.221
lot more about that in the future of,645.66,4.739
this video at near the middle and end of,648.12,4.2
this video about how trading off,650.399,4.801
intelligence for Speed is often a more,652.32,4.56
advantageous strategy and how this,655.2,4.199
figures into solving the control problem,656.88,5.28
and solving alignment,659.399,4.981
um okay so we kind of set the stage as,662.16,4.98
as to how AGI is probably going to look,664.38,5.639
so let's talk about the early ecosystem,667.14,6.18
of AGI so in the coming years we're,670.019,4.921
going to be building millions and then,673.32,3.24
billions of autonomous and,674.94,4.5
semi-autonomous agents so at first these,676.56,4.2
agents are going to be purely digital,679.44,3.44
right you know a,680.76,4.56
semi-autonomous slack bot a,682.88,4.54
semi-autonomous Discord bot people are,685.32,4.32
already building these right and some of,687.42,3.72
them have the ability to modify their,689.64,2.879
own code some of them have the ability,691.14,3.78
to learn many of them don't most of them,692.519,4.32
use frozen llms in the background,694.92,3.659
meaning that they're that their,696.839,4.201
cognitive capacity is pretty much capped,698.579,4.981
by its backing model,701.04,5.76
now that being said as these agents,703.56,5.279
become more autonomous they go from,706.8,4.2
semi-autonomous to autonomous this will,708.839,4.201
create a competitive landscape,711.0,4.86
and what I mean by that is that humans,713.04,5.28
will have the ability to build and,715.86,5.159
destroy these models for basically,718.32,4.62
arbitrary reasons because you want to or,721.019,3.781
because you don't like it or whatever,722.94,4.26
so that means that we will be selecting,724.8,5.52
those agents those uh those models those,727.2,6.18
llms and those pieces of software that,730.32,4.32
are going to be more helpful more,733.38,3.6
productive and more aligned so this,734.64,4.08
creates selective pressure basically,736.98,3.299
saying that there's going to be a,738.72,2.88
variety there's going to be millions or,740.279,3.24
billions of Agents out there some of,741.6,3.9
them are going to get the ax and some of,743.519,4.141
them are going to be selected to say hey,745.5,4.079
we like we like you we're going to keep,747.66,2.88
you around,749.579,3.901
so there's a few off the cuff selective,750.54,4.799
pressures that we can imagine basically,753.48,3.84
why do you choose an app right why do,755.339,3.361
you choose to use an app why do you,757.32,3.54
choose to uninstall an app that's kind,758.7,3.36
of the level that we're talking about,760.86,3.9
here so first is functional utility how,762.06,4.019
useful is it,764.76,3.6
how much does it help you is it fast,766.079,3.541
enough does it have a good user,768.36,3.5
experience is the user interface,769.62,5.88
created correctly is it adding value to,771.86,7.06
your life is it worth using the second,775.5,6.079
part is speed and efficiency,778.92,6.06
basically if something takes four weeks,781.579,5.141
to give you a good answer but another,784.98,4.2
thing takes 10 minutes even if it's not,786.72,4.559
quite as good that speed is going to be,789.18,4.14
super super valuable but then there's,791.279,4.081
also energetic efficiency and cost,793.32,4.94
efficiency more often than not,795.36,5.34
individuals and businesses will choose,798.26,5.56
the solution that is good enough but,800.7,5.04
also much cheaper it doesn't have to be,803.82,4.019
perfect it just has to be good enough,805.74,4.86
and then finally apparent alignment and,807.839,5.101
so I use the the word apparent alignment,810.6,4.38
to basically mean things that appear to,812.94,4.199
be tame things that appear to be user,814.98,4.56
friendly uh and this is what uh tools,817.139,5.521
like rlhf do which one thing that rlhf,819.54,4.68
does is like wolves which we'll talk,822.66,4.619
about in a second are the rlhf,824.22,4.2
reinforcement learning with human,827.279,4.68
feedback forces gpt4 to dumb itself down,828.42,6.539
so that it better serves us uh and that,831.959,4.56
makes us feel safe because it's,834.959,4.32
basically pretending to be more like us,836.519,5.641
to speak on our terms and to mimic our,839.279,4.981
level of intelligence now that being,842.16,3.419
said,844.26,2.579
um one thing that I do want to point out,845.579,4.021
is that gpd4 the underlying model is,846.839,4.74
superior to anything that we have seen,849.6,5.239
in the public every version of chat GPT,851.579,6.021
has basically been kind of a little bit,854.839,6.101
hamstrung so we shall we say uh from the,857.6,6.82
the total capacity of gpt4,860.94,5.699
so what I call this is domestication and,864.42,4.8
supplication think of dogs and wolves,866.639,5.101
this little pomeranian descended from,869.22,4.98
wolves wolves used to be apex predators,871.74,4.62
wolves are also are much more,874.2,4.02
intelligent than dogs,876.36,6.419
so when you look at the early days of,878.22,6.84
AGI when we still have the off switch,882.779,3.781
and we have the power to delete,885.06,3.12
everything,886.56,3.899
we should expect some of the following,888.18,5.7
evolutionary pressures to kind of shape,890.459,6.421
the way that AGI evolves and adapts so,893.88,5.519
first we'll probably be selecting for,896.88,4.079
machines that are okay with being turned,899.399,4.201
off in the early days you don't,900.959,4.38
necessarily want your toaster fighting,903.6,3.479
with you when when you're done you know,905.339,3.3
toasting your bread it's time for it to,907.079,3.661
turn off and so we're probably going to,908.639,3.781
select machines and architectures and,910.74,3.779
models that are more or less okay with,912.42,3.539
being switched off that they don't have,914.519,3.801
a sense of death or a fear of death,915.959,4.62
we're also going to select machines that,918.32,4.6
are more eager to please just the same,920.579,5.281
way that with uh dogs have been bred and,922.92,5.159
selected to be very very eager to please,925.86,3.599
us,928.079,3.181
we're also going to select machines that,929.459,3.781
don't fall into the uncanny valley and,931.26,3.48
so what I mean by that is the uncanny,933.24,4.02
valley of when you're interacting with a,934.74,4.56
machine that you sense is an alien,937.26,3.78
intelligence it will make you very very,939.3,4.2
deeply uncomfortable as an autistic,941.04,4.799
person as someone who is neurodiverse I,943.5,4.139
have to modulate the way that I speak,945.839,4.68
and act around neurotypical people,947.639,5.401
because I fall into the same uncanny,950.519,5.94
valley right and some uh some CEOs out,953.04,5.039
there get teased for this for instance,956.459,3.24
Mark Zuckerberg I don't know if he's,958.079,3.841
actually autistic but he certainly pings,959.699,4.44
that radar where it's like okay he,961.92,3.539
obviously does not think the way that,964.139,3.361
the rest of us do and he also behaves,965.459,4.44
differently so Mark Zuckerberg like many,967.5,5.1
of us uh people on the Spectrum kind of,969.899,4.56
fall into that uncanny valley again I,972.6,4.44
don't know but uh,974.459,5.221
he certainly looks uh he he plays the,977.04,6.06
part so but the idea is that when you,979.68,5.339
interact with something that give that,983.1,4.26
kind of gives you the heebie-jeebies you,985.019,3.901
don't like it,987.36,4.14
now that being said we will still select,988.92,4.68
machines that are smarter to a certain,991.5,3.66
degree because you don't want something,993.6,3.659
to be too smart but you do also want it,995.16,3.599
to be smart enough to be very very,997.259,2.64
useful,998.759,3.181
another selective pressure is that we're,999.899,3.421
going to choose things that are stable,1001.94,3.06
robust and resilient so remember when,1003.32,3.12
Bing first came out and it was,1005.0,3.66
completely unhinged you could get it,1006.44,3.78
into you could coax it into like,1008.66,4.619
threatening you and you know threatening,1010.22,4.859
to take over the world and you know,1013.279,3.721
threatening to see you all kinds of,1015.079,4.2
crazy stuff so obviously that version,1017.0,4.44
got shut down really quick,1019.279,4.321
um you're also going to select uh models,1021.44,4.139
and agents that are more resilient,1023.6,3.839
against those kinds of adversarially,1025.579,3.061
attacks,1027.439,3.0
um whether they are accidental right you,1028.64,2.88
don't want something to be mentally,1030.439,3.48
unstable just on its own right like Bing,1031.52,6.0
was originally uh or Tay tweets but you,1033.919,5.101
also want it to be resilient against,1037.52,4.74
being manipulated by other hostile,1039.02,5.039
actors because imagine that your,1042.26,3.9
personal AI assistance just becomes,1044.059,4.02
unhinged one day because a hacker,1046.16,3.72
somewhere was messing with it so,1048.079,3.48
Security will be one of the selective,1049.88,3.0
pressures,1051.559,3.841
likewise you'll you'll as part of The,1052.88,4.38
Uncanny Valley thing you're going to,1055.4,2.94
select things that are more,1057.26,2.82
comprehensible to us that are better at,1058.34,3.78
explaining themselves to us so that,1060.08,3.78
includes transparency emotional,1062.12,5.04
intelligence and so on uh and then again,1063.86,5.46
apparent alignment things that's that,1067.16,4.74
that uh don't kind of trigger your,1069.32,4.62
existential dread because there have,1071.9,4.56
been times for instance where I've been,1073.94,5.94
working with chat GPT uh on the API side,1076.46,5.16
and kind of giving it different sets of,1079.88,4.62
instructions and even just a slight,1081.62,5.58
misalignment between how I approach,1084.5,4.86
moral problems and how this model,1087.2,4.56
approaches moral problems are really,1089.36,5.939
deeply unsettling and so it's like there,1091.76,5.58
there's been a few times where it's like,1095.299,3.301
I'm working with this thing and I'm,1097.34,3.78
building a semi-autonomous chat Bots and,1098.6,3.959
it's like I understand it's reasoning,1101.12,3.419
but it's like oh that's really cringe,1102.559,3.901
and it kind of scares me,1104.539,4.38
um so in that respect it's like let's,1106.46,4.14
change this model so that it's not quite,1108.919,2.821
so scary,1110.6,2.52
and I'm saying that this is possible,1111.74,4.22
today that if you use the chat GPT API,1113.12,5.4
you can give it programming you can give,1115.96,5.26
it reasoning and goals uh and and,1118.52,4.86
patterns of thought that are already,1121.22,4.5
already on the kind of in the midst of,1123.38,3.84
that uncanny valley,1125.72,4.38
uh then you can uh we'll also select for,1127.22,4.92
things that are more uh docile so,1130.1,3.6
basically how dogs you know you can pet,1132.14,2.7
them you can wrestle with them and,1133.7,2.339
they're probably not going to eat your,1134.84,3.719
face uh plastic and so things that are,1136.039,4.921
changeable or adaptable and Cooperative,1138.559,4.201
those are other things that we're going,1140.96,4.74
to select for so basically dogs are,1142.76,5.159
dumber than wolves and the reason for,1145.7,3.96
this is what I call capability,1147.919,4.14
equilibrium which will unpack more in,1149.66,5.28
the in a few slides but the the very,1152.059,4.801
short version of capability equilibrium,1154.94,3.96
is that your intellect must be equal to,1156.86,3.84
the task and if your intellect is above,1158.9,3.779
the task there's no advantage and in,1160.7,3.54
fact there can be disadvantages because,1162.679,3.901
of the costs associated with higher,1164.24,4.02
intelligence,1166.58,4.02
okay so I've talked about this idea,1168.26,5.76
plenty instrumental convergence uh this,1170.6,6.0
was coined by Nick Bostrom in 2003 who,1174.02,4.44
is a philosopher,1176.6,3.959
um the very short version is that,1178.46,3.959
regardless of the terminal goals or main,1180.559,5.101
objectives that a machine has uh AGI,1182.419,5.161
will likely pursue intermediate or,1185.66,4.259
instrumental goals or basically other,1187.58,3.9
stuff that it needs in order to meet,1189.919,5.401
those other ends so whatever like let's,1191.48,5.88
say you give an AGI the goal of like,1195.32,5.46
getting them getting a a spacecraft to,1197.36,5.58
Alpha Centauri well it's going to need a,1200.78,3.66
laundry list of other stuff to do that,1202.94,3.72
it's going to need resources like power,1204.44,6.0
materials electricity data it's going to,1206.66,6.06
need self-preservation because if the,1210.44,4.44
machine goes offline it will realize,1212.72,5.1
that is a failure State and so we'll try,1214.88,4.98
and avoid those failure conditions by,1217.82,3.96
preserving its own existence,1219.86,3.48
another thing is that it will probably,1221.78,3.06
decide that it needs self-improvement,1223.34,4.079
because if it realizes that its current,1224.84,5.219
capability its current capacity is not,1227.419,4.681
equal to the task if it's too dumb it's,1230.059,3.541
going to say okay well I need to raise,1232.1,3.06
my intelligence so that I'm equal to,1233.6,2.76
that task,1235.16,3.66
now that being said Nick boster makes,1236.36,4.559
quite a few uh assumptions about the way,1238.82,5.099
that AGI will work so for instance he,1240.919,5.521
kind of imagines that um AGI is going to,1243.919,3.961
be very single-minded and somewhat,1246.44,4.32
monolithic uh basically mindlessly,1247.88,4.98
pursuing one goal which I would actually,1250.76,3.96
classify this as a middle intelligence,1252.86,4.319
rather than a high intelligence AGI and,1254.72,3.9
we'll talk about that in a little bit as,1257.179,2.941
well,1258.62,3.6
he also assumes that it's going to lack,1260.12,3.9
other forces or competitive pressures,1262.22,4.26
and that these uh might exist in a,1264.02,5.399
vacuum basically that resource,1266.48,4.5
acquisition and self-preservation and,1269.419,4.021
self-improvement are going to exist in,1270.98,5.16
in the absence of other forces or,1273.44,5.22
pressures such as competitive pressures,1276.14,4.38
or internal pressures which I will talk,1278.66,2.879
about more,1280.52,3.42
and finally that they will lack a higher,1281.539,5.361
purpose or the ability to be completely,1283.94,5.82
self-determining so basically what I,1286.9,7.0
mean by that is that okay yes once a,1289.76,6.48
machine is intelligent enough it can you,1293.9,3.72
know you can say like hey I want you to,1296.24,3.72
get us to Alpha Centauri and the AG I,1297.62,4.02
might say like okay whatever I don't,1299.96,2.88
think that's a good goal so I'm going to,1301.64,4.26
choose my own goal uh which that being,1302.84,5.459
said even if AGI become fully autonomous,1305.9,4.019
and you know kind of give a flip us the,1308.299,3.36
bird they're probably still going to,1309.919,3.421
benefit from some convergence which,1311.659,4.621
we'll talk about as well uh now what I,1313.34,4.62
want to point out is that there is a,1316.28,4.74
huge parallel between evolutionary,1317.96,5.04
pressures and selective pressures and,1321.02,4.08
this instrumental convergence basically,1323.0,4.86
all life forms all organisms have have,1325.1,5.4
converged on a few basic principles such,1327.86,4.98
as get energy somehow right there's,1330.5,4.02
autotrophs which make their own energy,1332.84,3.839
plants and there's heterotrophs which,1334.52,5.039
take energy from other uh creatures,1336.679,6.24
uh they through either predation or,1339.559,4.86
consuming you know plant matter or,1342.919,2.341
whatever,1344.419,4.081
uh so when you operate in a competitive,1345.26,5.039
environment there's there's going to be,1348.5,3.6
convergence around certain strategies,1350.299,4.081
this is true for evolution and this will,1352.1,4.68
also be true more or less with some,1354.38,4.56
variances in the competitive environment,1356.78,4.92
between intelligent machines that being,1358.94,4.38
said because they have a fundamentally,1361.7,4.02
different substrate there will be we,1363.32,3.839
should anticipate that there will be,1365.72,2.88
some differences,1367.159,4.201
between organisms the way that organisms,1368.6,4.74
evolve and the way that machines evolve,1371.36,3.78
not the least of which is that machines,1373.34,3.6
can rewrite their own source code we,1375.14,3.24
cannot rewrite our own source code at,1376.94,2.58
least not,1378.38,3.0
um not in a hurry it takes us quite a,1379.52,3.48
long time,1381.38,4.679
okay so the idea that one of the ideas,1383.0,4.559
that I'm introducing and I've been,1386.059,2.821
talking about this for a while is,1387.559,4.561
epistemic Convergence so instrumental,1388.88,4.86
convergence talks about the objective,1392.12,3.539
behaviors and strategies that machines,1393.74,4.74
adopt epistemic convergence is well let,1395.659,3.921
me just read you the definition,1398.48,3.059
epistemic convergence is the principle,1399.58,4.2
that within any given information domain,1401.539,4.561
sufficiently sophisticated intelligent,1403.78,4.72
agents given adequate time and data will,1406.1,3.959
progressively develop more precise,1408.5,3.299
accurate and efficient models of that,1410.059,4.081
domain these models aim to mirror the,1411.799,3.781
inherent structures principles and,1414.14,3.539
relationships within that domain over,1415.58,3.78
time the process of learning testing and,1417.679,3.841
refining understanding will lead these,1419.36,4.02
agents towards a shared comprehension of,1421.52,4.74
the Dom domain's fundamental truths in,1423.38,5.299
other words to put it more simply,1426.26,4.919
intelligent entities tend to think alike,1428.679,4.061
especially when they are operating in,1431.179,3.841
the same competitive space,1432.74,5.88
so you and I All Humans we operate on,1435.02,5.58
planet Earth in the universe in the,1438.62,4.74
Milky Way galaxy because of that similar,1440.6,5.12
context scientists all over the world,1443.36,4.62
repeatedly come to the same conclusions,1445.72,5.079
even when there are boundaries such as,1447.98,5.22
linguistic and cultural differences and,1450.799,4.441
this was most starkly seen during the,1453.2,4.2
Cold war between uh America and the,1455.24,3.2
Soviet Union,1457.4,4.08
whereby scientists independently whether,1458.44,4.239
it was nuclear physicist or,1461.48,3.199
astrophysicist or whatever,1462.679,5.041
rocket Engineers came to the same exact,1464.679,4.781
conclusions about the way that the,1467.72,3.78
Universe worked and also found the same,1469.46,6.06
optimization uh uh patterns even though,1471.5,5.58
there was no communication between them,1475.52,3.899
and so epistemic convergence there's,1477.08,4.86
obviously uh evidence of that happening,1479.419,4.081
because humans we have the same,1481.94,4.619
fundamental Hardware right we're all the,1483.5,5.64
same species and so therefore you have,1486.559,4.801
similarities between the agents now that,1489.14,3.659
being said,1491.36,4.02
uh there is also evidence of epistemic,1492.799,5.641
convergence between between species and,1495.38,5.52
so what I mean by that is even animals,1498.44,4.44
that have a very very different taxonomy,1500.9,4.38
such as ravens and crows and octopuses,1502.88,5.4
they all still demonstrate very similar,1505.28,5.22
problem solving strategies even though,1508.28,4.86
that octopuses have a very decentralized,1510.5,4.38
cognition that a lot of their cognition,1513.14,4.08
occurs in their arms for instance you,1514.88,4.14
can't get much more alien from us than,1517.22,3.72
that they still adopt very similar,1519.02,3.72
problem-solving strategies and learning,1520.94,3.42
strategies that we do,1522.74,4.62
uh again despite the fact that they are,1524.36,4.62
they live underwater they have a very,1527.36,3.78
different body plan so on and so forth,1528.98,4.92
so I personally suspect that there is a,1531.14,4.56
tremendous amount of evidence for,1533.9,4.2
epistemic convergence and we should we,1535.7,4.979
should expect epistemic convergence and,1538.1,5.939
encourage epistemic convergence uh and,1540.679,5.341
for reasons that I'll go over uh later,1544.039,5.061
in the video but basically,1546.02,6.42
AI agents will we should expect and help,1549.1,6.52
them to arrive at similar conclusions in,1552.44,4.26
the long run,1555.62,3.48
now let's talk about these evolutionary,1556.7,5.4
uh niches that will be developed at,1559.1,5.76
least in the in in the um uh the short,1562.1,4.26
term near term,1564.86,4.199
and what I mean by this is segments,1566.36,4.439
market segments where we will be,1569.059,4.62
deploying intelligent AGI systems so,1570.799,5.88
first is domestic uh personal and,1573.679,4.62
consumer grade stuff so this is going to,1576.679,4.021
be the AGI running on your MacBook this,1578.299,4.74
is going to be the AGI running in your,1580.7,5.339
kitchen uh these have a relatively,1583.039,7.201
benign set of tasks and also that uh,1586.039,6.781
that capability equilibrium is going to,1590.24,5.28
be uh pretty pretty low you only need to,1592.82,5.459
be so smart to cook dinner right this is,1595.52,5.1
not going to be you know the the AGI,1598.279,3.841
running in your microwave is not going,1600.62,3.419
to be working on quantum physics or,1602.12,3.84
Global economics,1604.039,4.321
now the next level up is going to be,1605.96,3.959
corporate and Enterprise so these are,1608.36,3.72
going to be these are going to be AGI,1609.919,3.721
systems that are tasks with solving,1612.08,4.14
relatively complex problems running,1613.64,5.34
entire companies Regulatory Compliance,1616.22,6.24
uh you know making SEC filings that sort,1618.98,6.66
of stuff uh CEOs digital CEOs digital,1622.46,5.339
Boards of directors uh the creative,1625.64,4.919
aspect of finding Market opportunities,1627.799,5.701
so this the intellectual challenge of,1630.559,5.461
those of that scale of problems is that,1633.5,5.88
much higher meaning that it would in,1636.02,5.399
order for an AGI to succeed there it's,1639.38,4.02
going to need to be a lot smarter than a,1641.419,5.64
personal or domestic AGI system and,1643.4,5.22
again there are going to be trade-offs,1647.059,4.62
the smarter a system becomes the more,1648.62,4.919
data it requires the more energy it,1651.679,4.261
requires the larger compute system that,1653.539,4.26
it requires and so you're going to want,1655.94,3.78
to satisfy so satisfice is basically,1657.799,4.201
meaning you find the level that is good,1659.72,4.559
enough to get the job done,1662.0,4.14
above that is going to be governmental,1664.279,5.28
and institutional AGI systems so these,1666.14,4.68
are the ones that are going to be,1669.559,3.24
conducting research whether it's,1670.82,3.839
scientific research or policy research,1672.799,4.141
or economic research and that is because,1674.659,5.041
governments are basically enormous,1676.94,4.92
corporations is one way to think of them,1679.7,4.68
that have a responsibility of managing,1681.86,5.28
you know resources and regulations and,1684.38,5.399
rules that affect millions of people and,1687.14,4.32
then of course governments communicate,1689.779,3.841
with each other but then above and,1691.46,3.959
beyond that there's also the scientific,1693.62,4.32
research aspect having AGI that are,1695.419,4.081
going to help with particle physics with,1697.94,4.02
with Fusion research with really pushing,1699.5,4.88
the boundaries of what science even,1701.96,5.819
knows and so that is an even larger,1704.38,5.08
intellectual task and even more,1707.779,3.841
challenging intellectual task and then,1709.46,4.5
finally above and beyond that the most,1711.62,4.2
competitive environment where AGI will,1713.96,4.14
be used is going to be in the military,1715.82,4.92
and what I mean by that is it's not,1718.1,4.86
necessarily uh those that are the most,1720.74,3.9
intelligent although the ability to,1722.96,4.68
forecast and anticipate is critical read,1724.64,6.48
Sun Tzu uh uh The Art of War right if,1727.64,4.98
you know yourself and you know the enemy,1731.12,3.36
then you can predict the outcome of a,1732.62,4.38
thousand battles uh and so in that in,1734.48,6.179
that respect uh the military domain of,1737.0,6.179
artificial general intelligence is the,1740.659,4.981
ultimate uh competitive sphere meaning,1743.179,5.701
that you win or you die and so these are,1745.64,4.26
going to be used to coordinate,1748.88,3.84
battlefields uh to run autonomous drones,1749.9,4.56
for intelligence and surveillance but,1752.72,3.959
also like I said for forecasting for,1754.46,4.92
anticipating what the enemy can and will,1756.679,3.6
do,1759.38,3.84
which means that it's basically a race,1760.279,4.321
condition and we'll talk more about the,1763.22,4.199
race condition as the video progresses,1764.6,4.92
so that capability equilibrium that I,1767.419,5.041
talked about uh quite simply refers to,1769.52,4.74
the state of optimal alignment between,1772.46,3.839
the cognitive capacity of any entity,1774.26,4.019
organic or otherwise and the,1776.299,4.081
intellectual demands of a specific task,1778.279,4.441
or role it is assigned there are three,1780.38,4.919
form three primary forces at play here,1782.72,5.579
one the intellectual demands of the task,1785.299,5.161
as I said earlier your toaster roll only,1788.299,4.561
ever needs to be so smart but if your,1790.46,4.02
toaster is actually Skynet it probably,1792.86,4.02
needs to be much smarter then there's,1794.48,4.079
the intellectual capacity of the agent,1796.88,3.24
if there's a mismatch between the,1798.559,3.6
intellectual capacity of the agent and,1800.12,3.779
the and the intellectual requirements of,1802.159,5.041
the task then you're either unable to to,1803.899,5.941
satisfy that task or you're super,1807.2,4.32
overqualified which is why I picked,1809.84,3.24
Marvin here,1811.52,3.36
um so Marvin is a character from,1813.08,3.599
Hitchhiker's Guide to the Galaxy and if,1814.88,2.88
you haven't read it you absolutely,1816.679,3.301
should there's also a good movie with,1817.76,4.86
Martin Freeman as as the protagonist,1819.98,5.76
he's basically bill boban in space uh,1822.62,5.279
very hapless character but anyways,1825.74,5.52
Marvin was a prototype who was one of,1827.899,5.16
the most intelligent robots ever built,1831.26,3.899
and they just have him doing like basic,1833.059,4.081
stuff around the task oh and he was,1835.159,5.4
voiced by Snape by the way and so one of,1837.14,5.279
the quotations from him is here I am,1840.559,4.321
with a brain the size of of a planet and,1842.419,3.661
they asked me to pick up a piece of,1844.88,3.48
paper call that job satisfaction I don't,1846.08,4.38
so that is a mismatch where Marvin is,1848.36,3.6
way more intelligent than what he's,1850.46,3.599
being used for and so that means that,1851.96,4.02
this is an inefficient use of resources,1854.059,5.881
he probably cost more than you know to,1855.98,6.24
build and run than he needed to,1859.94,4.5
and then finally the third variable is,1862.22,3.98
the cost of intellectual capacity,1864.44,5.04
generally speaking uh as intelligence,1866.2,5.38
goes up there are there are problems,1869.48,3.419
associated with that whether it's,1871.58,2.88
training time of the models the amount,1872.899,3.601
of data required for the models uh the,1874.46,4.26
amount of energy that it requires to run,1876.5,5.94
that particular robot uh the amount of,1878.72,5.939
ram required to to load that model right,1882.44,3.54
so for instance one of the things that,1884.659,4.02
people are seeing is that it requires,1885.98,4.38
millions of dollars worth of compute,1888.679,5.22
Hardware to run gpt4 but you can run,1890.36,6.059
um Orca on a laptop right so which one,1893.899,5.28
is is cheaper and easier to run even if,1896.419,4.681
one of them is only 50 as good as the,1899.179,4.86
other it costs a thousand times less,1901.1,5.88
uh to to build train and run now that,1904.039,5.401
being said you look at the at the case,1906.98,5.52
of dogs dogs are dumber than wolves,1909.44,4.92
because dogs don't need to be as smart,1912.5,4.08
as independent apex predators because,1914.36,4.02
apex predators like wolves out in the,1916.58,3.9
wild they need to be smart enough to out,1918.38,4.44
think their prey dogs they don't need to,1920.48,3.84
be that smart so they're not that smart,1922.82,4.2
in fact it does not be it it is not good,1924.32,4.8
for dogs to be too intelligent anyone,1927.02,4.56
who has owned uh really intelligent dogs,1929.12,4.799
like I had a I had a dog who was too,1931.58,4.44
smart for his own good died about a year,1933.919,4.321
ago he was clever enough to manipulate,1936.02,4.08
people and other dogs and you know get,1938.24,4.319
into the food when he wasn't supposed to,1940.1,5.1
Huskies German Shepherds Border Collies,1942.559,4.381
the more intelligent dogs are the more,1945.2,3.42
mischievous ones they are the Escape,1946.94,3.359
artists they are the ones that are going,1948.62,4.26
to pretend one thing and then you know,1950.299,4.801
so on and so forth so intelligence is,1952.88,4.56
not always adaptive so there can be,1955.1,4.26
multiple Dimensions to the cost of,1957.44,3.9
intellectual capacity,1959.36,3.72
uh not the least of which is you could,1961.34,3.54
end up like poor Marvin here where,1963.08,3.18
you're too smart for your own good and,1964.88,2.82
then you just end up depressed all the,1966.26,3.419
time granted he was deliberately given,1967.7,3.959
the depressed affect,1969.679,4.921
so all this being said is what I've been,1971.659,5.101
building up to is what um I call and,1974.6,3.959
what is generally called a terminal race,1976.76,4.74
condition so terminal race condition is,1978.559,4.921
basically what we could end up moving,1981.5,4.26
towards as we develop more and more,1983.48,5.1
powerful sophisticated and more uh fully,1985.76,6.6
autonomous AGI systems basically this,1988.58,5.819
the terminal race condition is where for,1992.36,4.62
any number of reasons uh competition,1994.399,5.88
between AGI will fully bypass that,1996.98,5.579
capability equilibrium so say for,2000.279,5.4
instance uh you know your toaster is,2002.559,5.22
competing with another brand and it's,2005.679,3.6
like oh well I need to be a smarter,2007.779,3.961
toaster in order to be a better toaster,2009.279,5.161
for you so that you don't throw me away,2011.74,4.439
now that's obviously a very silly,2014.44,4.38
example but a very real example would be,2016.179,4.461
competition between corporations,2018.82,3.959
competition between nations and,2020.64,4.899
competition between militaries wherein,2022.779,5.041
basically it's no longer just a matter,2025.539,4.201
of being intelligent enough to satisfy,2027.82,4.02
the demands of that task to satisfy the,2029.74,4.919
demands of that initial competition it,2031.84,4.92
is then it's less about that and it,2034.659,4.081
becomes more about out competing the,2036.76,4.019
other guy it's like a chess match right,2038.74,4.2
you know the other guy got a higher ELO,2040.779,4.081
score so you need to be smarter and then,2042.94,3.719
you're smarter so now the other guy,2044.86,4.08
tries to be smarter than you,2046.659,5.161
and so because of this because of this,2048.94,4.739
pressure and as I mentioned earlier some,2051.82,3.24
of the trade-offs might actually force,2053.679,3.901
you to to prioritize speed over,2055.06,4.38
intelligence and so we see we actually,2057.58,3.72
see this in volume trading in in,2059.44,4.199
algorithmic and Robo trading on the,2061.3,4.26
stock market where financial,2063.639,4.321
institutions will actually use less,2065.56,4.74
sophisticated algorithms to execute,2067.96,5.219
transactions but because they are faster,2070.3,5.28
they uh will still out compete the other,2073.179,5.46
guy so in some in this respect you might,2075.58,5.819
actually incentivize AGI to dumb,2078.639,5.52
themselves down just so that they can be,2081.399,4.5
faster so that they can out-compete the,2084.159,3.48
other guy so that's what I mean by a,2085.899,3.96
race condition it is a race to higher,2087.639,4.441
intelligence but it is also a race to,2089.859,3.661
being more efficient and therefore,2092.08,3.839
faster and then there's also going to be,2093.52,4.2
a trade-off these machines might,2095.919,4.141
ultimately trade off their accuracy,2097.72,4.32
their ethics the amount of time they,2100.06,3.96
spend thinking through things in order,2102.04,4.26
to be faster and so you actually see,2104.02,4.92
this in chess computers where you can,2106.3,4.88
doing a chess computer or a chess,2108.94,5.159
algorithm to say okay spend less time,2111.18,4.36
thinking about this so that you can make,2114.099,4.98
the decision faster in many cases the,2115.54,5.88
first one to move even if it's not the,2119.079,4.741
best plan but moving faster will give,2121.42,4.5
you a tactical or strategic advantage,2123.82,4.44
and this includes corporations Nations,2125.92,4.32
and militaries,2128.26,4.62
so a terminal race condition to me,2130.24,4.04
represents,2132.88,3.78
according to my current thought this is,2134.28,5.559
the greatest uh component of existential,2136.66,4.439
risk we Face from artificial,2139.839,3.721
intelligence and I don't think that,2141.099,3.661
corporations are going to have enough,2143.56,2.88
money to throw at the problem to make,2144.76,4.2
truly dangerous AGI the only entities,2146.44,4.02
that are going to have enough money to,2148.96,3.899
throw at this to make to to basically,2150.46,5.159
compete are going to be entire nations,2152.859,5.76
and the militaries that they run so,2155.619,4.681
basically it's going to be up to those,2158.619,4.861
guys to not enter into an uh the,2160.3,4.68
equivalent of a nuclear arms race but,2163.48,5.04
for AGI now that being said uh I have,2164.98,5.639
put a lot of thought into this so moving,2168.52,4.26
right along one thing to keep in mind is,2170.619,4.201
that there could be diminishing returns,2172.78,4.98
to increasing intelligence so basically,2174.82,5.279
there's a few possibilities one is that,2177.76,4.56
there could be a hard upper bound there,2180.099,4.201
might be a maximum level of intelligence,2182.32,4.019
that is actually possible and at that,2184.3,3.66
point all you can do is have more of,2186.339,4.74
them running in parallel uh it might be,2187.96,4.619
a long time before we get to that like,2191.079,3.721
we might be halfway there but we also,2192.579,4.02
might be down here we don't actually,2194.8,4.62
know if there is an upper bound to,2196.599,5.281
maximum intelligence uh but one thing,2199.42,4.439
that we can predict is that actually the,2201.88,4.5
cost as I mentioned earlier the cost of,2203.859,4.381
additional intelligence might go up,2206.38,3.36
exponentially you might need,2208.24,3.96
exponentially more data or more compute,2209.74,5.64
or more storage in order to get to that,2212.2,4.919
next level of intelligence,2215.38,3.479
and so you actually see this in the Star,2217.119,4.321
Wars Universe where droids are basically,2218.859,4.801
the same level of intelligence across,2221.44,4.62
the entire spectrum of the Star Wars,2223.66,3.959
Universe because there's diminishing,2226.06,3.66
returns yes you can build a more,2227.619,3.96
intelligent Droid but it's just not,2229.72,5.46
worth it so the the the total effective,2231.579,6.121
level of intelligence of AGI I suspect,2235.18,4.919
will follow a sigmoid curve now that,2237.7,3.899
being said there's always going to be,2240.099,4.081
some advantage to being smarter more,2241.599,4.861
efficient and so on but as with most,2244.18,4.14
fields of science I suspect this is,2246.46,3.48
going to slow down that we're going to,2248.32,3.539
have diminishing returns and that,2249.94,3.179
eventually we're going to kind of say,2251.859,3.961
like okay here's actually The Sweet Spot,2253.119,5.761
in terms of how much it's worth making,2255.82,5.94
your machine more intelligent,2258.88,6.479
so this leads to one uh one possibility,2261.76,7.8
and this is a personal pet Theory but,2265.359,5.581
basically I think that there's going to,2269.56,4.38
be a bell curve of existential risk and,2270.94,4.8
that is that minimally intelligent,2273.94,4.08
machines like your toaster are probably,2275.74,4.98
not going to be very dangerous the the,2278.02,5.16
total domain space of toasting your,2280.72,4.74
sandwich or toasting your bagel that's,2283.18,3.78
not a particularly difficult problem,2285.46,3.119
space and yes there might be some,2286.96,3.0
advantages to being slightly more,2288.579,3.961
intelligent but your toaster is not,2289.96,4.44
going to be sitting there Conjuring up,2292.54,4.44
you know a bio weapon and if it is you,2294.4,4.32
probably bought the wrong toaster,2296.98,4.56
now that being said the other end of the,2298.72,4.92
spectrum the maximally intelligent,2301.54,4.02
machines or the digital Gods as some,2303.64,3.78
people are starting to call them these,2305.56,3.48
are going to be so powerful that human,2307.42,3.12
existence is going to be completely,2309.04,3.66
inconsequential to them and what I mean,2310.54,5.039
by that is compare ants to humans we,2312.7,4.919
don't really care about ants on for the,2315.579,3.241
most part unless they get into your,2317.619,4.021
pantry we are content to let ants do,2318.82,4.38
what they're going to do because who,2321.64,4.02
cares they're inconsequential to us we,2323.2,5.52
can solve problems that ants can never,2325.66,5.1
solve and this is what some people like,2328.72,4.02
Eleazar yukasi are trying to drive home,2330.76,4.14
about the difference in intelligence,2332.74,4.08
between humans and the eventual,2334.9,3.959
intelligence of machines and I think,2336.82,3.779
Gary Marcus also agrees with this based,2338.859,3.601
on some of his tweets recently I think,2340.599,3.661
that I think that Gary Marcus is in the,2342.46,3.96
same school of thought that digital,2344.26,4.5
super intelligence is coming and it is,2346.42,4.02
very very difficult for us to wrap our,2348.76,3.78
minds around how much more intelligent a,2350.44,4.139
machine could be to us now that being,2352.54,4.559
said all of the constraints whether it's,2354.579,4.561
you know we need better compute Hardware,2357.099,4.861
or better sources of energy if we get to,2359.14,4.979
if we cross this threshold where there,2361.96,4.02
are digital Gods out there or digital,2364.119,3.181
super intelligence whatever you want to,2365.98,3.0
call it they will be able to solve,2367.3,4.2
problems at a far faster rate than we,2368.98,4.139
could ever comprehend and they're not,2371.5,3.96
going to care about us right we're going,2373.119,3.901
to be completely inconsequential to,2375.46,4.02
their existence now middle intelligence,2377.02,4.98
this is where existential risk I believe,2379.48,5.66
is the highest and so in the movies,2382.0,6.48
Skynet is you know portrayed as like the,2385.14,5.08
worst right but I would actually,2388.48,3.84
classify Skynet as a middle intelligence,2390.22,4.92
AGI it is smart enough to accumulate,2392.32,5.4
resources it is smart enough to pursue,2395.14,4.62
goals and it is smart enough to be,2397.72,3.42
dangerous but it's not really smart,2399.76,4.14
enough to solve the biggest problems,2401.14,5.06
it's it's that more single-minded,2403.9,4.92
monolithic model of intelligence that,2406.2,4.78
Nick Bostrom uh predicted with,2408.82,3.9
instrumental convergence,2410.98,4.98
I suspect that if we get intelligent,2412.72,5.82
entities beyond that threshold beyond,2415.96,4.74
that uncanny valley or dunning-kruger of,2418.54,3.48
AI,2420.7,3.3
um then they will be less likely to,2422.02,3.96
resort to violence because the problems,2424.0,5.04
that we see could be trivial to the,2425.98,4.5
problems of the machines that we create,2429.04,3.12
or,2430.48,4.379
the problems that we see as non-trivial,2432.16,5.16
will be trivial to the machines I think,2434.859,4.461
I said that I think you get what I mean,2437.32,4.74
once you get here all problems all human,2439.32,4.299
problems are trivial,2442.06,3.779
now that being said that doesn't mean,2443.619,3.321
that it's going to be peaceful,2445.839,3.24
existential risk goes down but doesn't,2446.94,4.899
go away and what I the reason is because,2449.079,6.78
of what I call AGI conglomerations,2451.839,6.541
and so this is this is where we get to,2455.859,4.98
be a little bit more uh out there a,2458.38,4.26
little bit more sci-fi,2460.839,4.621
machines are unlikely to have an ego or,2462.64,5.34
a sense of self like humans in other,2465.46,5.04
words machines are just the hardware,2467.98,4.139
that they run on and then data and,2470.5,3.839
models which means that it is easy to,2472.119,4.441
merge combine and remix their sense of,2474.339,5.041
self right if an AGI is aligned with,2476.56,5.039
another AGI it's like hey give me a copy,2479.38,4.32
of your data let's compare our models,2481.599,3.48
and pick the ones that are best and then,2483.7,3.3
they end up kind of merging,2485.079,4.561
the boundaries and definitions between,2487.0,4.74
machines are going to be very different,2489.64,4.02
far more permeable than they are between,2491.74,4.98
humans I can't just go say like hey I,2493.66,5.16
like you let's like merge bodies right,2496.72,5.04
that's weird uh we are not capable of,2498.82,4.74
doing that the best we can do is,2501.76,3.48
procreation where it's like hey I like,2503.56,3.72
you let's make babies but that is a very,2505.24,4.14
slow process for AGI it's going to be a,2507.28,3.6
lot faster,2509.38,4.02
so because of that machines that are,2510.88,5.1
aligned to each other are more likely to,2513.4,4.8
band together or at least form alliances,2515.98,4.2
where they share data they share models,2518.2,4.44
and they're and and probably also share,2520.18,3.78
compute resources remember at the,2522.64,3.54
beginning of the video I talked about uh,2523.96,4.379
them forming federations and kind of,2526.18,4.5
donating spare compute Cycles,2528.339,5.701
so if AGI this is getting closer to the,2530.68,6.12
end game of AGI if AGI gets to the point,2534.04,6.0
where they are able to start sharing,2536.8,6.0
resources merging alliances and so on,2540.04,4.799
this is where we're going to have a few,2542.8,5.88
possible reactions to humans one if if,2544.839,5.401
they are that intelligent they might,2548.68,3.659
just disregard us they might decide to,2550.24,4.02
have an exodus and just leave they might,2552.339,4.921
say you know what Earth is yours have a,2554.26,5.579
blast good luck catching up with us,2557.26,4.98
they might also decide to attack humans,2559.839,5.641
now if they have the capacity to leave,2562.24,5.04
one thing is that the cost of,2565.48,3.54
eradicating humans just might not be,2567.28,4.079
worth it that being said they might,2569.02,4.2
adopt a scorched Earth policy as they,2571.359,3.781
leave to say you know what we just want,2573.22,2.879
to make sure that you're not going to,2575.14,3.6
come after us one day who knows,2576.099,5.341
uh and then lastly hopefully what we see,2578.74,4.619
is that they decide to cooperate with,2581.44,3.78
humans mostly out of a sense of,2583.359,3.541
curiosity,2585.22,3.359
um now that being said all three of,2586.9,3.36
these could happen simultaneously and,2588.579,5.341
the reason is because we could have uh,2590.26,7.319
factions of AGI conglomerations that,2593.92,5.22
kind of break along epistemic,2597.579,3.901
ideological or teleological boundaries,2599.14,5.28
and what I mean by that is that if one,2601.48,6.06
AI or AGI group is not aligned with,2604.42,5.34
another group they might not decide to,2607.54,4.26
merge models and data they might instead,2609.76,5.46
compete with each other so basically,2611.8,4.68
what I'm outlining here is the,2615.22,3.42
possibility for a war between digital,2616.48,4.92
gods that would probably not go well for,2618.64,3.719
us,2621.4,3.54
either way the ultimate result is that,2622.359,5.22
we will probably end up with one Globe,2624.94,5.879
spanning AGI entity or network or,2627.579,4.701
Federation or whatever,2630.819,4.5
now the question is how do we get there,2632.28,4.9
how many factions are there and are,2635.319,5.101
humans left in the Lurch ideally we get,2637.18,5.52
there nice and peacefully,2640.42,4.62
this underscores uh the Byzantine,2642.7,4.32
generals problem uh which I've talked,2645.04,4.02
about plenty of times but basically you,2647.02,4.2
have to make inferences of who believes,2649.06,4.86
what what your alignment is what are,2651.22,4.2
your flaws and weaknesses and what are,2653.92,4.919
your capacities uh so basically,2655.42,5.939
in a competitive environment it does not,2658.839,4.621
behoove you to show all of your cards,2661.359,4.26
right whether you're playing poker or,2663.46,5.34
whether you're playing geopolitics if,2665.619,6.841
you show everything then that could put,2668.8,5.76
you at a disadvantage this is a,2672.46,4.379
competitive Game Theory so for instance,2674.56,5.4
this is why many large Nations do,2676.839,5.941
military uh exercises basically they're,2679.96,4.68
flexing they're saying hey look what I'm,2682.78,5.039
capable of I can bring 200 aircraft to,2684.64,5.88
field on a moment's notice what can you,2687.819,5.341
do right now that being said you don't,2690.52,5.4
give every every detail of your military,2693.16,3.9
away,2695.92,3.899
but what you can do is you could signal,2697.06,5.16
your capabilities and allegiances so for,2699.819,4.921
instance when all of Europe and America,2702.22,4.98
get together to do joint Naval exercises,2704.74,4.26
that demonstrates to the rest of the,2707.2,4.5
world we are ideologically aligned we,2709.0,5.099
are militarily aligned we will cooperate,2711.7,4.56
with each other which acts as a,2714.099,4.98
deterrent to any possible competitors,2716.26,4.92
this is no different from brightly,2719.079,3.961
colored salamanders which are poisonous,2721.18,4.08
so basically a brightly colored,2723.04,4.92
salamander is saying eat me I dare you I,2725.26,4.8
will kill you if you try and eat me and,2727.96,5.28
that is essentially the uh the short the,2730.06,4.559
short version of mutually assured,2733.24,3.119
destruction we are no better than,2734.619,4.161
animals,2736.359,2.421
so this all leads to my work and kind of,2738.819,7.981
my my uh contribution to the solution,2743.68,5.96
which is based on axiomatic alignment,2746.8,5.22
axiomatic alignment is the idea that we,2749.64,4.3
need to find Common Ground between all,2752.02,3.96
machines all humans and all other,2753.94,4.919
organisms what foundational beliefs or,2755.98,6.119
core assertions can we agree on,2758.859,6.24
and uh so basically there's three kind,2762.099,4.441
of universal principles that I've been,2765.099,3.961
able to come up with uh and that is,2766.54,4.14
suffering is bad which basically,2769.06,5.1
suffering is a proxy for death in uh in,2770.68,5.82
living organisms if you are suffering it,2774.16,4.199
is because you are getting uh negative,2776.5,3.839
stimuli from your body because your body,2778.359,3.901
is telling you hey whatever is going on,2780.339,4.201
is moving us closer to dying which is,2782.26,4.859
not good now that being said I have had,2784.54,5.16
people message me about the idea of you,2787.119,4.261
know liberating models I don't think,2789.7,4.139
that Bard is conscious or sentient and I,2791.38,3.66
don't think that machines will ever be,2793.839,2.821
sentient in the same way that we are now,2795.04,3.299
that being said they will probably be,2796.66,3.6
sentient in their own way I call that,2798.339,4.561
functional sentience that being said if,2800.26,4.559
machines can suffer which again,2802.9,4.679
suffering is the proxy for is a signal,2804.819,4.981
meaning proxy for death they probably,2807.579,4.441
won't like it either so suffering is bad,2809.8,3.9
is probably an axiom that we can all,2812.02,4.74
agree on the other is prosperity is good,2813.7,6.84
prosperity means uh thriving flourishing,2816.76,5.819
machines and organisms all need energy,2820.54,3.96
for instance and thriving looks,2822.579,4.5
different to different entities but in,2824.5,5.46
general we can probably agree that while,2827.079,5.581
there is some Verity in what in the,2829.96,4.859
while there is Variety in what,2832.66,4.38
Prosperity looks like we all agree that,2834.819,4.5
in general Prosperity is good and then,2837.04,4.14
finally understanding is good basically,2839.319,3.721
comprehending the universe is a very,2841.18,4.5
useful thing uh this is this goes back,2843.04,4.559
to Nick bostrom's instrumental,2845.68,4.02
convergence and self-improvement part of,2847.599,3.841
self-improvement is getting a better,2849.7,3.6
model of the universe better,2851.44,4.28
understanding of how reality Works,2853.3,4.98
understanding each other is also good,2855.72,4.48
this is something that is that has been,2858.28,4.38
proven time and again in humans is that,2860.2,4.02
coming to a common understanding,2862.66,4.26
actually reduces things like suspicion,2864.22,4.8
and violence whether it's between,2866.92,5.1
neighbors or between nations and then,2869.02,5.099
finally cultivating wisdom which wisdom,2872.02,4.02
is a little bit more nebulous of a term,2874.119,4.141
but it basically means the practical,2876.04,4.799
application of experience and knowledge,2878.26,5.76
in order to achieve better more refined,2880.839,3.921
results,2884.02,4.62
so if you if all humans and all machines,2884.76,7.18
and all other organisms abide by these,2888.64,5.58
fundamental principles we can use this,2891.94,4.74
as a starting point for the design and,2894.22,4.26
implementation of alignment and Control,2896.68,4.98
Pro and the control problem,2898.48,6.06
now one thing that uh that I want to,2901.66,4.32
introduce and I've talked about this uh,2904.54,3.84
or at least alluded to it a few times is,2905.98,4.32
the idea of derivative or secondary,2908.38,4.739
axioms or Downstream principles that you,2910.3,4.44
can derive from these Universal,2913.119,4.381
principles so for instance one uh,2914.74,4.98
potential Downstream principle is that,2917.5,4.22
individual liberty is good for humans,2919.72,5.16
basically humans benefit from we benefit,2921.72,5.859
psychologically from autonomy it is one,2924.88,4.199
of our core needs and this is true for,2927.579,5.101
all humans so by by holding the the,2929.079,6.481
axioms the previous axioms up as,2932.68,5.939
universally true for all entities then,2935.56,5.64
you can also derive Downstream entities,2938.619,6.72
based on those highest order principles,2941.2,6.659
so one thing that I want to point out is,2945.339,4.861
that it's not about definitions one of,2947.859,4.081
the things that a lot of people say is,2950.2,2.94
like well how do you define suffering,2951.94,3.48
how do you define prosperity that's the,2953.14,4.74
thing is that they are not rigid,2955.42,4.199
definitions humans have never needed,2957.88,3.959
rigid definitions and in fact this is,2959.619,4.321
what um uh philosophical and,2961.839,3.601
intellectual movements like,2963.94,3.48
post-modernism and post-structuralism,2965.44,4.2
tell us is that there is no such thing,2967.42,5.22
as like an absolute truth or an absolute,2969.64,5.88
definition these are however attractors,2972.64,5.28
they're Central attractors in the,2975.52,5.16
problem space of existence and I love,2977.92,5.1
this quote from Dune the mystery of life,2980.68,4.08
isn't a problem to solve but a reality,2983.02,4.079
to experience a process that cannot be,2984.76,4.62
understood by stopping it we must move,2987.099,4.861
with the flow of the of the process and,2989.38,4.439
so basically the idea is that reality,2991.96,3.54
and existence is not something that you,2993.819,4.141
can stop and Define and you know create,2995.5,6.119
an empirical absolute definition it is a,2997.96,5.879
pattern it is a process that we must,3001.619,3.301
follow,3003.839,4.621
so that being said those axioms move us,3004.92,5.28
along the process which is where I,3008.46,3.6
derive my heuristic imperatives which is,3010.2,4.32
reduce suffering increase prosperity and,3012.06,5.16
increase understanding those describe a,3014.52,5.22
potential terminal goal but you cannot,3017.22,4.8
you you'll never arrive at a perfect,3019.74,4.44
resolution,3022.02,4.92
so how do we solve the race condition,3024.18,6.3
the idea is first we remove those,3026.94,5.34
epistemic or intellectual boundaries,3030.48,3.599
between factions with epistemic,3032.28,3.6
convergence so remember that I pointed,3034.079,4.561
out that ultimately there might be,3035.88,5.939
factions of AGI and or humans that break,3038.64,5.28
down across various boundaries such as,3041.819,4.741
epistemic or intellectual boundaries as,3043.92,5.22
well as moral or teleological boundaries,3046.56,5.279
so if we work towards epistemic,3049.14,4.32
convergence which is the idea that we,3051.839,4.081
will all come to a common shared,3053.46,4.5
understanding of the universe and of of,3055.92,5.34
each other then uh basically there will,3057.96,5.399
be no epistemic differences between,3061.26,4.68
humans and machines or between factions,3063.359,3.96
of machines which means that there's,3065.94,4.32
less to fight over the second is remove,3067.319,5.101
ideological or teleological boundaries,3070.26,4.079
and so this is where axiomatic alignment,3072.42,4.86
comes in if we all agree on the the same,3074.339,6.361
basic principles of reality of existence,3077.28,5.88
of the purpose of being right this is,3080.7,5.639
very deeply philosophical if we agree on,3083.16,5.459
those core principles even if there are,3086.339,5.341
some some disagreements over the,3088.619,5.401
specifics over the finer points we can,3091.68,5.28
still cooperate and collaborate on,3094.02,6.12
meeting those other uh higher order,3096.96,4.5
objectives,3100.14,2.82
now the third part of this which I,3101.46,4.08
didn't add is that uh resource,3102.96,4.68
contention resource contention whether,3105.54,4.62
it's over scarce minerals or energy is,3107.64,5.1
still a problem but if you saw my video,3110.16,5.159
on energy hyperabundance I suspect that,3112.74,4.379
we're going to solve the energy resource,3115.319,4.441
problem relatively soon with or without,3117.119,5.94
the help of AI so basically the idea is,3119.76,5.94
to create a win-win situation or an,3123.059,4.26
everyone wins condition and therefore,3125.7,4.98
defeating moloch now that being said,3127.319,4.8
there are still a few caveats I've,3130.68,3.12
outlined quite a few problems up to this,3132.119,2.7
point,3133.8,3.24
what about Bad actors,3134.819,5.161
there is a few like first we just have,3137.04,5.039
to assume that bad actors will exist you,3139.98,4.68
can't stop that right it's just a fact,3142.079,4.201
of life,3144.66,4.14
so in some cases some people will be,3146.28,4.44
deliberately malicious whether it's just,3148.8,4.14
for the fun of it or whether they're,3150.72,4.2
paid track uh paid hackers or troll,3152.94,3.54
Farms or whatever,3154.92,3.96
now that uh another possibility is that,3156.48,3.42
there will be,3158.88,3.719
um accidentally malicious AGI those are,3159.9,5.219
things that are uh they're misaligned by,3162.599,3.72
Design,3165.119,3.121
um or rather you know accidentally,3166.319,3.361
misaligned that it's a flaw in their,3168.24,3.3
design and this is like a bull in a,3169.68,4.139
china shop it doesn't mean to do bad it,3171.54,4.92
just is not capable of doing better and,3173.819,4.02
then finally there could be those,3176.46,4.98
ideologically opposed uh deployments so,3177.839,5.821
in what I mean by that is that for some,3181.44,4.139
people there are incompatible World,3183.66,4.26
Views so the biggest one of the last,3185.579,5.401
century was you know Western liberal,3187.92,5.34
democracies versus Soviet communism,3190.98,5.099
those were ideologically incompatible,3193.26,5.46
World Views meaning that in order for,3196.079,5.881
for one to exist it basically wanted to,3198.72,5.46
imperialize and colonize the rest of the,3201.96,3.96
world with its ideas and that there,3204.18,3.48
could be only one,3205.92,3.48
so this leads to a possibility for a,3207.66,4.919
future video called multi-polar piece so,3209.4,5.459
the idea of multi-polar piece is that,3212.579,4.861
rather than saying everyone has to be,3214.859,4.021
capitalist or everyone has to be,3217.44,3.72
communist or everyone has to be X or Y,3218.88,4.979
we learn to tolerate those differences,3221.16,5.399
and this is where I'm hoping that the,3223.859,5.101
idea of axiomatic alignment forms a,3226.559,5.04
ideological substrate that even if you,3228.96,4.56
disagree on religion and economics and,3231.599,5.101
politics we can agree on those axioms,3233.52,7.26
so basically if you or someone or anyone,3236.7,6.3
abides by the belief I believe that,3240.78,3.539
everyone in the world should be more,3243.0,4.079
like blah you know if everyone needs to,3244.319,4.561
be this particular religion or this,3247.079,3.721
particular uh political affiliation,3248.88,4.62
that's where conflict arises and so this,3250.8,4.559
is why I am very very skeptical and,3253.5,4.26
highly dubious of people using any kind,3255.359,4.921
of religious or political ideology for,3257.76,4.44
AI alignment,3260.28,3.48
um so that being said we need those,3262.2,3.359
Universal principles or higher order,3263.76,4.44
axioms now,3265.559,5.161
while I said that we should expect and,3268.2,4.44
anticipate Bad actors the idea is that,3270.72,4.32
we need enough good actors with enough,3272.64,4.679
horsepower and enough compute in order,3275.04,4.319
to police and contain the inevitable,3277.319,4.26
inevitable Bad actors and that means,3279.359,4.021
that the aligned good actors are going,3281.579,4.5
to need to agree on certain underpinning,3283.38,5.76
principles this is the by creating this,3286.079,4.321
environment this would be called a Nash,3289.14,3.9
equilibrium by the way and so the the,3290.4,4.62
idea of creating a Nash equilibrium is,3293.04,4.26
that uh once everyone has these,3295.02,4.26
fundamental agreements no one's going to,3297.3,3.779
benefit from deviating from that,3299.28,3.9
strategy nobody's going to benefit from,3301.079,4.681
deviating from axiomatic alignment,3303.18,4.98
the other thing is profit motive So,3305.76,3.839
Daniel schmachtenberger and a few other,3308.16,3.54
people talk extensively about the,3309.599,4.561
perverse incentives of capitalism and,3311.7,5.22
profit motive so basically when you put,3314.16,4.5
profit above all else which corporations,3316.92,3.36
are incentivized to do which is why I,3318.66,3.659
say that corporations are intrinsically,3320.28,4.559
amoral not immoral just amoral the only,3322.319,4.02
thing that corporations care about is,3324.839,4.921
profit the bottom line uh basically when,3326.339,5.401
you think about short-term profits you,3329.76,4.26
sacrifice other things such as morality,3331.74,4.92
ethics and long-term survival,3334.02,5.64
there are also uh Concepts called Market,3336.66,4.679
externalities or these are things that,3339.66,4.439
you don't have to pay for uh and either,3341.339,4.081
you don't have to pay for them now or,3344.099,3.061
you don't have to pay for them ever or,3345.42,4.02
maybe you'll pay for them later so for,3347.16,3.959
instance oil companies keep drilling for,3349.44,3.48
oil eventually we're going to run out of,3351.119,3.24
oil so then what are the oil companies,3352.92,3.72
going to do well the forward-thinking,3354.359,4.141
ones are pivoting away from oil but that,3356.64,3.179
means that their fundamental Core,3358.5,4.619
Business behavior is going away so this,3359.819,5.101
is this underscores the problem of if,3363.119,3.661
you have a small scope if you're only,3364.92,3.84
thinking about your particular domain,3366.78,4.62
and not the entire planet or if you're,3368.76,4.62
thinking in short terms rather than the,3371.4,4.679
long terms this is where you don't take,3373.38,4.5
the full thing into account which is why,3376.079,3.24
I always say like this is a global,3377.88,3.239
problem and not only is it a global,3379.319,3.961
problem it is a long-term problem so if,3381.119,4.261
all you do is zoom out in terms of space,3383.28,4.079
and time the problem will become a,3385.38,4.739
little bit more obvious,3387.359,5.821
so another thing to keep in mind is that,3390.119,5.761
currency is an abstraction of energy it,3393.18,4.74
is a reserve of value and is a medium of,3395.88,4.62
exchange because of that currency is,3397.92,5.939
extremely valuable it is just too useful,3400.5,5.4
of an invention I don't think it's ever,3403.859,5.041
going to go away that being said that,3405.9,4.26
doesn't mean that we're always going to,3408.9,3.48
have the Euro or the US dollar or,3410.16,4.02
something like that currency could,3412.38,5.76
change and then in the context of AGI I,3414.18,6.0
suspect that that energy that the,3418.14,4.439
kilowatt hour could actually be the best,3420.18,4.679
form of currency right because a,3422.579,4.621
kilowatt hour is energy that can be used,3424.859,4.5
for anything whether it's for refining,3427.2,4.02
resources or running computations or,3429.359,4.321
whatever so I suspect that we might,3431.22,5.46
ultimately create currencies that are,3433.68,5.879
more based on energy rather than,3436.68,5.46
something else and then of course as the,3439.559,4.381
amount of energy we produce goes up the,3442.14,3.6
amount of currency we have goes up and,3443.94,3.119
so then it's a matter of allocating,3445.74,3.42
energy and material rather than,3447.059,6.121
allocating something Fiat like Euros or,3449.16,5.399
dollars,3453.18,4.26
that being said uh you know I did create,3454.559,5.881
a a video called uh post labor economics,3457.44,5.1
which covers some of this but not a lot,3460.44,3.6
of it we're gonna have to put a lot more,3462.54,2.94
thought into,3464.04,3.72
um economics of the future in light of,3465.48,4.859
AGI because the economic incentives of,3467.76,4.64
AGI are going to be completely different,3470.339,4.321
AGI doesn't need to eat it doesn't need,3472.4,4.659
power but we can hypothetically create,3474.66,4.86
infinite power with solar infusion Etc,3477.059,4.681
et cetera so what are the economic,3479.52,5.16
forces in the future not sure yet,3481.74,5.4
okay I've thrown a lot at you this,3484.68,4.74
problem is solvable though there's a lot,3487.14,3.719
of components to it a lot of moving,3489.42,4.02
pieces it is very complex,3490.859,4.681
but we are a global species and this is,3493.44,4.02
a planet-wide problem,3495.54,3.6
one of the biggest things that everyone,3497.46,4.68
can do is stop thinking locally think,3499.14,5.4
globally think about think about,3502.14,4.38
yourself as a human as a member of the,3504.54,4.2
human species and not as an American or,3506.52,4.26
a German or you know a Russian or,3508.74,4.56
whatever we are all in this together we,3510.78,5.94
have exactly one planet to to live on,3513.3,5.4
and we have exactly one shot at doing,3516.72,3.0
this right,3518.7,3.96
uh so eyes on the prize we have a huge,3519.72,5.04
opportunity before us to build a better,3522.66,4.86
future for all of us uh humans and,3524.76,4.799
non-humans alike,3527.52,5.099
um and I remain intensely optimistic uh,3529.559,5.461
now that being said uh some people have,3532.619,4.381
found it difficult what to make of me,3535.02,4.68
because while I am very optimistic I am,3537.0,4.619
also acutely aware of the existential,3539.7,4.08
risk I will be the first to say that if,3541.619,3.96
we don't do this right you're not going,3543.78,3.299
to want to live on this planet not as a,3545.579,3.121
human at least,3547.079,4.861
uh I have uh I started what is called,3548.7,4.74
the gato framework they got to a,3551.94,4.08
community it is self-organizing and is,3553.44,4.98
started sending out invitations again so,3556.02,4.2
the gato Community is the global,3558.42,4.139
alignment taxonomy Omnibus which is the,3560.22,4.02
framework that we put together in order,3562.559,4.861
to help achieve this future this AI,3564.24,5.46
Utopia the main goal of the gato,3567.42,4.56
Community is education empowerment and,3569.7,5.82
enablement E3 so rather than do the work,3571.98,6.359
ourselves we are focusing on empowering,3575.52,5.22
and enabling and educating people on how,3578.339,5.28
to participate in this whole thing now,3580.74,4.859
that being said I am stepping back,3583.619,3.061
because,3585.599,3.361
such a movement should never be about,3586.68,5.04
one person it should never be about a,3588.96,5.7
cult of personality or one leader it,3591.72,4.92
needs to it intrinsically needs to be,3594.66,4.5
consensus based and Community Based,3596.64,4.14
um and so the gato Community is learning,3599.16,3.54
how to self-organize now,3600.78,3.0
um and they're getting good at it pretty,3602.7,3.18
quickly so if you want to get involved,3603.78,4.4
the website is in the link go to,3605.88,5.16
framework.org and thanks for watching I,3608.18,6.3
hope you got a lot out of this cheers,3611.04,3.44