davidshapiro_youtube_transcripts / AGI Unleashed Game Theory Byzantine Generals and the Heuristic Imperatives_transcript.csv
Stevross's picture
Upload 50 files
421fea8
raw
history blame contribute delete
No virus
40.3 kB
text,start,duration
hey everyone David Shapiro here with,1.319,5.161
today's video good morning,4.2,4.439
today's topic is going to be a little,6.48,4.199
bit more severe and a little bit more,8.639,4.981
intense so the title of today's video is,10.679,6.721
Agi Unleashed what happens when we lose,13.62,5.999
control,17.4,5.58
part one the coming storm,19.619,5.521
as everyone is aware things are ramping,22.98,4.379
up very quickly there are people calling,25.14,5.639
for moratoriums on AI research and while,27.359,5.641
some of us don't take it seriously there,30.779,4.321
are very legitimate concerns about,33.0,3.6
what's going on,35.1,4.139
and in other words we're in the end game,36.6,5.22
now we are in the ramp up towards AGI,39.239,5.521
and the singularity whether or not you,41.82,5.579
are emotionally ready for it,44.76,6.299
so just as a tiny bit of evidence this,47.399,5.961
paper from nature,51.059,4.441
demonstrates that we are in an,53.36,4.96
exponential ramp up on AI whatever else,55.5,5.34
is true the investment is there the,58.32,4.919
research is there it's happening and,60.84,5.22
it's not slowing down,63.239,5.401
now I'm probably preaching to the choir,66.06,4.44
and rehashing old,68.64,4.56
stuff that everyone knows but let's just,70.5,4.799
briefly talk about the existential risks,73.2,5.459
there are two overarching themes or,75.299,5.46
categories of the existential risks,78.659,7.441
risks for AGI one is the basically just,80.759,8.301
the deliberate weaponization of AI,86.1,6.839
some people are comparing AGI to nuclear,89.06,6.4
weapons beyond that there's the,92.939,5.581
potential of cyber warfare drone Warfare,95.46,8.24
autonomous uh tanks artillery aircraft,98.52,8.279
submarines so on and so forth many of,103.7,4.959
these systems are already being,106.799,3.86
developed and deployed,108.659,5.701
so basically ai ai is already being,110.659,5.621
weaponized it's just a matter of on what,114.36,2.88
scale,116.28,3.9
moreover the second category is,117.24,5.22
accidental outcomes or accident or,120.18,4.979
unintended consequences basically what,122.46,5.7
if the AGI escapes but there's a few,125.159,5.341
other possible avenues for this to,128.16,4.68
happen for instance runaway corporate,130.5,3.36
greed,132.84,2.94
you bet your bottom dollar that the,133.86,4.2
first corporation that can create AGI to,135.78,4.86
automate profits is going to do so,138.06,5.759
uh more uh beyond that there's political,140.64,4.8
corruption and just straight up,143.819,3.841
political incompetence this is something,145.44,3.84
that has been uh discussed actually,147.66,4.2
quite a bit in uh my comments section,149.28,4.76
for my most recent videos which is that,151.86,8.04
uh yes uh like Italy uh uh uh Britain,154.04,8.26
and a few other places are doing their,159.9,4.32
best to kind of get ahead of the curve,162.3,4.62
but the fact of the matter is is that,164.22,4.799
governments by and large are moving too,166.92,4.74
slow and politicians just don't get it,169.019,5.281
and then finally you can even have,171.66,6.0
situations where otherwise benign agis,174.3,6.18
collaborate and eventually turn on us,177.66,5.52
so the existential risks are there,180.48,5.46
and you guys most most of you know me I,183.18,4.62
am very optimistic I'm very sanguine,185.94,3.84
about all this and by the end of the,187.8,4.439
video I hope that many of you will see,189.78,4.739
where I'm coming from,192.239,5.881
now even if AGI doesn't wipe us out,194.519,7.14
there are still lots and lots of risks,198.12,5.759
um job loss economic changes social,201.659,5.16
upheaval so on and so forth mostly it,203.879,6.601
comes down to economic shifts who gets,206.819,5.761
to capture all of the tremendous wealth,210.48,5.58
generated by AGI and even if AGI is safe,212.58,5.519
and doesn't kill us there's still every,216.06,3.78
possibility that the corporatists and,218.099,3.841
capitalists will have us living in a,219.84,4.56
dystopian hell before too long,221.94,5.46
so that is also an intrinsic risk even,224.4,5.759
if it is not an existential risk so we,227.4,4.32
really need to be talking about what,230.159,3.72
happens when we get these super powerful,231.72,4.76
agis,233.879,2.601
part two autonomous AI,236.84,7.06
AGI is kind of a loaded term there's a,240.12,5.1
lot of,243.9,4.38
debate over what AGI even means so this,245.22,4.92
is why I have started saying autonomous,248.28,4.319
AI I don't care how smart it is,250.14,5.459
the the point being is that AI becoming,252.599,5.04
autonomous is really what we're talking,255.599,5.941
about and as we talk more about what is,257.639,6.601
the path to AGI autonomy is rapidly,261.54,6.18
becoming part of the conversation now,264.24,5.82
that being said there is a huge,267.72,4.68
disconnect between what,270.06,5.4
us proletariat are talking about and,272.4,6.54
what is being released from the halls of,275.46,5.64
power and what the academic,278.94,4.14
establishment is talking about,281.1,4.68
there is a lot of gatekeeping and a lot,283.08,5.52
of this is cloistered and that I think,285.78,4.74
is probably one of the biggest problems,288.6,3.9
which is honestly why I'm making these,290.52,4.22
videos,292.5,2.24
autonomous AI is coming whether or not,294.96,5.22
you like it whether or not you're,297.96,4.5
emotionally ready for it autonomous AI,300.18,4.56
is easier to make than you think people,302.46,3.959
are starting to realize this especially,304.74,4.38
with the release of chat GPT plugins and,306.419,4.381
the ability for language models to use,309.12,3.859
apis,310.8,5.88
to be fair the Department of Defense and,312.979,5.741
universities and tech companies are all,316.68,5.04
working on autonomous AI systems they,318.72,4.94
are kind of cloistering and and and,321.72,4.319
closing their research it's not as open,323.66,6.16
as it once was and in part I I totally,326.039,6.181
understand that because I think that,329.82,4.56
some of these folks realize how close we,332.22,3.0
are,334.38,3.659
and that scares them and so they're,335.22,4.62
they're basically playing the CL playing,338.039,4.741
the cards very close to their vest until,339.84,5.699
they get a better read on the situation,342.78,5.46
now taken one step further,345.539,5.641
or or looking at at this problem no one,348.24,5.04
has actually proposed a comprehensive,351.18,4.44
framework uh there's been plenty of,353.28,4.28
books on it there's been a few papers,355.62,5.82
but there's not really anything that is,357.56,7.18
a fully realized solution and I don't in,361.44,5.28
my opinion nobody has even fully,364.74,4.98
articulated the danger yet what's going,366.72,5.759
on and I'm hoping that this video will,369.72,4.44
advance that conversation just a little,372.479,3.481
bit,374.16,3.78
part of what happens in this comment,375.96,3.9
comes up on my videos a lot is a lot of,377.94,3.479
people are afraid to even talk about,379.86,4.02
this out of fear of ridicule,381.419,4.021
those of us that are paying attention,383.88,3.06
those of us that are doing the work,385.44,3.84
those of us that are working on it we,386.94,4.979
all see what's possible but Society by,389.28,4.56
and large is not ready for it they're,391.919,4.381
not ready for the conversation and so,393.84,4.32
there's a lot of ridicule whether it's,396.3,4.26
in Private Industry whether it's in the,398.16,4.68
halls of Academia government military,400.56,3.68
and so on,402.84,5.34
uh so this basically comes down to,404.24,5.32
something that I call institutional,408.18,3.359
codependence which is that the,409.56,3.24
establishment believes that the,411.539,3.061
establishment is always right and the,412.8,4.519
establishment will use any tactic attack,414.6,6.06
tactic or technique sorry uh to control,417.319,6.16
the conversation such as shame bullying,420.66,7.14
and otherwise silencing techniques uh to,423.479,6.72
stymie the conversation again this is,427.8,5.78
why I've created my YouTube channel,430.199,3.381
instead who gets to dominate the,433.88,5.219
conversation fiction and billionaires,436.68,5.639
Skynet Elon Musk The Establishment has,439.099,5.38
abdicated responsibility for this now,442.319,4.741
that being said I am absolutely 100,444.479,6.72
certain that the halls of power and and,447.06,6.24
the establishment is doing the research,451.199,4.801
behind closed doors,453.3,5.58
but they have abdicated responsibility,456.0,5.099
for controlling the narrative and,458.88,5.36
guiding the conversation,461.099,3.141
so let's talk about a couple aspects of,464.639,4.68
the control problem so for those not in,467.039,4.56
the know the control problem is the,469.319,5.28
category of problems of basically once,471.599,5.04
the AI becomes super intelligent how do,474.599,3.301
you control it,476.639,3.481
there's a few Concepts in here I'm not,477.9,3.72
going to go into every single one of,480.12,3.72
them but a few of them that you may or,481.62,4.62
may not know about one is convergent,483.84,4.68
instrumental values the idea is that any,486.24,4.92
sufficiently intelligent agent will come,488.52,4.679
up with certain instrumental values or,491.16,5.039
goals in service to those other goals,493.199,4.081
whatever the whether whatever whatever,496.199,3.541
its primary purpose happens to be such,497.28,4.02
as self-preservation resource,499.74,3.899
acquisition and so on,501.3,4.679
so basically in order to further any,503.639,5.28
goal whatever it happens to be your AI,505.979,5.101
might come to the conclusion that it,508.919,3.901
needs to preserve itself in order to,511.08,3.959
continue furthering its goal that's a,512.82,4.68
pretty reasonable uh thought to have I'm,515.039,4.92
not going to comment one way or another,517.5,4.38
um on on whether or not I think that's,519.959,3.781
going to happen ultimately I don't think,521.88,3.6
it's relevant and you'll see why a,523.74,3.539
little bit later second is the,525.48,3.84
orthogonality thesis which basically it,527.279,3.781
says very simply there is no correlation,529.32,4.199
between an ai's level of intelligence,531.06,4.8
and the values that it pursues,533.519,4.561
the treacherous turn this is a,535.86,4.8
hypothetic hypothetical situation in,538.08,7.199
which a apparently benign AGI suddenly,540.66,7.44
or apparently turns on its creators,545.279,5.101
because it came to some conclusion that,548.1,4.32
we don't understand,550.38,4.139
the courage ability problem which,552.42,5.039
basically says that uh your AI might,554.519,5.82
might not remain open to correction or,557.459,4.981
feedback or it might resist being shut,560.339,2.961
down,562.44,3.6
so basically you lose control of it just,563.3,5.92
because uh it says I'm not I'm not down,566.04,4.5
with that anymore,569.22,3.54
and then finally the value loading,570.54,4.32
problem which is how do you specify,572.76,4.68
human wave values in such a way that the,574.86,5.22
AI can understand and act on them and,577.44,4.74
then the very next follow-up question is,580.08,4.68
who gets to Define them anyways,582.18,4.14
so again these are a few of the,584.76,4.079
assertions and these and hypotheses this,586.32,4.56
is not an exhaustive list but you kind,588.839,4.68
of get an idea there are a lot of ideas,590.88,4.5
out there and not a whole lot of,593.519,3.241
solutions,595.38,4.56
now speaking of there are some solutions,596.76,5.4
out there and these are more like broad,599.94,3.72
categories,602.16,3.0
um rather than,603.66,3.84
um rather than comprehensive Frameworks,605.16,6.119
so one thing is kill switch Solutions,607.5,5.76
um pretty self-explanatory this is a,611.279,4.141
broad category of ideas,613.26,4.139
um I just saw on the internet someone,615.42,5.94
proposed that we we put uh bombs like,617.399,5.821
remotely triggered bombs in every data,621.36,4.38
center so that we can immediately shut,623.22,5.1
down every data center if we need to,625.74,5.539
okay sure that,628.32,5.76
doesn't sound like a very reasonable uh,631.279,5.62
direction to go for me but hey worst,634.08,5.699
comes worst case scenario maybe we do,636.899,5.641
uh courage ability which basically is,639.779,4.381
just the idea of just make the machine,642.54,4.799
responsive to feedback uh but what if,644.16,4.98
the feedback mechanism one doesn't work,647.339,4.081
or two the machine shuts it down or,649.14,5.04
three the feedback mechanism,651.42,6.359
um doesn't have the intended uh efficacy,654.18,5.279
that we want it to have,657.779,3.421
uh and then there's various kinds of,659.459,3.961
reinforcement learning inverse,661.2,4.44
reinforcement learning and passive and,663.42,5.4
blah blah basically just include an,665.64,6.6
algorithm so that the machine will,668.82,5.759
automatically autonomously learn the,672.24,4.68
values that we want one this is,674.579,4.141
difficult to do in the first place and,676.92,3.479
two what if the machine rewrites itself,678.72,4.7
or accident or even accidentally,680.399,5.341
nullifies uh those reinforcement,683.42,4.06
learning signals,685.74,4.08
finally values alignment what if you,687.48,4.56
build it with the human friendly values,689.82,4.139
in the first place,692.04,2.64
um,693.959,2.761
still you run into the same problem one,694.68,4.14
how do you implement it two what if it,696.72,4.08
changes its mind later and three what if,698.82,4.32
the values that you gave it are poorly,700.8,4.56
defined or intrinsically flawed or,703.14,5.879
broken now that that being said there is,705.36,5.659
a paper that literally just came out,709.019,4.201
called the capacity for moral,711.019,3.88
self-correction in large language models,713.22,3.66
so I'm really glad to see that the,714.899,3.901
establishment is,716.88,4.32
at the very beginning of talking about,718.8,4.38
this uh soberly,721.2,4.379
so the link is here but you can just,723.18,3.779
search that,725.579,3.421
um I I believe this paper was published,726.959,3.901
at least in part by the folks over at,729.0,4.68
anthropic still these are not complete,730.86,5.099
Solutions and we are literally months,733.68,4.74
away from from fully autonomous AGI,735.959,5.281
systems the conversation is not going,738.42,4.56
fast enough,741.24,4.2
so if you haven't heard found it or seen,742.98,4.26
it just do a Google search for bad,745.44,3.959
alignment take Bingo,747.24,4.38
um these are these were circulated on,749.399,4.581
Twitter probably more than a year ago,751.62,5.279
these will address and and argue against,753.98,4.72
many of the things that people say out,756.899,3.421
there so I'm not gonna I'm not gonna,758.7,3.6
rehash it or read all of them to you but,760.32,4.199
I just wanted to show you that like some,762.3,3.719
people are treating it like a joke it's,764.519,3.12
not really a joke but this is a good way,766.019,3.241
of just showing like yeah the thing that,767.639,4.861
you thought has already been addressed,769.26,7.079
part three the AGI landscape,772.5,5.7
um the biggest takeaway here is that,776.339,3.841
there's not going to be one AGI there's,778.2,4.62
going to be a few features of how AGI is,780.18,5.219
implemented so let's talk about that,782.82,5.519
first and foremost intelligence is not,785.399,4.861
binary it's not like you're going to,788.339,3.901
flip the switch in one day it's AGI but,790.26,3.92
the day before it wasn't,792.24,5.52
basically intelligence uh and and the,794.18,5.92
sophistication of AGI systems will,797.76,4.68
evolve over time there are going to be,800.1,4.88
various constraints such as time energy,802.44,5.76
data and all of that basically means,804.98,5.74
that the the level of power of your AGI,808.2,4.74
system is going to be on a sliding scale,810.72,5.04
so for instance even the most evil,812.94,5.28
machines might not be that powerful and,815.76,4.019
they're going to be constrained based on,818.22,3.179
you know the processing power of the,819.779,3.24
computers that they're running on the,821.399,3.841
network speed that they have so on and,823.019,4.5
so forth when you look at intelligence,825.24,3.779
there are literally thousands of,827.519,4.861
dimensions of intelligence that that the,829.019,5.161
types of intelligence out there are huge,832.38,4.32
and so AGI is not going to master all of,834.18,5.099
them immediately it's going to take time,836.7,4.62
and then as I just mentioned there are,839.279,4.201
gonna there are going to be numerous,841.32,4.38
limiting factors or constraints such as,843.48,3.84
the underlying Hardware the training,845.7,4.98
data and energy requirements of course,847.32,6.8
that is going to change quickly as,850.68,5.88
basically the underlying Hardware ramps,854.12,4.48
up exponentially the amount of data that,856.56,4.92
is available ramps up exponentially and,858.6,6.72
then the underlying machine learning,861.48,5.94
models the neural networks also get,865.32,3.959
exponentially more sophisticated and,867.42,3.539
larger,869.279,3.961
now as I mentioned most importantly,870.959,5.041
there won't just be one Skynet the,873.24,3.96
reason that we think that there's going,876.0,2.88
to be just one is because it is,877.2,3.379
convenient from a narrative perspective,878.88,4.44
in Terminator it's easy to just say,880.579,5.981
there's one big bad there's one Skynet,883.32,5.16
um but that's not how it's going to,886.56,4.38
happen there's going to be hundreds,888.48,4.44
thousands millions it's going to ramp up,890.94,3.24
very quickly,892.92,5.52
so what this results in is a sort of,894.18,6.899
arms race amongst in between the agis,898.44,4.98
themselves as well as the sponsors or,901.079,3.721
the people trying to build and control,903.42,4.02
them which results in a survival of the,904.8,5.399
fittest situation or a race condition,907.44,5.459
where basically the most aggressive and,910.199,5.221
sophisticated and and Powerful agis are,912.899,4.161
the ones who win,915.42,4.32
which that could be bad because then,917.06,4.899
you're basically selecting for the most,919.74,5.279
aggressive and hostile agis,921.959,5.82
the high velocity of AGI cyber warfare,925.019,6.18
will probably require our AGI systems to,927.779,6.481
be partially or fully autonomous,931.199,5.461
so basically what that means is that in,934.26,5.4
order to match the arms race in cyber,936.66,4.26
warfare,939.66,3.6
the agis that we built will probably,940.92,5.52
need to be evolving which means that,943.26,4.68
they'll spawn off copies of themselves,946.44,4.44
they'll be polymorphic they will recode,947.94,6.24
themselves so on and so forth and also,950.88,5.94
when you look at AGI in the context of,954.18,4.92
cyber warfare they will explicitly,956.82,4.759
require adversarial objective functions,959.1,4.979
this is what was explored in Skynet,961.579,4.361
which basically the objective function,964.079,4.141
of Skynet was probably like maximize,965.94,5.22
military power or so on,968.22,5.46
So In This Global AGR arms race there's,971.16,4.32
going to be numerous copies they're all,973.68,3.48
going to be changing which results in,975.48,3.479
the Byzantine generals problem so the,977.16,3.96
Byzantine generals problem is a cyber,978.959,4.261
security thought experiment where,981.12,5.159
wherein the idea is you have numerous,983.22,4.979
generals and you don't know their,986.279,3.36
allegiance you don't know their loyalty,988.199,3.601
and you don't know their plans either so,989.639,3.721
how do those how do those generals,991.8,3.96
communicate with each other in such a,993.36,4.919
way that they can understand who's on,995.76,4.499
whose side and also come to consensus on,998.279,4.261
what the plan is assuming that there are,1000.259,5.101
hostile or adversarial actors,1002.54,5.76
now thinking of this in terms of three,1005.36,5.46
to five entities is difficult enough but,1008.3,3.779
we're going to be talking about a,1010.82,3.36
situation where there are millions or,1012.079,5.281
billions of agis all of them with,1014.18,5.579
unknown objective functions,1017.36,5.399
as autonomous agents lastly they will,1019.759,4.92
form alliances with each other,1022.759,4.32
by some means or other they will,1024.679,4.081
communicate they will establish their,1027.079,4.201
intentions and allegiances,1028.76,4.679
um and they will spend more time talking,1031.28,4.08
with each other than they willed with us,1033.439,4.561
this was something that is um that,1035.36,4.38
people are starting to talk about some,1038.0,4.079
of the folks that I I'm working with on,1039.74,4.62
cognitive architecture we're realizing,1042.079,5.1
that the very instant that you create a,1044.36,5.64
cognitive architecture it can talk 24 7.,1047.179,5.581
we can't talk 24 7 so even just by,1050.0,4.62
virtue of experimenting with cognitive,1052.76,3.6
architectures it makes sense to have,1054.62,4.16
them talking with each other,1056.36,6.179
uh and having agis talk with each other,1058.78,5.139
and come to agreements and,1062.539,3.0
understandings,1063.919,3.841
um this is going to happen even with the,1065.539,4.621
most benign benevolent outcomes of AGI,1067.76,5.64
now what these what these autonomous AI,1070.16,5.879
systems agree and disagree on will,1073.4,5.279
likely determine the overall outcome of,1076.039,4.981
what happens with the singularity with,1078.679,4.62
AGI and with,1081.02,5.64
um the basically the fate of humanity,1083.299,7.141
part four AGI Unleashed now given,1086.66,5.22
everything that I've outlined the,1090.44,3.06
question remains how do you control the,1091.88,2.46
machine,1093.5,5.36
my answer is maybe you don't,1094.34,4.52
the reason that I believe this is,1099.2,2.88
because the genie is out of the bottle,1100.82,4.08
open source models are proliferating you,1102.08,4.92
can already run a 30 billion parameter,1104.9,4.44
model on a laptop with six gigabytes of,1107.0,4.44
memory that paper just came out what,1109.34,3.78
yesterday or today,1111.44,3.84
Global deployments of AI are rising,1113.12,4.62
Federal and Military investment globally,1115.28,4.139
is also Rising,1117.74,3.9
because of this centralized alignment,1119.419,4.681
research is completely irrelevant it,1121.64,5.159
doesn't matter how responsible the most,1124.1,5.04
responsible actors are there are hostile,1126.799,3.721
actors out there with malevolent,1129.14,3.539
intentions and they have lots of funding,1130.52,4.26
not only that the AI systems are,1132.679,5.101
becoming much more accessible,1134.78,5.7
because of that distributed cooperation,1137.78,5.399
is now required alignment is not just,1140.48,4.86
about creating an individual Ai and if,1143.179,4.081
you go look at the alignment the bad,1145.34,4.62
alignment take Bingo none of those talk,1147.26,5.34
about distribution collaboration or,1149.96,4.8
collective intelligence or Collective,1152.6,4.62
processing all of the all of the,1154.76,4.26
conversations today are still talking,1157.22,4.56
about individual agis as if they're,1159.02,5.1
going to exist in a vacuum so far as I,1161.78,5.1
know no one is talking about this in the,1164.12,6.059
context of Game Theory and competition,1166.88,5.64
so because of this we need an alignment,1170.179,4.021
scheme that can create open source,1172.52,3.06
collaboration amongst numerous,1174.2,4.5
autonomous AGI entities such a framework,1175.58,5.459
needs to be simple robust and easy to,1178.7,3.719
implement,1181.039,4.821
we'll get to that in just a minute,1182.419,3.441
so,1186.02,3.72
what I'm basically proposing is a,1187.82,3.9
collective control scheme which might,1189.74,3.54
sound impossible,1191.72,3.9
creating one benevolent stable super,1193.28,4.2
intelligence is hard enough and now I'm,1195.62,3.24
saying we need to create millions of,1197.48,3.96
them billions of them,1198.86,4.8
what I'm saying is not that we need to,1201.44,3.84
we might not have a choice in the matter,1203.66,4.68
this might be the only path forward,1205.28,4.92
now if you're familiar with the work of,1208.34,4.92
John Nash and Game Theory you might be,1210.2,5.099
able to think about this in terms of,1213.26,4.14
okay let's just imagine for a minute,1215.299,3.901
that there are millions of agis out,1217.4,2.94
there,1219.2,3.08
with many of them with unknown,1220.34,7.64
intentions given a game theory,1222.28,8.019
dilemmas like the prisoner's dilemma and,1227.98,4.72
so on if you think about this in that,1230.299,4.981
perspective it may be possible to devise,1232.7,4.979
rules or assumptions that enable the AI,1235.28,4.98
the agis to reach consensus on their,1237.679,5.641
behavior even with the presence of,1240.26,4.98
malicious and faulty actors,1243.32,5.28
so what kinds of rules or assumptions,1245.24,6.0
could we give our AGI systems that we're,1248.6,5.22
all going to be developing independently,1251.24,6.059
excuse me so that they arrive at this,1253.82,6.18
equilibrium this Nash equilibrium that,1257.299,4.861
we're looking for how do we ensure that,1260.0,4.02
this that the millions and billions of,1262.16,4.2
agis that are coming arrive at the,1264.02,5.399
consensus we want them to,1266.36,4.319
part five,1269.419,3.481
heuristic imperatives,1270.679,4.38
so now we're going to talk about the,1272.9,4.38
work that I have done on this problem,1275.059,5.221
and this is not just hypothetical these,1277.28,4.8
there are also experiments that I've,1280.28,3.66
done that are documented and I'll link,1282.08,4.2
to those as well,1283.94,4.08
so the heuristic imperatives that I have,1286.28,4.139
come up with are quite simply one reduce,1288.02,4.62
suffering in the universe two increase,1290.419,4.14
prosperity in the universe and three,1292.64,4.019
increase understanding in the universe,1294.559,3.48
and I've been I've been talking about,1296.659,4.861
these uh much more frequently lately,1298.039,5.461
so let's take a deeper dive into these,1301.52,3.18
imperatives,1303.5,3.48
so first what is a heuristic imperative,1304.7,4.26
it's a set of principles that can be,1306.98,4.319
embedded into autonomous AI that,1308.96,4.079
basically takes the place of intrinsic,1311.299,4.021
motivations now what I want to point out,1313.039,4.921
is that the gpt4 paper that Microsoft,1315.32,5.16
published did mention intrinsic,1317.96,4.86
motivation so again The Establishment is,1320.48,3.66
starting to come around and I'm sure,1322.82,2.58
they've had more conversations,1324.14,3.6
internally that they are not revealing,1325.4,4.86
yet but they are setting the stage to,1327.74,4.919
talk about what intrinsic motivations do,1330.26,3.539
we give them,1332.659,2.88
so in the case of the heuristic,1333.799,4.62
imperatives these are imperatives that,1335.539,6.061
are uh basically provided a moral and,1338.419,4.861
ethical framework as well as those,1341.6,3.78
intrinsic motivations because very early,1343.28,3.899
on in my research I realized that there,1345.38,4.679
is no difference between an intrinsic,1347.179,4.681
motivation and a moral and ethical,1350.059,4.081
framework you have to have some impetus,1351.86,4.799
some motivation behind and reasoning,1354.14,6.36
behind all behavior and all reasoning,1356.659,5.941
so why these three why suffering and,1360.5,4.98
prosperity and understanding first it's,1362.6,5.76
a holistic approach it uh it's a it's a,1365.48,5.28
flexible framework that provides a very,1368.36,5.34
Broad and yet simple to implement,1370.76,5.58
framework it also balances trade-offs,1373.7,4.5
remember these heuristic imperatives,1376.34,4.38
have to be implemented simultaneously,1378.2,6.479
and in lockstep so this forces the AI to,1380.72,5.819
reason through and balance trade-offs,1384.679,3.36
between,1386.539,3.841
um between these objectives,1388.039,4.681
they're also very adaptable and context,1390.38,5.039
sensitive and basically what I mean by,1392.72,4.68
that is that large language models today,1395.419,5.821
like gpt4 are very very aware of the,1397.4,6.18
fact that these that these general,1401.24,4.08
principles these heuristic imperatives,1403.58,4.8
are not the be-all end-all but they are,1405.32,5.82
guidelines they're they're uh they're,1408.38,4.14
shorthand,1411.14,3.539
um ways of basically implementing,1412.52,4.32
intuition in order to quickly make,1414.679,5.041
decisions uh that adhere to a general,1416.84,5.52
principle or a moral compass and then,1419.72,5.28
evaluate that uh based against the,1422.36,4.1
context that it's in,1425.0,4.14
there's two other things that emerged,1426.46,4.719
during my most recent experiments with,1429.14,4.14
the heuristic imperatives and that is,1431.179,3.961
that the heuristic imperatives promote,1433.28,5.04
individual autonomy uh basically chat,1435.14,6.72
gpt4 realized that in order to reduce,1438.32,5.64
suffering of people you need to protect,1441.86,5.04
individual autonomy ditto for Prosperity,1443.96,4.92
that if you control people they're not,1446.9,3.06
going to be happy and they're not going,1448.88,3.299
to be prosperous so that was an emergent,1449.96,4.44
quality of the heuristic imperatives,1452.179,4.74
that surprised me and made me realize,1454.4,7.08
that chat gpd4 is already capable of a,1456.919,7.981
very very highly nuanced reasoning the,1461.48,5.16
other emerging quality that I did,1464.9,4.62
anticipate was fostering Trust,1466.64,6.06
basically when you have an AI equipped,1469.52,5.1
with these heuristic imperatives it,1472.7,4.92
understands that um fermenting trust or,1474.62,5.46
fostering trust with people is actually,1477.62,4.86
critical as a subsidiary goal or an,1480.08,4.74
auxiliary goal of these because if if,1482.48,4.679
humans don't trust the AI the rest of,1484.82,5.7
its imperatives are made irrelevant,1487.159,6.301
finally there are a lot of what about,1490.52,4.92
isms yeah but what about there's a lot,1493.46,3.54
of protests which of course this is part,1495.44,3.0
of the conversation,1497.0,3.84
so the most con these are some of the,1498.44,5.04
most common protests that I get when I,1500.84,4.26
talk about the heuristic imperatives one,1503.48,4.079
is won't reduce suffering result in the,1505.1,4.319
extermination of all life the short,1507.559,4.221
answer is yes if you only have that one,1509.419,5.221
which is why I spent two years working,1511.78,5.56
on the other two heuristic imperatives,1514.64,5.1
to counterbalance them because I realize,1517.34,4.86
that any single objective function is,1519.74,3.84
always going to be intrinsically,1522.2,4.74
unstable you must have a system that,1523.58,6.42
balances multiple sometimes antagonistic,1526.94,5.58
functions against each other in order to,1530.0,5.4
stabilize and reach that equilibrium,1532.52,5.279
number two yeah but who gets to Define,1535.4,4.139
suffering prosperity and understanding,1537.799,4.26
the short answer is nobody that is the,1539.539,4.201
point of of implementing it as a,1542.059,4.561
heuristic the machine learns as it goes,1543.74,6.299
and anyways llms like gpt4 already have,1546.62,5.1
a far more nuanced understanding,1550.039,4.321
understanding of the concept of,1551.72,4.92
suffering prosperity and understanding,1554.36,5.04
um than any individual human does and,1556.64,5.22
also humans have never needed perfect,1559.4,4.32
definitions we learn as we go as well,1561.86,4.08
and we get by,1563.72,4.74
number three well what about uh cultural,1565.94,5.16
biases and individual differences as I,1568.46,5.579
just mentioned in the last slide gpd4,1571.1,4.62
already understands the importance of,1574.039,4.081
individual liberty and autonomy as well,1575.72,5.579
as how critical self-determination is to,1578.12,5.82
suffering or to reduce suffering and,1581.299,4.461
increase prosperity,1583.94,5.04
so because of that and also because it,1585.76,5.14
is aware of context the importance of,1588.98,3.26
context,1590.9,3.54
issue number three is actually less of,1592.24,4.72
an issue than you might think and,1594.44,5.4
finally number four uh and most,1596.96,4.98
importantly why would the machine hold,1599.84,4.079
to these imperatives in the first place,1601.94,4.5
and we will get into this in a lot more,1603.919,3.601
detail,1606.44,4.26
but the tldr is that with Game Theory,1607.52,5.159
and thinking of it in terms of the,1610.7,4.02
Byzantine generals problems,1612.679,4.38
all of the agis equipped with the,1614.72,3.66
heuristic imperatives would be,1617.059,3.48
incentivized to cooperate Not only would,1618.38,3.659
they be incentivized to cooperate with,1620.539,3.24
each other they'll be incentivized to,1622.039,3.321
cooperate with us,1623.779,5.78
and that results in a collective,1625.36,7.54
equilibrium in which the Hostile and,1629.559,5.321
malicious agis are going to be the,1632.9,5.279
pariahs so basically the benevolent,1634.88,5.52
machines are stronger together than the,1638.179,5.961
Hostile actors are individually,1640.4,3.74
okay great,1644.24,3.78
assuming that you're on board how do you,1645.799,3.541
implement this this sounds too,1648.02,3.18
complicated well fortunately it's,1649.34,3.66
actually not that complicated,1651.2,4.8
first is constitutional AI so I proposed,1653.0,4.74
a constitution in my book natural,1656.0,4.08
language cognitive architecture back in,1657.74,4.86
the summer of 2021 almost two years ago,1660.08,5.28
right after that anthropic AI came out,1662.6,4.38
and they did their own version of,1665.36,3.26
constitutional AI which was reduce,1666.98,5.16
harmfulness or achieve harmlessness,1668.62,5.62
I don't think anthropic's core objective,1672.14,3.899
function is good because the most,1674.24,4.439
harmless AGI is not going to be one that,1676.039,5.88
fights other malicious agis at least I,1678.679,4.98
don't think so,1681.919,4.021
um another way but still the premise of,1683.659,4.5
of implementing it in a Constitution,1685.94,3.42
which is just a natural language,1688.159,3.481
document saying how the AI should behave,1689.36,4.38
does seem to work,1691.64,4.32
reinforcement learning the heuristic,1693.74,4.02
imperatives can make a really good,1695.96,4.74
reinforcement learning signal similar to,1697.76,4.14
reinforcement learning with human,1700.7,3.599
feedback but instead use the heuristic,1701.9,4.58
imperatives as feedback so it'd be,1704.299,4.321
rlhi reinforcement learning with,1706.48,4.54
heuristic imperatives so it's just a,1708.62,4.62
different reward system this also tends,1711.02,3.779
to work pretty well I've tested it with,1713.24,5.039
fine tuning it works pretty well,1714.799,5.821
um number three planning cognitive,1718.279,3.9
control task management and,1720.62,3.36
prioritization these heuristic,1722.179,3.541
imperatives work really well with,1723.98,3.78
Frameworks such as atom which atom is a,1725.72,3.78
framework that I recently wrote about,1727.76,3.48
called autonomous task orchestration,1729.5,5.82
manager so basically as your AI system,1731.24,6.66
is coming up with and executing tasks,1735.32,4.68
you use the heuristic imperatives to,1737.9,4.32
plan the tasks to choose which tasks to,1740.0,4.559
do to prioritize them and also choose,1742.22,4.559
which tasks not to do,1744.559,4.441
and then finally for review assessment,1746.779,4.26
and self-evaluation online learning,1749.0,4.14
systems that use the heuristic,1751.039,4.461
imperatives are super easy to implement,1753.14,5.88
and and are very flexible and that can,1755.5,5.26
also allow you to label data for,1759.02,5.159
training and future decision making,1760.76,5.7
so if you're on board with all this and,1764.179,3.841
you want to read more,1766.46,3.62
um I've got it all for free on GitHub,1768.02,4.139
I've also got a few books that are on,1770.08,3.52
Barnes and Noble but most people just,1772.159,4.321
use the the free ones anyways so the,1773.6,5.22
most recent work is on my GitHub under,1776.48,5.34
Dave shop heuristic imperatives this is,1778.82,5.459
a white paper that was almost entirely,1781.82,4.32
written by gpt4 so you can see how,1784.279,3.841
nuanced gpt4's understanding of the,1786.14,3.72
problem is,1788.12,3.48
um about a year ago I published a book,1789.86,3.66
called benevolent by Design which is the,1791.6,3.84
first book that fully promotes uh,1793.52,4.74
proposes this framework and explores,1795.44,4.859
different ways to implement it and then,1798.26,4.44
finally also very recently I proposed,1800.299,4.201
the atom framework which includes the,1802.7,3.78
heuristic imperatives for task,1804.5,3.299
orchestration,1806.48,4.02
but also moreover I encourage you to,1807.799,4.26
just have a conversation with chatgpt,1810.5,3.84
about these uh plenty of people on,1812.059,3.961
Reddit and other and Discord and other,1814.34,3.78
places have tested the heuristic,1816.02,3.56
imperatives they've tried to break them,1818.12,4.98
and they and you know they use the one,1819.58,6.16
one interesting conversation was someone,1823.1,5.1
used chat GPT to try and come up with,1825.74,4.919
the the pitfalls of the heuristic,1828.2,4.62
imperatives and I said yeah like that,1830.659,3.841
just goes to show that it has a more,1832.82,3.66
nuanced understanding of the risks and,1834.5,3.72
the implementation than you do and,1836.48,3.299
they're like okay yeah I guess I see,1838.22,2.76
what you mean,1839.779,5.4
okay so part six conclusion,1840.98,6.059
as far as I can tell the problem is,1845.179,4.801
solved but there's still a lot of work,1847.039,5.161
to do,1849.98,4.74
so the problem comes down to twofold one,1852.2,5.099
is dissemination and experimentation the,1854.72,4.14
perfect solution doesn't matter if no,1857.299,3.541
one knows about it so we need to spread,1858.86,3.24
the word,1860.84,3.42
um this is why I created my YouTube,1862.1,3.66
channel,1864.26,4.26
um and even if my heuristic comparatives,1865.76,4.38
are not perfect it's the best we've got,1868.52,3.3
so far,1870.14,2.279
um,1871.82,2.88
yeah so I've been working pretty much a,1872.419,3.841
year straight to get my YouTube channel,1874.7,3.18
as big as possible,1876.26,6.36
to achieve to arrive at this moment,1877.88,6.36
another problem is that there's only so,1882.62,3.299
much experimentation I can do on my own,1884.24,3.659
now that being said lots of other people,1885.919,3.781
have started experimenting I'm working,1887.899,4.861
with various cognitive architects who,1889.7,4.56
have put the heuristic imperatives into,1892.76,4.44
their machines and again they have,1894.26,6.18
discovered that yes it is one very easy,1897.2,4.74
to implement the heuristic imperatives,1900.44,4.2
and two it does seem to drive curiosity,1901.94,5.28
and a few other uh beneficial behaviors,1904.64,4.259
for the machine it makes them very,1907.22,3.54
thoughtful,1908.899,3.061
um there's a few places that you can,1910.76,2.7
join the conversation,1911.96,2.64
um all the links are in the description,1913.46,3.42
so I just created a new subreddit called,1914.6,4.02
heuristic imperatives so that we can,1916.88,3.659
talk about these and share our work,1918.62,4.14
there's also a Discord Community,1920.539,4.26
um also Link in the description but I've,1922.76,4.26
been working on this since 2019 when,1924.799,4.321
gpt2 came out,1927.02,4.56
um and you know I will be the first to,1929.12,4.439
admit there's a lot of ways to skin this,1931.58,4.56
cat maybe my heuristic imperatives,1933.559,4.921
aren't even the best but at least now,1936.14,4.019
you're aware of the concept and you know,1938.48,4.5
how easy it is to implement so maybe the,1940.159,4.201
rest of us can collectively work,1942.98,3.66
together and implement this situation,1944.36,5.52
where even in an uncertain environment,1946.64,4.86
with potentially hostile actors the,1949.88,3.899
Byzantine generals environment we can,1951.5,4.26
have agis that will cooperate and,1953.779,4.26
collaborate and that will ultimately end,1955.76,6.0
up in a very safe and stable environment,1958.039,6.0
so all that being said thank you for,1961.76,4.38
watching please jump in the comments the,1964.039,4.441
conversation Discord and Reddit and do,1966.14,4.68
the experiments yourself I promise it's,1968.48,3.54
pretty easy,1970.82,5.0
all right that's it,1972.02,3.8