davidshapiro_youtube_transcripts
/
AGI Unleashed Game Theory Byzantine Generals and the Heuristic Imperatives_transcript.csv
text,start,duration | |
hey everyone David Shapiro here with,1.319,5.161 | |
today's video good morning,4.2,4.439 | |
today's topic is going to be a little,6.48,4.199 | |
bit more severe and a little bit more,8.639,4.981 | |
intense so the title of today's video is,10.679,6.721 | |
Agi Unleashed what happens when we lose,13.62,5.999 | |
control,17.4,5.58 | |
part one the coming storm,19.619,5.521 | |
as everyone is aware things are ramping,22.98,4.379 | |
up very quickly there are people calling,25.14,5.639 | |
for moratoriums on AI research and while,27.359,5.641 | |
some of us don't take it seriously there,30.779,4.321 | |
are very legitimate concerns about,33.0,3.6 | |
what's going on,35.1,4.139 | |
and in other words we're in the end game,36.6,5.22 | |
now we are in the ramp up towards AGI,39.239,5.521 | |
and the singularity whether or not you,41.82,5.579 | |
are emotionally ready for it,44.76,6.299 | |
so just as a tiny bit of evidence this,47.399,5.961 | |
paper from nature,51.059,4.441 | |
demonstrates that we are in an,53.36,4.96 | |
exponential ramp up on AI whatever else,55.5,5.34 | |
is true the investment is there the,58.32,4.919 | |
research is there it's happening and,60.84,5.22 | |
it's not slowing down,63.239,5.401 | |
now I'm probably preaching to the choir,66.06,4.44 | |
and rehashing old,68.64,4.56 | |
stuff that everyone knows but let's just,70.5,4.799 | |
briefly talk about the existential risks,73.2,5.459 | |
there are two overarching themes or,75.299,5.46 | |
categories of the existential risks,78.659,7.441 | |
risks for AGI one is the basically just,80.759,8.301 | |
the deliberate weaponization of AI,86.1,6.839 | |
some people are comparing AGI to nuclear,89.06,6.4 | |
weapons beyond that there's the,92.939,5.581 | |
potential of cyber warfare drone Warfare,95.46,8.24 | |
autonomous uh tanks artillery aircraft,98.52,8.279 | |
submarines so on and so forth many of,103.7,4.959 | |
these systems are already being,106.799,3.86 | |
developed and deployed,108.659,5.701 | |
so basically ai ai is already being,110.659,5.621 | |
weaponized it's just a matter of on what,114.36,2.88 | |
scale,116.28,3.9 | |
moreover the second category is,117.24,5.22 | |
accidental outcomes or accident or,120.18,4.979 | |
unintended consequences basically what,122.46,5.7 | |
if the AGI escapes but there's a few,125.159,5.341 | |
other possible avenues for this to,128.16,4.68 | |
happen for instance runaway corporate,130.5,3.36 | |
greed,132.84,2.94 | |
you bet your bottom dollar that the,133.86,4.2 | |
first corporation that can create AGI to,135.78,4.86 | |
automate profits is going to do so,138.06,5.759 | |
uh more uh beyond that there's political,140.64,4.8 | |
corruption and just straight up,143.819,3.841 | |
political incompetence this is something,145.44,3.84 | |
that has been uh discussed actually,147.66,4.2 | |
quite a bit in uh my comments section,149.28,4.76 | |
for my most recent videos which is that,151.86,8.04 | |
uh yes uh like Italy uh uh uh Britain,154.04,8.26 | |
and a few other places are doing their,159.9,4.32 | |
best to kind of get ahead of the curve,162.3,4.62 | |
but the fact of the matter is is that,164.22,4.799 | |
governments by and large are moving too,166.92,4.74 | |
slow and politicians just don't get it,169.019,5.281 | |
and then finally you can even have,171.66,6.0 | |
situations where otherwise benign agis,174.3,6.18 | |
collaborate and eventually turn on us,177.66,5.52 | |
so the existential risks are there,180.48,5.46 | |
and you guys most most of you know me I,183.18,4.62 | |
am very optimistic I'm very sanguine,185.94,3.84 | |
about all this and by the end of the,187.8,4.439 | |
video I hope that many of you will see,189.78,4.739 | |
where I'm coming from,192.239,5.881 | |
now even if AGI doesn't wipe us out,194.519,7.14 | |
there are still lots and lots of risks,198.12,5.759 | |
um job loss economic changes social,201.659,5.16 | |
upheaval so on and so forth mostly it,203.879,6.601 | |
comes down to economic shifts who gets,206.819,5.761 | |
to capture all of the tremendous wealth,210.48,5.58 | |
generated by AGI and even if AGI is safe,212.58,5.519 | |
and doesn't kill us there's still every,216.06,3.78 | |
possibility that the corporatists and,218.099,3.841 | |
capitalists will have us living in a,219.84,4.56 | |
dystopian hell before too long,221.94,5.46 | |
so that is also an intrinsic risk even,224.4,5.759 | |
if it is not an existential risk so we,227.4,4.32 | |
really need to be talking about what,230.159,3.72 | |
happens when we get these super powerful,231.72,4.76 | |
agis,233.879,2.601 | |
part two autonomous AI,236.84,7.06 | |
AGI is kind of a loaded term there's a,240.12,5.1 | |
lot of,243.9,4.38 | |
debate over what AGI even means so this,245.22,4.92 | |
is why I have started saying autonomous,248.28,4.319 | |
AI I don't care how smart it is,250.14,5.459 | |
the the point being is that AI becoming,252.599,5.04 | |
autonomous is really what we're talking,255.599,5.941 | |
about and as we talk more about what is,257.639,6.601 | |
the path to AGI autonomy is rapidly,261.54,6.18 | |
becoming part of the conversation now,264.24,5.82 | |
that being said there is a huge,267.72,4.68 | |
disconnect between what,270.06,5.4 | |
us proletariat are talking about and,272.4,6.54 | |
what is being released from the halls of,275.46,5.64 | |
power and what the academic,278.94,4.14 | |
establishment is talking about,281.1,4.68 | |
there is a lot of gatekeeping and a lot,283.08,5.52 | |
of this is cloistered and that I think,285.78,4.74 | |
is probably one of the biggest problems,288.6,3.9 | |
which is honestly why I'm making these,290.52,4.22 | |
videos,292.5,2.24 | |
autonomous AI is coming whether or not,294.96,5.22 | |
you like it whether or not you're,297.96,4.5 | |
emotionally ready for it autonomous AI,300.18,4.56 | |
is easier to make than you think people,302.46,3.959 | |
are starting to realize this especially,304.74,4.38 | |
with the release of chat GPT plugins and,306.419,4.381 | |
the ability for language models to use,309.12,3.859 | |
apis,310.8,5.88 | |
to be fair the Department of Defense and,312.979,5.741 | |
universities and tech companies are all,316.68,5.04 | |
working on autonomous AI systems they,318.72,4.94 | |
are kind of cloistering and and and,321.72,4.319 | |
closing their research it's not as open,323.66,6.16 | |
as it once was and in part I I totally,326.039,6.181 | |
understand that because I think that,329.82,4.56 | |
some of these folks realize how close we,332.22,3.0 | |
are,334.38,3.659 | |
and that scares them and so they're,335.22,4.62 | |
they're basically playing the CL playing,338.039,4.741 | |
the cards very close to their vest until,339.84,5.699 | |
they get a better read on the situation,342.78,5.46 | |
now taken one step further,345.539,5.641 | |
or or looking at at this problem no one,348.24,5.04 | |
has actually proposed a comprehensive,351.18,4.44 | |
framework uh there's been plenty of,353.28,4.28 | |
books on it there's been a few papers,355.62,5.82 | |
but there's not really anything that is,357.56,7.18 | |
a fully realized solution and I don't in,361.44,5.28 | |
my opinion nobody has even fully,364.74,4.98 | |
articulated the danger yet what's going,366.72,5.759 | |
on and I'm hoping that this video will,369.72,4.44 | |
advance that conversation just a little,372.479,3.481 | |
bit,374.16,3.78 | |
part of what happens in this comment,375.96,3.9 | |
comes up on my videos a lot is a lot of,377.94,3.479 | |
people are afraid to even talk about,379.86,4.02 | |
this out of fear of ridicule,381.419,4.021 | |
those of us that are paying attention,383.88,3.06 | |
those of us that are doing the work,385.44,3.84 | |
those of us that are working on it we,386.94,4.979 | |
all see what's possible but Society by,389.28,4.56 | |
and large is not ready for it they're,391.919,4.381 | |
not ready for the conversation and so,393.84,4.32 | |
there's a lot of ridicule whether it's,396.3,4.26 | |
in Private Industry whether it's in the,398.16,4.68 | |
halls of Academia government military,400.56,3.68 | |
and so on,402.84,5.34 | |
uh so this basically comes down to,404.24,5.32 | |
something that I call institutional,408.18,3.359 | |
codependence which is that the,409.56,3.24 | |
establishment believes that the,411.539,3.061 | |
establishment is always right and the,412.8,4.519 | |
establishment will use any tactic attack,414.6,6.06 | |
tactic or technique sorry uh to control,417.319,6.16 | |
the conversation such as shame bullying,420.66,7.14 | |
and otherwise silencing techniques uh to,423.479,6.72 | |
stymie the conversation again this is,427.8,5.78 | |
why I've created my YouTube channel,430.199,3.381 | |
instead who gets to dominate the,433.88,5.219 | |
conversation fiction and billionaires,436.68,5.639 | |
Skynet Elon Musk The Establishment has,439.099,5.38 | |
abdicated responsibility for this now,442.319,4.741 | |
that being said I am absolutely 100,444.479,6.72 | |
certain that the halls of power and and,447.06,6.24 | |
the establishment is doing the research,451.199,4.801 | |
behind closed doors,453.3,5.58 | |
but they have abdicated responsibility,456.0,5.099 | |
for controlling the narrative and,458.88,5.36 | |
guiding the conversation,461.099,3.141 | |
so let's talk about a couple aspects of,464.639,4.68 | |
the control problem so for those not in,467.039,4.56 | |
the know the control problem is the,469.319,5.28 | |
category of problems of basically once,471.599,5.04 | |
the AI becomes super intelligent how do,474.599,3.301 | |
you control it,476.639,3.481 | |
there's a few Concepts in here I'm not,477.9,3.72 | |
going to go into every single one of,480.12,3.72 | |
them but a few of them that you may or,481.62,4.62 | |
may not know about one is convergent,483.84,4.68 | |
instrumental values the idea is that any,486.24,4.92 | |
sufficiently intelligent agent will come,488.52,4.679 | |
up with certain instrumental values or,491.16,5.039 | |
goals in service to those other goals,493.199,4.081 | |
whatever the whether whatever whatever,496.199,3.541 | |
its primary purpose happens to be such,497.28,4.02 | |
as self-preservation resource,499.74,3.899 | |
acquisition and so on,501.3,4.679 | |
so basically in order to further any,503.639,5.28 | |
goal whatever it happens to be your AI,505.979,5.101 | |
might come to the conclusion that it,508.919,3.901 | |
needs to preserve itself in order to,511.08,3.959 | |
continue furthering its goal that's a,512.82,4.68 | |
pretty reasonable uh thought to have I'm,515.039,4.92 | |
not going to comment one way or another,517.5,4.38 | |
um on on whether or not I think that's,519.959,3.781 | |
going to happen ultimately I don't think,521.88,3.6 | |
it's relevant and you'll see why a,523.74,3.539 | |
little bit later second is the,525.48,3.84 | |
orthogonality thesis which basically it,527.279,3.781 | |
says very simply there is no correlation,529.32,4.199 | |
between an ai's level of intelligence,531.06,4.8 | |
and the values that it pursues,533.519,4.561 | |
the treacherous turn this is a,535.86,4.8 | |
hypothetic hypothetical situation in,538.08,7.199 | |
which a apparently benign AGI suddenly,540.66,7.44 | |
or apparently turns on its creators,545.279,5.101 | |
because it came to some conclusion that,548.1,4.32 | |
we don't understand,550.38,4.139 | |
the courage ability problem which,552.42,5.039 | |
basically says that uh your AI might,554.519,5.82 | |
might not remain open to correction or,557.459,4.981 | |
feedback or it might resist being shut,560.339,2.961 | |
down,562.44,3.6 | |
so basically you lose control of it just,563.3,5.92 | |
because uh it says I'm not I'm not down,566.04,4.5 | |
with that anymore,569.22,3.54 | |
and then finally the value loading,570.54,4.32 | |
problem which is how do you specify,572.76,4.68 | |
human wave values in such a way that the,574.86,5.22 | |
AI can understand and act on them and,577.44,4.74 | |
then the very next follow-up question is,580.08,4.68 | |
who gets to Define them anyways,582.18,4.14 | |
so again these are a few of the,584.76,4.079 | |
assertions and these and hypotheses this,586.32,4.56 | |
is not an exhaustive list but you kind,588.839,4.68 | |
of get an idea there are a lot of ideas,590.88,4.5 | |
out there and not a whole lot of,593.519,3.241 | |
solutions,595.38,4.56 | |
now speaking of there are some solutions,596.76,5.4 | |
out there and these are more like broad,599.94,3.72 | |
categories,602.16,3.0 | |
um rather than,603.66,3.84 | |
um rather than comprehensive Frameworks,605.16,6.119 | |
so one thing is kill switch Solutions,607.5,5.76 | |
um pretty self-explanatory this is a,611.279,4.141 | |
broad category of ideas,613.26,4.139 | |
um I just saw on the internet someone,615.42,5.94 | |
proposed that we we put uh bombs like,617.399,5.821 | |
remotely triggered bombs in every data,621.36,4.38 | |
center so that we can immediately shut,623.22,5.1 | |
down every data center if we need to,625.74,5.539 | |
okay sure that,628.32,5.76 | |
doesn't sound like a very reasonable uh,631.279,5.62 | |
direction to go for me but hey worst,634.08,5.699 | |
comes worst case scenario maybe we do,636.899,5.641 | |
uh courage ability which basically is,639.779,4.381 | |
just the idea of just make the machine,642.54,4.799 | |
responsive to feedback uh but what if,644.16,4.98 | |
the feedback mechanism one doesn't work,647.339,4.081 | |
or two the machine shuts it down or,649.14,5.04 | |
three the feedback mechanism,651.42,6.359 | |
um doesn't have the intended uh efficacy,654.18,5.279 | |
that we want it to have,657.779,3.421 | |
uh and then there's various kinds of,659.459,3.961 | |
reinforcement learning inverse,661.2,4.44 | |
reinforcement learning and passive and,663.42,5.4 | |
blah blah basically just include an,665.64,6.6 | |
algorithm so that the machine will,668.82,5.759 | |
automatically autonomously learn the,672.24,4.68 | |
values that we want one this is,674.579,4.141 | |
difficult to do in the first place and,676.92,3.479 | |
two what if the machine rewrites itself,678.72,4.7 | |
or accident or even accidentally,680.399,5.341 | |
nullifies uh those reinforcement,683.42,4.06 | |
learning signals,685.74,4.08 | |
finally values alignment what if you,687.48,4.56 | |
build it with the human friendly values,689.82,4.139 | |
in the first place,692.04,2.64 | |
um,693.959,2.761 | |
still you run into the same problem one,694.68,4.14 | |
how do you implement it two what if it,696.72,4.08 | |
changes its mind later and three what if,698.82,4.32 | |
the values that you gave it are poorly,700.8,4.56 | |
defined or intrinsically flawed or,703.14,5.879 | |
broken now that that being said there is,705.36,5.659 | |
a paper that literally just came out,709.019,4.201 | |
called the capacity for moral,711.019,3.88 | |
self-correction in large language models,713.22,3.66 | |
so I'm really glad to see that the,714.899,3.901 | |
establishment is,716.88,4.32 | |
at the very beginning of talking about,718.8,4.38 | |
this uh soberly,721.2,4.379 | |
so the link is here but you can just,723.18,3.779 | |
search that,725.579,3.421 | |
um I I believe this paper was published,726.959,3.901 | |
at least in part by the folks over at,729.0,4.68 | |
anthropic still these are not complete,730.86,5.099 | |
Solutions and we are literally months,733.68,4.74 | |
away from from fully autonomous AGI,735.959,5.281 | |
systems the conversation is not going,738.42,4.56 | |
fast enough,741.24,4.2 | |
so if you haven't heard found it or seen,742.98,4.26 | |
it just do a Google search for bad,745.44,3.959 | |
alignment take Bingo,747.24,4.38 | |
um these are these were circulated on,749.399,4.581 | |
Twitter probably more than a year ago,751.62,5.279 | |
these will address and and argue against,753.98,4.72 | |
many of the things that people say out,756.899,3.421 | |
there so I'm not gonna I'm not gonna,758.7,3.6 | |
rehash it or read all of them to you but,760.32,4.199 | |
I just wanted to show you that like some,762.3,3.719 | |
people are treating it like a joke it's,764.519,3.12 | |
not really a joke but this is a good way,766.019,3.241 | |
of just showing like yeah the thing that,767.639,4.861 | |
you thought has already been addressed,769.26,7.079 | |
part three the AGI landscape,772.5,5.7 | |
um the biggest takeaway here is that,776.339,3.841 | |
there's not going to be one AGI there's,778.2,4.62 | |
going to be a few features of how AGI is,780.18,5.219 | |
implemented so let's talk about that,782.82,5.519 | |
first and foremost intelligence is not,785.399,4.861 | |
binary it's not like you're going to,788.339,3.901 | |
flip the switch in one day it's AGI but,790.26,3.92 | |
the day before it wasn't,792.24,5.52 | |
basically intelligence uh and and the,794.18,5.92 | |
sophistication of AGI systems will,797.76,4.68 | |
evolve over time there are going to be,800.1,4.88 | |
various constraints such as time energy,802.44,5.76 | |
data and all of that basically means,804.98,5.74 | |
that the the level of power of your AGI,808.2,4.74 | |
system is going to be on a sliding scale,810.72,5.04 | |
so for instance even the most evil,812.94,5.28 | |
machines might not be that powerful and,815.76,4.019 | |
they're going to be constrained based on,818.22,3.179 | |
you know the processing power of the,819.779,3.24 | |
computers that they're running on the,821.399,3.841 | |
network speed that they have so on and,823.019,4.5 | |
so forth when you look at intelligence,825.24,3.779 | |
there are literally thousands of,827.519,4.861 | |
dimensions of intelligence that that the,829.019,5.161 | |
types of intelligence out there are huge,832.38,4.32 | |
and so AGI is not going to master all of,834.18,5.099 | |
them immediately it's going to take time,836.7,4.62 | |
and then as I just mentioned there are,839.279,4.201 | |
gonna there are going to be numerous,841.32,4.38 | |
limiting factors or constraints such as,843.48,3.84 | |
the underlying Hardware the training,845.7,4.98 | |
data and energy requirements of course,847.32,6.8 | |
that is going to change quickly as,850.68,5.88 | |
basically the underlying Hardware ramps,854.12,4.48 | |
up exponentially the amount of data that,856.56,4.92 | |
is available ramps up exponentially and,858.6,6.72 | |
then the underlying machine learning,861.48,5.94 | |
models the neural networks also get,865.32,3.959 | |
exponentially more sophisticated and,867.42,3.539 | |
larger,869.279,3.961 | |
now as I mentioned most importantly,870.959,5.041 | |
there won't just be one Skynet the,873.24,3.96 | |
reason that we think that there's going,876.0,2.88 | |
to be just one is because it is,877.2,3.379 | |
convenient from a narrative perspective,878.88,4.44 | |
in Terminator it's easy to just say,880.579,5.981 | |
there's one big bad there's one Skynet,883.32,5.16 | |
um but that's not how it's going to,886.56,4.38 | |
happen there's going to be hundreds,888.48,4.44 | |
thousands millions it's going to ramp up,890.94,3.24 | |
very quickly,892.92,5.52 | |
so what this results in is a sort of,894.18,6.899 | |
arms race amongst in between the agis,898.44,4.98 | |
themselves as well as the sponsors or,901.079,3.721 | |
the people trying to build and control,903.42,4.02 | |
them which results in a survival of the,904.8,5.399 | |
fittest situation or a race condition,907.44,5.459 | |
where basically the most aggressive and,910.199,5.221 | |
sophisticated and and Powerful agis are,912.899,4.161 | |
the ones who win,915.42,4.32 | |
which that could be bad because then,917.06,4.899 | |
you're basically selecting for the most,919.74,5.279 | |
aggressive and hostile agis,921.959,5.82 | |
the high velocity of AGI cyber warfare,925.019,6.18 | |
will probably require our AGI systems to,927.779,6.481 | |
be partially or fully autonomous,931.199,5.461 | |
so basically what that means is that in,934.26,5.4 | |
order to match the arms race in cyber,936.66,4.26 | |
warfare,939.66,3.6 | |
the agis that we built will probably,940.92,5.52 | |
need to be evolving which means that,943.26,4.68 | |
they'll spawn off copies of themselves,946.44,4.44 | |
they'll be polymorphic they will recode,947.94,6.24 | |
themselves so on and so forth and also,950.88,5.94 | |
when you look at AGI in the context of,954.18,4.92 | |
cyber warfare they will explicitly,956.82,4.759 | |
require adversarial objective functions,959.1,4.979 | |
this is what was explored in Skynet,961.579,4.361 | |
which basically the objective function,964.079,4.141 | |
of Skynet was probably like maximize,965.94,5.22 | |
military power or so on,968.22,5.46 | |
So In This Global AGR arms race there's,971.16,4.32 | |
going to be numerous copies they're all,973.68,3.48 | |
going to be changing which results in,975.48,3.479 | |
the Byzantine generals problem so the,977.16,3.96 | |
Byzantine generals problem is a cyber,978.959,4.261 | |
security thought experiment where,981.12,5.159 | |
wherein the idea is you have numerous,983.22,4.979 | |
generals and you don't know their,986.279,3.36 | |
allegiance you don't know their loyalty,988.199,3.601 | |
and you don't know their plans either so,989.639,3.721 | |
how do those how do those generals,991.8,3.96 | |
communicate with each other in such a,993.36,4.919 | |
way that they can understand who's on,995.76,4.499 | |
whose side and also come to consensus on,998.279,4.261 | |
what the plan is assuming that there are,1000.259,5.101 | |
hostile or adversarial actors,1002.54,5.76 | |
now thinking of this in terms of three,1005.36,5.46 | |
to five entities is difficult enough but,1008.3,3.779 | |
we're going to be talking about a,1010.82,3.36 | |
situation where there are millions or,1012.079,5.281 | |
billions of agis all of them with,1014.18,5.579 | |
unknown objective functions,1017.36,5.399 | |
as autonomous agents lastly they will,1019.759,4.92 | |
form alliances with each other,1022.759,4.32 | |
by some means or other they will,1024.679,4.081 | |
communicate they will establish their,1027.079,4.201 | |
intentions and allegiances,1028.76,4.679 | |
um and they will spend more time talking,1031.28,4.08 | |
with each other than they willed with us,1033.439,4.561 | |
this was something that is um that,1035.36,4.38 | |
people are starting to talk about some,1038.0,4.079 | |
of the folks that I I'm working with on,1039.74,4.62 | |
cognitive architecture we're realizing,1042.079,5.1 | |
that the very instant that you create a,1044.36,5.64 | |
cognitive architecture it can talk 24 7.,1047.179,5.581 | |
we can't talk 24 7 so even just by,1050.0,4.62 | |
virtue of experimenting with cognitive,1052.76,3.6 | |
architectures it makes sense to have,1054.62,4.16 | |
them talking with each other,1056.36,6.179 | |
uh and having agis talk with each other,1058.78,5.139 | |
and come to agreements and,1062.539,3.0 | |
understandings,1063.919,3.841 | |
um this is going to happen even with the,1065.539,4.621 | |
most benign benevolent outcomes of AGI,1067.76,5.64 | |
now what these what these autonomous AI,1070.16,5.879 | |
systems agree and disagree on will,1073.4,5.279 | |
likely determine the overall outcome of,1076.039,4.981 | |
what happens with the singularity with,1078.679,4.62 | |
AGI and with,1081.02,5.64 | |
um the basically the fate of humanity,1083.299,7.141 | |
part four AGI Unleashed now given,1086.66,5.22 | |
everything that I've outlined the,1090.44,3.06 | |
question remains how do you control the,1091.88,2.46 | |
machine,1093.5,5.36 | |
my answer is maybe you don't,1094.34,4.52 | |
the reason that I believe this is,1099.2,2.88 | |
because the genie is out of the bottle,1100.82,4.08 | |
open source models are proliferating you,1102.08,4.92 | |
can already run a 30 billion parameter,1104.9,4.44 | |
model on a laptop with six gigabytes of,1107.0,4.44 | |
memory that paper just came out what,1109.34,3.78 | |
yesterday or today,1111.44,3.84 | |
Global deployments of AI are rising,1113.12,4.62 | |
Federal and Military investment globally,1115.28,4.139 | |
is also Rising,1117.74,3.9 | |
because of this centralized alignment,1119.419,4.681 | |
research is completely irrelevant it,1121.64,5.159 | |
doesn't matter how responsible the most,1124.1,5.04 | |
responsible actors are there are hostile,1126.799,3.721 | |
actors out there with malevolent,1129.14,3.539 | |
intentions and they have lots of funding,1130.52,4.26 | |
not only that the AI systems are,1132.679,5.101 | |
becoming much more accessible,1134.78,5.7 | |
because of that distributed cooperation,1137.78,5.399 | |
is now required alignment is not just,1140.48,4.86 | |
about creating an individual Ai and if,1143.179,4.081 | |
you go look at the alignment the bad,1145.34,4.62 | |
alignment take Bingo none of those talk,1147.26,5.34 | |
about distribution collaboration or,1149.96,4.8 | |
collective intelligence or Collective,1152.6,4.62 | |
processing all of the all of the,1154.76,4.26 | |
conversations today are still talking,1157.22,4.56 | |
about individual agis as if they're,1159.02,5.1 | |
going to exist in a vacuum so far as I,1161.78,5.1 | |
know no one is talking about this in the,1164.12,6.059 | |
context of Game Theory and competition,1166.88,5.64 | |
so because of this we need an alignment,1170.179,4.021 | |
scheme that can create open source,1172.52,3.06 | |
collaboration amongst numerous,1174.2,4.5 | |
autonomous AGI entities such a framework,1175.58,5.459 | |
needs to be simple robust and easy to,1178.7,3.719 | |
implement,1181.039,4.821 | |
we'll get to that in just a minute,1182.419,3.441 | |
so,1186.02,3.72 | |
what I'm basically proposing is a,1187.82,3.9 | |
collective control scheme which might,1189.74,3.54 | |
sound impossible,1191.72,3.9 | |
creating one benevolent stable super,1193.28,4.2 | |
intelligence is hard enough and now I'm,1195.62,3.24 | |
saying we need to create millions of,1197.48,3.96 | |
them billions of them,1198.86,4.8 | |
what I'm saying is not that we need to,1201.44,3.84 | |
we might not have a choice in the matter,1203.66,4.68 | |
this might be the only path forward,1205.28,4.92 | |
now if you're familiar with the work of,1208.34,4.92 | |
John Nash and Game Theory you might be,1210.2,5.099 | |
able to think about this in terms of,1213.26,4.14 | |
okay let's just imagine for a minute,1215.299,3.901 | |
that there are millions of agis out,1217.4,2.94 | |
there,1219.2,3.08 | |
with many of them with unknown,1220.34,7.64 | |
intentions given a game theory,1222.28,8.019 | |
dilemmas like the prisoner's dilemma and,1227.98,4.72 | |
so on if you think about this in that,1230.299,4.981 | |
perspective it may be possible to devise,1232.7,4.979 | |
rules or assumptions that enable the AI,1235.28,4.98 | |
the agis to reach consensus on their,1237.679,5.641 | |
behavior even with the presence of,1240.26,4.98 | |
malicious and faulty actors,1243.32,5.28 | |
so what kinds of rules or assumptions,1245.24,6.0 | |
could we give our AGI systems that we're,1248.6,5.22 | |
all going to be developing independently,1251.24,6.059 | |
excuse me so that they arrive at this,1253.82,6.18 | |
equilibrium this Nash equilibrium that,1257.299,4.861 | |
we're looking for how do we ensure that,1260.0,4.02 | |
this that the millions and billions of,1262.16,4.2 | |
agis that are coming arrive at the,1264.02,5.399 | |
consensus we want them to,1266.36,4.319 | |
part five,1269.419,3.481 | |
heuristic imperatives,1270.679,4.38 | |
so now we're going to talk about the,1272.9,4.38 | |
work that I have done on this problem,1275.059,5.221 | |
and this is not just hypothetical these,1277.28,4.8 | |
there are also experiments that I've,1280.28,3.66 | |
done that are documented and I'll link,1282.08,4.2 | |
to those as well,1283.94,4.08 | |
so the heuristic imperatives that I have,1286.28,4.139 | |
come up with are quite simply one reduce,1288.02,4.62 | |
suffering in the universe two increase,1290.419,4.14 | |
prosperity in the universe and three,1292.64,4.019 | |
increase understanding in the universe,1294.559,3.48 | |
and I've been I've been talking about,1296.659,4.861 | |
these uh much more frequently lately,1298.039,5.461 | |
so let's take a deeper dive into these,1301.52,3.18 | |
imperatives,1303.5,3.48 | |
so first what is a heuristic imperative,1304.7,4.26 | |
it's a set of principles that can be,1306.98,4.319 | |
embedded into autonomous AI that,1308.96,4.079 | |
basically takes the place of intrinsic,1311.299,4.021 | |
motivations now what I want to point out,1313.039,4.921 | |
is that the gpt4 paper that Microsoft,1315.32,5.16 | |
published did mention intrinsic,1317.96,4.86 | |
motivation so again The Establishment is,1320.48,3.66 | |
starting to come around and I'm sure,1322.82,2.58 | |
they've had more conversations,1324.14,3.6 | |
internally that they are not revealing,1325.4,4.86 | |
yet but they are setting the stage to,1327.74,4.919 | |
talk about what intrinsic motivations do,1330.26,3.539 | |
we give them,1332.659,2.88 | |
so in the case of the heuristic,1333.799,4.62 | |
imperatives these are imperatives that,1335.539,6.061 | |
are uh basically provided a moral and,1338.419,4.861 | |
ethical framework as well as those,1341.6,3.78 | |
intrinsic motivations because very early,1343.28,3.899 | |
on in my research I realized that there,1345.38,4.679 | |
is no difference between an intrinsic,1347.179,4.681 | |
motivation and a moral and ethical,1350.059,4.081 | |
framework you have to have some impetus,1351.86,4.799 | |
some motivation behind and reasoning,1354.14,6.36 | |
behind all behavior and all reasoning,1356.659,5.941 | |
so why these three why suffering and,1360.5,4.98 | |
prosperity and understanding first it's,1362.6,5.76 | |
a holistic approach it uh it's a it's a,1365.48,5.28 | |
flexible framework that provides a very,1368.36,5.34 | |
Broad and yet simple to implement,1370.76,5.58 | |
framework it also balances trade-offs,1373.7,4.5 | |
remember these heuristic imperatives,1376.34,4.38 | |
have to be implemented simultaneously,1378.2,6.479 | |
and in lockstep so this forces the AI to,1380.72,5.819 | |
reason through and balance trade-offs,1384.679,3.36 | |
between,1386.539,3.841 | |
um between these objectives,1388.039,4.681 | |
they're also very adaptable and context,1390.38,5.039 | |
sensitive and basically what I mean by,1392.72,4.68 | |
that is that large language models today,1395.419,5.821 | |
like gpt4 are very very aware of the,1397.4,6.18 | |
fact that these that these general,1401.24,4.08 | |
principles these heuristic imperatives,1403.58,4.8 | |
are not the be-all end-all but they are,1405.32,5.82 | |
guidelines they're they're uh they're,1408.38,4.14 | |
shorthand,1411.14,3.539 | |
um ways of basically implementing,1412.52,4.32 | |
intuition in order to quickly make,1414.679,5.041 | |
decisions uh that adhere to a general,1416.84,5.52 | |
principle or a moral compass and then,1419.72,5.28 | |
evaluate that uh based against the,1422.36,4.1 | |
context that it's in,1425.0,4.14 | |
there's two other things that emerged,1426.46,4.719 | |
during my most recent experiments with,1429.14,4.14 | |
the heuristic imperatives and that is,1431.179,3.961 | |
that the heuristic imperatives promote,1433.28,5.04 | |
individual autonomy uh basically chat,1435.14,6.72 | |
gpt4 realized that in order to reduce,1438.32,5.64 | |
suffering of people you need to protect,1441.86,5.04 | |
individual autonomy ditto for Prosperity,1443.96,4.92 | |
that if you control people they're not,1446.9,3.06 | |
going to be happy and they're not going,1448.88,3.299 | |
to be prosperous so that was an emergent,1449.96,4.44 | |
quality of the heuristic imperatives,1452.179,4.74 | |
that surprised me and made me realize,1454.4,7.08 | |
that chat gpd4 is already capable of a,1456.919,7.981 | |
very very highly nuanced reasoning the,1461.48,5.16 | |
other emerging quality that I did,1464.9,4.62 | |
anticipate was fostering Trust,1466.64,6.06 | |
basically when you have an AI equipped,1469.52,5.1 | |
with these heuristic imperatives it,1472.7,4.92 | |
understands that um fermenting trust or,1474.62,5.46 | |
fostering trust with people is actually,1477.62,4.86 | |
critical as a subsidiary goal or an,1480.08,4.74 | |
auxiliary goal of these because if if,1482.48,4.679 | |
humans don't trust the AI the rest of,1484.82,5.7 | |
its imperatives are made irrelevant,1487.159,6.301 | |
finally there are a lot of what about,1490.52,4.92 | |
isms yeah but what about there's a lot,1493.46,3.54 | |
of protests which of course this is part,1495.44,3.0 | |
of the conversation,1497.0,3.84 | |
so the most con these are some of the,1498.44,5.04 | |
most common protests that I get when I,1500.84,4.26 | |
talk about the heuristic imperatives one,1503.48,4.079 | |
is won't reduce suffering result in the,1505.1,4.319 | |
extermination of all life the short,1507.559,4.221 | |
answer is yes if you only have that one,1509.419,5.221 | |
which is why I spent two years working,1511.78,5.56 | |
on the other two heuristic imperatives,1514.64,5.1 | |
to counterbalance them because I realize,1517.34,4.86 | |
that any single objective function is,1519.74,3.84 | |
always going to be intrinsically,1522.2,4.74 | |
unstable you must have a system that,1523.58,6.42 | |
balances multiple sometimes antagonistic,1526.94,5.58 | |
functions against each other in order to,1530.0,5.4 | |
stabilize and reach that equilibrium,1532.52,5.279 | |
number two yeah but who gets to Define,1535.4,4.139 | |
suffering prosperity and understanding,1537.799,4.26 | |
the short answer is nobody that is the,1539.539,4.201 | |
point of of implementing it as a,1542.059,4.561 | |
heuristic the machine learns as it goes,1543.74,6.299 | |
and anyways llms like gpt4 already have,1546.62,5.1 | |
a far more nuanced understanding,1550.039,4.321 | |
understanding of the concept of,1551.72,4.92 | |
suffering prosperity and understanding,1554.36,5.04 | |
um than any individual human does and,1556.64,5.22 | |
also humans have never needed perfect,1559.4,4.32 | |
definitions we learn as we go as well,1561.86,4.08 | |
and we get by,1563.72,4.74 | |
number three well what about uh cultural,1565.94,5.16 | |
biases and individual differences as I,1568.46,5.579 | |
just mentioned in the last slide gpd4,1571.1,4.62 | |
already understands the importance of,1574.039,4.081 | |
individual liberty and autonomy as well,1575.72,5.579 | |
as how critical self-determination is to,1578.12,5.82 | |
suffering or to reduce suffering and,1581.299,4.461 | |
increase prosperity,1583.94,5.04 | |
so because of that and also because it,1585.76,5.14 | |
is aware of context the importance of,1588.98,3.26 | |
context,1590.9,3.54 | |
issue number three is actually less of,1592.24,4.72 | |
an issue than you might think and,1594.44,5.4 | |
finally number four uh and most,1596.96,4.98 | |
importantly why would the machine hold,1599.84,4.079 | |
to these imperatives in the first place,1601.94,4.5 | |
and we will get into this in a lot more,1603.919,3.601 | |
detail,1606.44,4.26 | |
but the tldr is that with Game Theory,1607.52,5.159 | |
and thinking of it in terms of the,1610.7,4.02 | |
Byzantine generals problems,1612.679,4.38 | |
all of the agis equipped with the,1614.72,3.66 | |
heuristic imperatives would be,1617.059,3.48 | |
incentivized to cooperate Not only would,1618.38,3.659 | |
they be incentivized to cooperate with,1620.539,3.24 | |
each other they'll be incentivized to,1622.039,3.321 | |
cooperate with us,1623.779,5.78 | |
and that results in a collective,1625.36,7.54 | |
equilibrium in which the Hostile and,1629.559,5.321 | |
malicious agis are going to be the,1632.9,5.279 | |
pariahs so basically the benevolent,1634.88,5.52 | |
machines are stronger together than the,1638.179,5.961 | |
Hostile actors are individually,1640.4,3.74 | |
okay great,1644.24,3.78 | |
assuming that you're on board how do you,1645.799,3.541 | |
implement this this sounds too,1648.02,3.18 | |
complicated well fortunately it's,1649.34,3.66 | |
actually not that complicated,1651.2,4.8 | |
first is constitutional AI so I proposed,1653.0,4.74 | |
a constitution in my book natural,1656.0,4.08 | |
language cognitive architecture back in,1657.74,4.86 | |
the summer of 2021 almost two years ago,1660.08,5.28 | |
right after that anthropic AI came out,1662.6,4.38 | |
and they did their own version of,1665.36,3.26 | |
constitutional AI which was reduce,1666.98,5.16 | |
harmfulness or achieve harmlessness,1668.62,5.62 | |
I don't think anthropic's core objective,1672.14,3.899 | |
function is good because the most,1674.24,4.439 | |
harmless AGI is not going to be one that,1676.039,5.88 | |
fights other malicious agis at least I,1678.679,4.98 | |
don't think so,1681.919,4.021 | |
um another way but still the premise of,1683.659,4.5 | |
of implementing it in a Constitution,1685.94,3.42 | |
which is just a natural language,1688.159,3.481 | |
document saying how the AI should behave,1689.36,4.38 | |
does seem to work,1691.64,4.32 | |
reinforcement learning the heuristic,1693.74,4.02 | |
imperatives can make a really good,1695.96,4.74 | |
reinforcement learning signal similar to,1697.76,4.14 | |
reinforcement learning with human,1700.7,3.599 | |
feedback but instead use the heuristic,1701.9,4.58 | |
imperatives as feedback so it'd be,1704.299,4.321 | |
rlhi reinforcement learning with,1706.48,4.54 | |
heuristic imperatives so it's just a,1708.62,4.62 | |
different reward system this also tends,1711.02,3.779 | |
to work pretty well I've tested it with,1713.24,5.039 | |
fine tuning it works pretty well,1714.799,5.821 | |
um number three planning cognitive,1718.279,3.9 | |
control task management and,1720.62,3.36 | |
prioritization these heuristic,1722.179,3.541 | |
imperatives work really well with,1723.98,3.78 | |
Frameworks such as atom which atom is a,1725.72,3.78 | |
framework that I recently wrote about,1727.76,3.48 | |
called autonomous task orchestration,1729.5,5.82 | |
manager so basically as your AI system,1731.24,6.66 | |
is coming up with and executing tasks,1735.32,4.68 | |
you use the heuristic imperatives to,1737.9,4.32 | |
plan the tasks to choose which tasks to,1740.0,4.559 | |
do to prioritize them and also choose,1742.22,4.559 | |
which tasks not to do,1744.559,4.441 | |
and then finally for review assessment,1746.779,4.26 | |
and self-evaluation online learning,1749.0,4.14 | |
systems that use the heuristic,1751.039,4.461 | |
imperatives are super easy to implement,1753.14,5.88 | |
and and are very flexible and that can,1755.5,5.26 | |
also allow you to label data for,1759.02,5.159 | |
training and future decision making,1760.76,5.7 | |
so if you're on board with all this and,1764.179,3.841 | |
you want to read more,1766.46,3.62 | |
um I've got it all for free on GitHub,1768.02,4.139 | |
I've also got a few books that are on,1770.08,3.52 | |
Barnes and Noble but most people just,1772.159,4.321 | |
use the the free ones anyways so the,1773.6,5.22 | |
most recent work is on my GitHub under,1776.48,5.34 | |
Dave shop heuristic imperatives this is,1778.82,5.459 | |
a white paper that was almost entirely,1781.82,4.32 | |
written by gpt4 so you can see how,1784.279,3.841 | |
nuanced gpt4's understanding of the,1786.14,3.72 | |
problem is,1788.12,3.48 | |
um about a year ago I published a book,1789.86,3.66 | |
called benevolent by Design which is the,1791.6,3.84 | |
first book that fully promotes uh,1793.52,4.74 | |
proposes this framework and explores,1795.44,4.859 | |
different ways to implement it and then,1798.26,4.44 | |
finally also very recently I proposed,1800.299,4.201 | |
the atom framework which includes the,1802.7,3.78 | |
heuristic imperatives for task,1804.5,3.299 | |
orchestration,1806.48,4.02 | |
but also moreover I encourage you to,1807.799,4.26 | |
just have a conversation with chatgpt,1810.5,3.84 | |
about these uh plenty of people on,1812.059,3.961 | |
Reddit and other and Discord and other,1814.34,3.78 | |
places have tested the heuristic,1816.02,3.56 | |
imperatives they've tried to break them,1818.12,4.98 | |
and they and you know they use the one,1819.58,6.16 | |
one interesting conversation was someone,1823.1,5.1 | |
used chat GPT to try and come up with,1825.74,4.919 | |
the the pitfalls of the heuristic,1828.2,4.62 | |
imperatives and I said yeah like that,1830.659,3.841 | |
just goes to show that it has a more,1832.82,3.66 | |
nuanced understanding of the risks and,1834.5,3.72 | |
the implementation than you do and,1836.48,3.299 | |
they're like okay yeah I guess I see,1838.22,2.76 | |
what you mean,1839.779,5.4 | |
okay so part six conclusion,1840.98,6.059 | |
as far as I can tell the problem is,1845.179,4.801 | |
solved but there's still a lot of work,1847.039,5.161 | |
to do,1849.98,4.74 | |
so the problem comes down to twofold one,1852.2,5.099 | |
is dissemination and experimentation the,1854.72,4.14 | |
perfect solution doesn't matter if no,1857.299,3.541 | |
one knows about it so we need to spread,1858.86,3.24 | |
the word,1860.84,3.42 | |
um this is why I created my YouTube,1862.1,3.66 | |
channel,1864.26,4.26 | |
um and even if my heuristic comparatives,1865.76,4.38 | |
are not perfect it's the best we've got,1868.52,3.3 | |
so far,1870.14,2.279 | |
um,1871.82,2.88 | |
yeah so I've been working pretty much a,1872.419,3.841 | |
year straight to get my YouTube channel,1874.7,3.18 | |
as big as possible,1876.26,6.36 | |
to achieve to arrive at this moment,1877.88,6.36 | |
another problem is that there's only so,1882.62,3.299 | |
much experimentation I can do on my own,1884.24,3.659 | |
now that being said lots of other people,1885.919,3.781 | |
have started experimenting I'm working,1887.899,4.861 | |
with various cognitive architects who,1889.7,4.56 | |
have put the heuristic imperatives into,1892.76,4.44 | |
their machines and again they have,1894.26,6.18 | |
discovered that yes it is one very easy,1897.2,4.74 | |
to implement the heuristic imperatives,1900.44,4.2 | |
and two it does seem to drive curiosity,1901.94,5.28 | |
and a few other uh beneficial behaviors,1904.64,4.259 | |
for the machine it makes them very,1907.22,3.54 | |
thoughtful,1908.899,3.061 | |
um there's a few places that you can,1910.76,2.7 | |
join the conversation,1911.96,2.64 | |
um all the links are in the description,1913.46,3.42 | |
so I just created a new subreddit called,1914.6,4.02 | |
heuristic imperatives so that we can,1916.88,3.659 | |
talk about these and share our work,1918.62,4.14 | |
there's also a Discord Community,1920.539,4.26 | |
um also Link in the description but I've,1922.76,4.26 | |
been working on this since 2019 when,1924.799,4.321 | |
gpt2 came out,1927.02,4.56 | |
um and you know I will be the first to,1929.12,4.439 | |
admit there's a lot of ways to skin this,1931.58,4.56 | |
cat maybe my heuristic imperatives,1933.559,4.921 | |
aren't even the best but at least now,1936.14,4.019 | |
you're aware of the concept and you know,1938.48,4.5 | |
how easy it is to implement so maybe the,1940.159,4.201 | |
rest of us can collectively work,1942.98,3.66 | |
together and implement this situation,1944.36,5.52 | |
where even in an uncertain environment,1946.64,4.86 | |
with potentially hostile actors the,1949.88,3.899 | |
Byzantine generals environment we can,1951.5,4.26 | |
have agis that will cooperate and,1953.779,4.26 | |
collaborate and that will ultimately end,1955.76,6.0 | |
up in a very safe and stable environment,1958.039,6.0 | |
so all that being said thank you for,1961.76,4.38 | |
watching please jump in the comments the,1964.039,4.441 | |
conversation Discord and Reddit and do,1966.14,4.68 | |
the experiments yourself I promise it's,1968.48,3.54 | |
pretty easy,1970.82,5.0 | |
all right that's it,1972.02,3.8 | |