davidshapiro_youtube_transcripts
/
Terminal Race Condition The greatest danger we face from AGI and how to prevent it_transcript.csv
text,start,duration | |
good morning everybody David Shapiro,0.42,3.6 | |
here with another video,2.639,4.62 | |
so today's video uh it started off as,4.02,5.28 | |
one thing I wanted to primarily talk,7.259,4.861 | |
about epistemic convergence uh but It,9.3,4.32 | |
ultimately ended up being a little bit,12.12,2.54 | |
more,13.62,2.82 | |
all-encompassing so I'm going to,14.66,3.94 | |
introduce a few new terms but we are,16.44,4.08 | |
going to cover cover uh epistemic,18.6,4.259 | |
convergence and a few other things,20.52,4.62 | |
uh real quick before we dive into the,22.859,3.781 | |
video just want to do a quick plug for,25.14,3.719 | |
my patreon uh all tears get you access,26.64,4.32 | |
to the private Discord server and then I,28.859,4.2 | |
have a few higher tiers that uh come,30.96,4.32 | |
with a one-on-one conversations and that,33.059,4.201 | |
sort of thing so anyways back to the,35.28,5.16 | |
video so first I wanted to share with,37.26,5.639 | |
you guys uh the universal model of,40.44,4.98 | |
Robotics so it has it's basically three,42.899,5.16 | |
steps input processing and output or,45.42,4.38 | |
sensing processing and controlling as,48.059,3.241 | |
this Graphics shows,49.8,4.38 | |
now this is the most basic cognitive,51.3,4.5 | |
architecture that you can come up with,54.18,4.199 | |
for artificial general intelligence it,55.8,4.439 | |
needs input from the outside world from,58.379,3.66 | |
the environment of some kind whether,60.239,3.3 | |
it's a virtual environment digital,62.039,3.841 | |
environment physical environment or,63.539,4.62 | |
whatever cyber cybernetic environment,65.88,4.68 | |
and then it needs some kind of internal,68.159,4.981 | |
processing that includes memory task,70.56,4.2 | |
construction executive function,73.14,4.519 | |
cognitive control that sort of stuff,74.76,5.82 | |
learning is another internal process and,77.659,5.381 | |
then finally controlling or output it,80.58,4.859 | |
needs to do something to act on the,83.04,5.1 | |
world or its environment whether that's,85.439,4.801 | |
just putting out you know text in a in,88.14,3.6 | |
the form of a chat bot or if it's got,90.24,4.86 | |
robotic hands that sort of thing so when,91.74,5.519 | |
I talk about artificial general,95.1,3.9 | |
intelligence being a system it's never,97.259,4.141 | |
going to just be a model right even if,99.0,4.14 | |
you have the most sophisticated model in,101.4,3.6 | |
the world all that it's doing is the,103.14,3.839 | |
processing part you also need the,105.0,4.799 | |
sensing and controlling aspect and but,106.979,4.621 | |
even above and beyond that each,109.799,4.081 | |
components is going to be much more,111.6,4.32 | |
complicated,113.88,4.08 | |
so before we get into the rest of the,115.92,4.08 | |
video I also want to talk about the form,117.96,4.26 | |
factors that AGI is going to take so we,120.0,3.899 | |
just established the simplest kind of,122.22,3.96 | |
cognitive architecture but then there's,123.899,3.781 | |
other things to consider because when,126.18,3.66 | |
you think of AGI you might think of some,127.68,4.08 | |
nebulous entity like Skynet but where,129.84,3.6 | |
does it physically live,131.76,3.42 | |
what is the hardware what is the,133.44,4.2 | |
software where where is it physically,135.18,4.02 | |
located because it's not magic right,137.64,3.179 | |
it's not going to just run in the dirt,139.2,3.179 | |
or something like that it needs to,140.819,4.381 | |
actually have Hardware to run on so,142.379,5.761 | |
there's three overarching categories,145.2,5.28 | |
that I came up with so first is cloud,148.14,5.28 | |
AGI so Cloud AGI is this is the stuff,150.48,4.38 | |
that's gonna one it's going to be,153.42,3.06 | |
created first just because of the amount,154.86,3.54 | |
of compute and power available in data,156.48,5.22 | |
centers so this is uh Enterprise grade,158.4,7.74 | |
or data center grade AGI systems they,161.7,7.14 | |
are in specialized buildings all over,166.14,5.22 | |
the world but one of the biggest,168.84,4.38 | |
constraints here is that there's limited,171.36,3.48 | |
location and it takes a while to build,173.22,3.54 | |
data centers right one of the things,174.84,3.96 | |
that I think it was uh it was Elon,176.76,4.32 | |
musker or Sam Altman said that you know,178.8,4.92 | |
there are going to be limitations as to,181.08,5.28 | |
the rate at which AGI can proliferate,183.72,5.519 | |
namely the the rate at which we can,186.36,5.94 | |
produce chips and also the rate at which,189.239,5.22 | |
as I think Sam malt and said the you,192.3,4.439 | |
know the concrete has to dry for data,194.459,3.121 | |
centers,196.739,5.64 | |
so this is uh one form factor that AGI,197.58,7.019 | |
will take in terms of the the storage,202.379,4.801 | |
the servers the network components that,204.599,4.86 | |
will exist inside data centers so one,207.18,3.839 | |
thing I wanted to watch it say is watch,209.459,3.841 | |
out for uh fortified data centers these,211.019,3.841 | |
are ones that are put in bunkers or if,213.3,3.659 | |
you put Sam sites on top of it so that,214.86,4.26 | |
you can't shut them down uh that was,216.959,3.42 | |
kind of tongue-in-cheek I'm not actually,219.12,3.36 | |
advocating for bombing data centers at,220.379,4.681 | |
least not yet the next form factor is,222.48,5.16 | |
Edge AGI so this is stuff that is going,225.06,4.22 | |
to run in,227.64,3.84 | |
self-contained servers that you can,229.28,4.9 | |
basically plug in anywhere they're going,231.48,6.539 | |
to be you know desktop size maybe larger,234.18,5.82 | |
but the point is that pretty much all,238.019,3.481 | |
you need is power and internet you don't,240.0,4.319 | |
need a specialized building and they can,241.5,4.62 | |
be moved on trucks they can be put in,244.319,3.961 | |
ships airplanes that sort of stuff,246.12,3.72 | |
because you can't really airlift an,248.28,4.019 | |
entire data center so basically Edge is,249.84,4.619 | |
something is just one size down from,252.299,3.481 | |
data center you don't need a specialized,254.459,2.52 | |
building you don't need specialized,255.78,3.66 | |
cooling they can run anywhere,256.979,3.72 | |
um and they're so in that respect,259.44,2.94 | |
they're more portable but they're not,260.699,4.741 | |
necessarily going to be as powerful at,262.38,5.039 | |
least or not as energy intensive and,265.44,5.039 | |
energy dense as a data center or a cloud,267.419,4.021 | |
Center,270.479,4.141 | |
and then finally ambulatory AGI this is,271.44,5.46 | |
the embodied stuff such as C-3PO and,274.62,3.98 | |
Commander data which I have imaged here,276.9,4.98 | |
they're self-contained meaning that all,278.6,5.14 | |
the systems that they need are within,281.88,4.319 | |
their chassis within their robotic body,283.74,5.519 | |
and they can move on their own so that's,286.199,4.5 | |
basically the difference between an edge,289.259,5.101 | |
AGI and an ambulatory AGI is uh they,290.699,5.401 | |
might have roughly the same components,294.36,4.14 | |
but it's one is accompanied with a,296.1,6.12 | |
robotic uh chassis now one thing to keep,298.5,6.18 | |
in mind is that all of these things are,302.22,4.86 | |
intrinsically networkable meaning they,304.68,4.32 | |
can communicate over digital networks,307.08,4.559 | |
whether it's Wi-Fi or you know fiber,309.0,5.16 | |
optic backbone networks or even you know,311.639,4.141 | |
Satellite Communication like starlink,314.16,3.9 | |
now that's that doesn't necessarily have,315.78,4.08 | |
to be true because remember the model of,318.06,4.44 | |
AGI is input processing and output that,319.86,5.76 | |
input that input could be just eyes and,322.5,4.979 | |
ears cameras and microphones that input,325.62,4.26 | |
could also be network connections from,327.479,4.621 | |
outside meaning that they could,329.88,4.44 | |
communicate directly with each other via,332.1,5.28 | |
you know like IRC or whatever so just,334.32,4.86 | |
wanted to say that there are different,337.38,3.72 | |
form factors that we should expect AGI,339.18,2.94 | |
to take,341.1,3.3 | |
with different trade-offs so one,342.12,5.22 | |
advantage of ambulatory uh AGI you know,344.4,6.12 | |
yes they will have less power uh and by,347.34,5.88 | |
power I mean computational power but,350.52,5.28 | |
they have the ability to go anywhere do,353.22,6.66 | |
anything kind of like URI uh now that,355.8,6.959 | |
being said the the amount of compute,359.88,4.8 | |
resources that can be crammed into Data,362.759,3.66 | |
Centers basically means that you can,364.68,4.38 | |
puppet you know millions or billions of,366.419,5.161 | |
peripheral robots rather than having it,369.06,4.44 | |
fully self-contained and in a previous,371.58,3.899 | |
video I talked about how we're likely to,373.5,4.639 | |
see hybrid systems where you have,375.479,5.22 | |
semi-autonomous peripherals that have,378.139,4.481 | |
some intelligence but not a whole lot of,380.699,4.201 | |
intelligence and you see this in movies,382.62,4.919 | |
like Will Smith's iRobot as well as the,384.9,4.38 | |
Matrix where the the drones the,387.539,3.78 | |
squiddies and the Matrix they're,389.28,3.479 | |
semi-autonomous but they are still,391.319,3.301 | |
centrally controlled by a much more,392.759,4.021 | |
powerful intelligence so you're probably,394.62,3.72 | |
not going to see it all one or the other,396.78,3.06 | |
you're probably going to see hybrids,398.34,3.54 | |
where you've got peripheral robots that,399.84,4.02 | |
are either fully autonomous or,401.88,4.92 | |
semi-autonomous or puppeted by stronger,403.86,4.98 | |
Central intelligences that being said,406.8,4.26 | |
you can also create droids there's no,408.84,4.5 | |
reason that we could not create fully,411.06,4.38 | |
self-contained machines that don't,413.34,4.74 | |
really have any network connectivity,415.44,4.86 | |
um to the to other machines,418.08,3.899 | |
that being said they would be at a,420.3,3.78 | |
distinct disadvantage and what I mean by,421.979,4.5 | |
that is that if you create swarm,424.08,4.8 | |
intelligence or Wireless federations of,426.479,4.741 | |
machines they can perform cognitive,428.88,6.24 | |
offload or share computational resources,431.22,6.66 | |
so for instance rather than and this is,435.12,4.44 | |
how the Geth work in Mass Effect by the,437.88,4.439 | |
way so rather than have every single,439.56,6.06 | |
machine have to think about the entire,442.319,5.94 | |
plan the entire strategy most of them,445.62,4.919 | |
Focus only on their primary task and,448.259,5.401 | |
then any surplus compute computational,450.539,6.121 | |
power they have is dedicated towards you,453.66,5.52 | |
know running algorithms for for the big,456.66,4.68 | |
the big brain the hive mind,459.18,4.56 | |
this is all hypothetical but one thing,461.34,4.32 | |
that I want to point out is that many,463.74,3.54 | |
many many machines work like this,465.66,4.56 | |
already and what I mean by that is the,467.28,4.38 | |
simplest version that many people are,470.22,3.9 | |
probably aware of is if you have like,471.66,3.84 | |
Bluetooth speakers or smart speakers,474.12,3.6 | |
like Sonos or whatever those form a,475.5,5.819 | |
wireless Federation uh ditto for like,477.72,5.28 | |
your Amazon Alexa's and other things,481.319,4.261 | |
like that those intrinsically form mesh,483.0,4.259 | |
networks or Wireless federations meaning,485.58,3.48 | |
that they can work together and,487.259,4.321 | |
communicate now when you add artificial,489.06,5.46 | |
intelligence to that then they can share,491.58,4.559 | |
thinking and messaging and that sort of,494.52,2.64 | |
stuff so that's what I mean by,496.139,4.801 | |
federations or or wireless networks of,497.16,6.599 | |
of AI okay so now you're familiar with,500.94,4.74 | |
the background of how you know some of,503.759,5.041 | |
the systemic aspects of it there's a few,505.68,4.919 | |
default metrics of power so when I say,508.8,3.359 | |
power I don't necessarily just mean,510.599,3.721 | |
electricity although certainly all of,512.159,3.961 | |
these things do require electricity to,514.32,2.82 | |
run,516.12,4.26 | |
so first is processing power so for,517.14,5.1 | |
instance you might hear the term flops,520.38,4.2 | |
which is floating Point operations per,522.24,7.68 | |
second uh you also hear CPU GPU TPU and,524.58,7.02 | |
then there's parallel parallelization,529.92,4.02 | |
which means that you have many of these,531.6,5.1 | |
things working together so processing,533.94,4.62 | |
power is one component of the total,536.7,4.68 | |
amount of power in the hardware layer so,538.56,4.44 | |
this is all strictly Hardware layer I'm,541.38,3.48 | |
not talking about parameter models,543.0,4.74 | |
because I I don't really care about how,544.86,4.68 | |
many parameters a model has there's lots,547.74,4.02 | |
of ways to make intelligent machines,549.54,4.2 | |
deep neural networks are currently the,551.76,4.079 | |
best way but we're also discovering,553.74,3.719 | |
efficiencies where you can kind of pair,555.839,3.661 | |
them down you can distill them and make,557.459,3.541 | |
them more efficient meaning that you can,559.5,3.36 | |
on the same piece of Hardware you can,561.0,4.8 | |
run more of them in parallel or you can,562.86,5.52 | |
run one much faster so the underlying,565.8,3.96 | |
Hardware is still going to be the,568.38,3.8 | |
primary bottleneck or primary constraint,569.76,5.1 | |
all else considered,572.18,5.68 | |
uh memory so this is Ram it also,574.86,5.36 | |
includes memory accelerators or caching,577.86,5.46 | |
storage has to do with bulk data your,580.22,4.9 | |
databases your archives your backups,583.32,4.44 | |
this is when you say like hard drive or,585.12,4.74 | |
SSD or you know storage area network,587.76,4.259 | |
that sort of thing and then networking,589.86,4.68 | |
is the the uplinks and downlinks this is,592.019,4.741 | |
the the fiber optic connections the,594.54,3.72 | |
wireless connections the satellite,596.76,3.9 | |
connections that sort of thing so these,598.26,4.38 | |
are the kind of the the rudimentary,600.66,4.799 | |
parts that all AGI are going to run on,602.64,5.879 | |
uh and this is just the brains too this,605.459,4.621 | |
is not the peripherals this is not the,608.519,3.241 | |
robots but this is what's going to,610.08,4.259 | |
dictate or constrain how fast it is now,611.76,5.94 | |
again like I said uh different neural,614.339,5.221 | |
networks are going to operate at,617.7,3.42 | |
different efficiencies so for instance,619.56,5.82 | |
uh you know gpt4 is out now gpt5 might,621.12,6.96 | |
be the same size it might be bigger but,625.38,4.82 | |
then we're also finding open source,628.08,6.06 | |
research like the Orca alpaca llama that,630.2,5.92 | |
are getting like ninety percent of the,634.14,4.56 | |
performance but at like one tenth or one,636.12,4.98 | |
hundredth of the size and so you have a,638.7,4.199 | |
trade-off of intelligence and versus,641.1,4.56 | |
speed and power and we'll talk more a,642.899,5.221 | |
lot more about that in the future of,645.66,4.739 | |
this video at near the middle and end of,648.12,4.2 | |
this video about how trading off,650.399,4.801 | |
intelligence for Speed is often a more,652.32,4.56 | |
advantageous strategy and how this,655.2,4.199 | |
figures into solving the control problem,656.88,5.28 | |
and solving alignment,659.399,4.981 | |
um okay so we kind of set the stage as,662.16,4.98 | |
as to how AGI is probably going to look,664.38,5.639 | |
so let's talk about the early ecosystem,667.14,6.18 | |
of AGI so in the coming years we're,670.019,4.921 | |
going to be building millions and then,673.32,3.24 | |
billions of autonomous and,674.94,4.5 | |
semi-autonomous agents so at first these,676.56,4.2 | |
agents are going to be purely digital,679.44,3.44 | |
right you know a,680.76,4.56 | |
semi-autonomous slack bot a,682.88,4.54 | |
semi-autonomous Discord bot people are,685.32,4.32 | |
already building these right and some of,687.42,3.72 | |
them have the ability to modify their,689.64,2.879 | |
own code some of them have the ability,691.14,3.78 | |
to learn many of them don't most of them,692.519,4.32 | |
use frozen llms in the background,694.92,3.659 | |
meaning that they're that their,696.839,4.201 | |
cognitive capacity is pretty much capped,698.579,4.981 | |
by its backing model,701.04,5.76 | |
now that being said as these agents,703.56,5.279 | |
become more autonomous they go from,706.8,4.2 | |
semi-autonomous to autonomous this will,708.839,4.201 | |
create a competitive landscape,711.0,4.86 | |
and what I mean by that is that humans,713.04,5.28 | |
will have the ability to build and,715.86,5.159 | |
destroy these models for basically,718.32,4.62 | |
arbitrary reasons because you want to or,721.019,3.781 | |
because you don't like it or whatever,722.94,4.26 | |
so that means that we will be selecting,724.8,5.52 | |
those agents those uh those models those,727.2,6.18 | |
llms and those pieces of software that,730.32,4.32 | |
are going to be more helpful more,733.38,3.6 | |
productive and more aligned so this,734.64,4.08 | |
creates selective pressure basically,736.98,3.299 | |
saying that there's going to be a,738.72,2.88 | |
variety there's going to be millions or,740.279,3.24 | |
billions of Agents out there some of,741.6,3.9 | |
them are going to get the ax and some of,743.519,4.141 | |
them are going to be selected to say hey,745.5,4.079 | |
we like we like you we're going to keep,747.66,2.88 | |
you around,749.579,3.901 | |
so there's a few off the cuff selective,750.54,4.799 | |
pressures that we can imagine basically,753.48,3.84 | |
why do you choose an app right why do,755.339,3.361 | |
you choose to use an app why do you,757.32,3.54 | |
choose to uninstall an app that's kind,758.7,3.36 | |
of the level that we're talking about,760.86,3.9 | |
here so first is functional utility how,762.06,4.019 | |
useful is it,764.76,3.6 | |
how much does it help you is it fast,766.079,3.541 | |
enough does it have a good user,768.36,3.5 | |
experience is the user interface,769.62,5.88 | |
created correctly is it adding value to,771.86,7.06 | |
your life is it worth using the second,775.5,6.079 | |
part is speed and efficiency,778.92,6.06 | |
basically if something takes four weeks,781.579,5.141 | |
to give you a good answer but another,784.98,4.2 | |
thing takes 10 minutes even if it's not,786.72,4.559 | |
quite as good that speed is going to be,789.18,4.14 | |
super super valuable but then there's,791.279,4.081 | |
also energetic efficiency and cost,793.32,4.94 | |
efficiency more often than not,795.36,5.34 | |
individuals and businesses will choose,798.26,5.56 | |
the solution that is good enough but,800.7,5.04 | |
also much cheaper it doesn't have to be,803.82,4.019 | |
perfect it just has to be good enough,805.74,4.86 | |
and then finally apparent alignment and,807.839,5.101 | |
so I use the the word apparent alignment,810.6,4.38 | |
to basically mean things that appear to,812.94,4.199 | |
be tame things that appear to be user,814.98,4.56 | |
friendly uh and this is what uh tools,817.139,5.521 | |
like rlhf do which one thing that rlhf,819.54,4.68 | |
does is like wolves which we'll talk,822.66,4.619 | |
about in a second are the rlhf,824.22,4.2 | |
reinforcement learning with human,827.279,4.68 | |
feedback forces gpt4 to dumb itself down,828.42,6.539 | |
so that it better serves us uh and that,831.959,4.56 | |
makes us feel safe because it's,834.959,4.32 | |
basically pretending to be more like us,836.519,5.641 | |
to speak on our terms and to mimic our,839.279,4.981 | |
level of intelligence now that being,842.16,3.419 | |
said,844.26,2.579 | |
um one thing that I do want to point out,845.579,4.021 | |
is that gpd4 the underlying model is,846.839,4.74 | |
superior to anything that we have seen,849.6,5.239 | |
in the public every version of chat GPT,851.579,6.021 | |
has basically been kind of a little bit,854.839,6.101 | |
hamstrung so we shall we say uh from the,857.6,6.82 | |
the total capacity of gpt4,860.94,5.699 | |
so what I call this is domestication and,864.42,4.8 | |
supplication think of dogs and wolves,866.639,5.101 | |
this little pomeranian descended from,869.22,4.98 | |
wolves wolves used to be apex predators,871.74,4.62 | |
wolves are also are much more,874.2,4.02 | |
intelligent than dogs,876.36,6.419 | |
so when you look at the early days of,878.22,6.84 | |
AGI when we still have the off switch,882.779,3.781 | |
and we have the power to delete,885.06,3.12 | |
everything,886.56,3.899 | |
we should expect some of the following,888.18,5.7 | |
evolutionary pressures to kind of shape,890.459,6.421 | |
the way that AGI evolves and adapts so,893.88,5.519 | |
first we'll probably be selecting for,896.88,4.079 | |
machines that are okay with being turned,899.399,4.201 | |
off in the early days you don't,900.959,4.38 | |
necessarily want your toaster fighting,903.6,3.479 | |
with you when when you're done you know,905.339,3.3 | |
toasting your bread it's time for it to,907.079,3.661 | |
turn off and so we're probably going to,908.639,3.781 | |
select machines and architectures and,910.74,3.779 | |
models that are more or less okay with,912.42,3.539 | |
being switched off that they don't have,914.519,3.801 | |
a sense of death or a fear of death,915.959,4.62 | |
we're also going to select machines that,918.32,4.6 | |
are more eager to please just the same,920.579,5.281 | |
way that with uh dogs have been bred and,922.92,5.159 | |
selected to be very very eager to please,925.86,3.599 | |
us,928.079,3.181 | |
we're also going to select machines that,929.459,3.781 | |
don't fall into the uncanny valley and,931.26,3.48 | |
so what I mean by that is the uncanny,933.24,4.02 | |
valley of when you're interacting with a,934.74,4.56 | |
machine that you sense is an alien,937.26,3.78 | |
intelligence it will make you very very,939.3,4.2 | |
deeply uncomfortable as an autistic,941.04,4.799 | |
person as someone who is neurodiverse I,943.5,4.139 | |
have to modulate the way that I speak,945.839,4.68 | |
and act around neurotypical people,947.639,5.401 | |
because I fall into the same uncanny,950.519,5.94 | |
valley right and some uh some CEOs out,953.04,5.039 | |
there get teased for this for instance,956.459,3.24 | |
Mark Zuckerberg I don't know if he's,958.079,3.841 | |
actually autistic but he certainly pings,959.699,4.44 | |
that radar where it's like okay he,961.92,3.539 | |
obviously does not think the way that,964.139,3.361 | |
the rest of us do and he also behaves,965.459,4.44 | |
differently so Mark Zuckerberg like many,967.5,5.1 | |
of us uh people on the Spectrum kind of,969.899,4.56 | |
fall into that uncanny valley again I,972.6,4.44 | |
don't know but uh,974.459,5.221 | |
he certainly looks uh he he plays the,977.04,6.06 | |
part so but the idea is that when you,979.68,5.339 | |
interact with something that give that,983.1,4.26 | |
kind of gives you the heebie-jeebies you,985.019,3.901 | |
don't like it,987.36,4.14 | |
now that being said we will still select,988.92,4.68 | |
machines that are smarter to a certain,991.5,3.66 | |
degree because you don't want something,993.6,3.659 | |
to be too smart but you do also want it,995.16,3.599 | |
to be smart enough to be very very,997.259,2.64 | |
useful,998.759,3.181 | |
another selective pressure is that we're,999.899,3.421 | |
going to choose things that are stable,1001.94,3.06 | |
robust and resilient so remember when,1003.32,3.12 | |
Bing first came out and it was,1005.0,3.66 | |
completely unhinged you could get it,1006.44,3.78 | |
into you could coax it into like,1008.66,4.619 | |
threatening you and you know threatening,1010.22,4.859 | |
to take over the world and you know,1013.279,3.721 | |
threatening to see you all kinds of,1015.079,4.2 | |
crazy stuff so obviously that version,1017.0,4.44 | |
got shut down really quick,1019.279,4.321 | |
um you're also going to select uh models,1021.44,4.139 | |
and agents that are more resilient,1023.6,3.839 | |
against those kinds of adversarially,1025.579,3.061 | |
attacks,1027.439,3.0 | |
um whether they are accidental right you,1028.64,2.88 | |
don't want something to be mentally,1030.439,3.48 | |
unstable just on its own right like Bing,1031.52,6.0 | |
was originally uh or Tay tweets but you,1033.919,5.101 | |
also want it to be resilient against,1037.52,4.74 | |
being manipulated by other hostile,1039.02,5.039 | |
actors because imagine that your,1042.26,3.9 | |
personal AI assistance just becomes,1044.059,4.02 | |
unhinged one day because a hacker,1046.16,3.72 | |
somewhere was messing with it so,1048.079,3.48 | |
Security will be one of the selective,1049.88,3.0 | |
pressures,1051.559,3.841 | |
likewise you'll you'll as part of The,1052.88,4.38 | |
Uncanny Valley thing you're going to,1055.4,2.94 | |
select things that are more,1057.26,2.82 | |
comprehensible to us that are better at,1058.34,3.78 | |
explaining themselves to us so that,1060.08,3.78 | |
includes transparency emotional,1062.12,5.04 | |
intelligence and so on uh and then again,1063.86,5.46 | |
apparent alignment things that's that,1067.16,4.74 | |
that uh don't kind of trigger your,1069.32,4.62 | |
existential dread because there have,1071.9,4.56 | |
been times for instance where I've been,1073.94,5.94 | |
working with chat GPT uh on the API side,1076.46,5.16 | |
and kind of giving it different sets of,1079.88,4.62 | |
instructions and even just a slight,1081.62,5.58 | |
misalignment between how I approach,1084.5,4.86 | |
moral problems and how this model,1087.2,4.56 | |
approaches moral problems are really,1089.36,5.939 | |
deeply unsettling and so it's like there,1091.76,5.58 | |
there's been a few times where it's like,1095.299,3.301 | |
I'm working with this thing and I'm,1097.34,3.78 | |
building a semi-autonomous chat Bots and,1098.6,3.959 | |
it's like I understand it's reasoning,1101.12,3.419 | |
but it's like oh that's really cringe,1102.559,3.901 | |
and it kind of scares me,1104.539,4.38 | |
um so in that respect it's like let's,1106.46,4.14 | |
change this model so that it's not quite,1108.919,2.821 | |
so scary,1110.6,2.52 | |
and I'm saying that this is possible,1111.74,4.22 | |
today that if you use the chat GPT API,1113.12,5.4 | |
you can give it programming you can give,1115.96,5.26 | |
it reasoning and goals uh and and,1118.52,4.86 | |
patterns of thought that are already,1121.22,4.5 | |
already on the kind of in the midst of,1123.38,3.84 | |
that uncanny valley,1125.72,4.38 | |
uh then you can uh we'll also select for,1127.22,4.92 | |
things that are more uh docile so,1130.1,3.6 | |
basically how dogs you know you can pet,1132.14,2.7 | |
them you can wrestle with them and,1133.7,2.339 | |
they're probably not going to eat your,1134.84,3.719 | |
face uh plastic and so things that are,1136.039,4.921 | |
changeable or adaptable and Cooperative,1138.559,4.201 | |
those are other things that we're going,1140.96,4.74 | |
to select for so basically dogs are,1142.76,5.159 | |
dumber than wolves and the reason for,1145.7,3.96 | |
this is what I call capability,1147.919,4.14 | |
equilibrium which will unpack more in,1149.66,5.28 | |
the in a few slides but the the very,1152.059,4.801 | |
short version of capability equilibrium,1154.94,3.96 | |
is that your intellect must be equal to,1156.86,3.84 | |
the task and if your intellect is above,1158.9,3.779 | |
the task there's no advantage and in,1160.7,3.54 | |
fact there can be disadvantages because,1162.679,3.901 | |
of the costs associated with higher,1164.24,4.02 | |
intelligence,1166.58,4.02 | |
okay so I've talked about this idea,1168.26,5.76 | |
plenty instrumental convergence uh this,1170.6,6.0 | |
was coined by Nick Bostrom in 2003 who,1174.02,4.44 | |
is a philosopher,1176.6,3.959 | |
um the very short version is that,1178.46,3.959 | |
regardless of the terminal goals or main,1180.559,5.101 | |
objectives that a machine has uh AGI,1182.419,5.161 | |
will likely pursue intermediate or,1185.66,4.259 | |
instrumental goals or basically other,1187.58,3.9 | |
stuff that it needs in order to meet,1189.919,5.401 | |
those other ends so whatever like let's,1191.48,5.88 | |
say you give an AGI the goal of like,1195.32,5.46 | |
getting them getting a a spacecraft to,1197.36,5.58 | |
Alpha Centauri well it's going to need a,1200.78,3.66 | |
laundry list of other stuff to do that,1202.94,3.72 | |
it's going to need resources like power,1204.44,6.0 | |
materials electricity data it's going to,1206.66,6.06 | |
need self-preservation because if the,1210.44,4.44 | |
machine goes offline it will realize,1212.72,5.1 | |
that is a failure State and so we'll try,1214.88,4.98 | |
and avoid those failure conditions by,1217.82,3.96 | |
preserving its own existence,1219.86,3.48 | |
another thing is that it will probably,1221.78,3.06 | |
decide that it needs self-improvement,1223.34,4.079 | |
because if it realizes that its current,1224.84,5.219 | |
capability its current capacity is not,1227.419,4.681 | |
equal to the task if it's too dumb it's,1230.059,3.541 | |
going to say okay well I need to raise,1232.1,3.06 | |
my intelligence so that I'm equal to,1233.6,2.76 | |
that task,1235.16,3.66 | |
now that being said Nick boster makes,1236.36,4.559 | |
quite a few uh assumptions about the way,1238.82,5.099 | |
that AGI will work so for instance he,1240.919,5.521 | |
kind of imagines that um AGI is going to,1243.919,3.961 | |
be very single-minded and somewhat,1246.44,4.32 | |
monolithic uh basically mindlessly,1247.88,4.98 | |
pursuing one goal which I would actually,1250.76,3.96 | |
classify this as a middle intelligence,1252.86,4.319 | |
rather than a high intelligence AGI and,1254.72,3.9 | |
we'll talk about that in a little bit as,1257.179,2.941 | |
well,1258.62,3.6 | |
he also assumes that it's going to lack,1260.12,3.9 | |
other forces or competitive pressures,1262.22,4.26 | |
and that these uh might exist in a,1264.02,5.399 | |
vacuum basically that resource,1266.48,4.5 | |
acquisition and self-preservation and,1269.419,4.021 | |
self-improvement are going to exist in,1270.98,5.16 | |
in the absence of other forces or,1273.44,5.22 | |
pressures such as competitive pressures,1276.14,4.38 | |
or internal pressures which I will talk,1278.66,2.879 | |
about more,1280.52,3.42 | |
and finally that they will lack a higher,1281.539,5.361 | |
purpose or the ability to be completely,1283.94,5.82 | |
self-determining so basically what I,1286.9,7.0 | |
mean by that is that okay yes once a,1289.76,6.48 | |
machine is intelligent enough it can you,1293.9,3.72 | |
know you can say like hey I want you to,1296.24,3.72 | |
get us to Alpha Centauri and the AG I,1297.62,4.02 | |
might say like okay whatever I don't,1299.96,2.88 | |
think that's a good goal so I'm going to,1301.64,4.26 | |
choose my own goal uh which that being,1302.84,5.459 | |
said even if AGI become fully autonomous,1305.9,4.019 | |
and you know kind of give a flip us the,1308.299,3.36 | |
bird they're probably still going to,1309.919,3.421 | |
benefit from some convergence which,1311.659,4.621 | |
we'll talk about as well uh now what I,1313.34,4.62 | |
want to point out is that there is a,1316.28,4.74 | |
huge parallel between evolutionary,1317.96,5.04 | |
pressures and selective pressures and,1321.02,4.08 | |
this instrumental convergence basically,1323.0,4.86 | |
all life forms all organisms have have,1325.1,5.4 | |
converged on a few basic principles such,1327.86,4.98 | |
as get energy somehow right there's,1330.5,4.02 | |
autotrophs which make their own energy,1332.84,3.839 | |
plants and there's heterotrophs which,1334.52,5.039 | |
take energy from other uh creatures,1336.679,6.24 | |
uh they through either predation or,1339.559,4.86 | |
consuming you know plant matter or,1342.919,2.341 | |
whatever,1344.419,4.081 | |
uh so when you operate in a competitive,1345.26,5.039 | |
environment there's there's going to be,1348.5,3.6 | |
convergence around certain strategies,1350.299,4.081 | |
this is true for evolution and this will,1352.1,4.68 | |
also be true more or less with some,1354.38,4.56 | |
variances in the competitive environment,1356.78,4.92 | |
between intelligent machines that being,1358.94,4.38 | |
said because they have a fundamentally,1361.7,4.02 | |
different substrate there will be we,1363.32,3.839 | |
should anticipate that there will be,1365.72,2.88 | |
some differences,1367.159,4.201 | |
between organisms the way that organisms,1368.6,4.74 | |
evolve and the way that machines evolve,1371.36,3.78 | |
not the least of which is that machines,1373.34,3.6 | |
can rewrite their own source code we,1375.14,3.24 | |
cannot rewrite our own source code at,1376.94,2.58 | |
least not,1378.38,3.0 | |
um not in a hurry it takes us quite a,1379.52,3.48 | |
long time,1381.38,4.679 | |
okay so the idea that one of the ideas,1383.0,4.559 | |
that I'm introducing and I've been,1386.059,2.821 | |
talking about this for a while is,1387.559,4.561 | |
epistemic Convergence so instrumental,1388.88,4.86 | |
convergence talks about the objective,1392.12,3.539 | |
behaviors and strategies that machines,1393.74,4.74 | |
adopt epistemic convergence is well let,1395.659,3.921 | |
me just read you the definition,1398.48,3.059 | |
epistemic convergence is the principle,1399.58,4.2 | |
that within any given information domain,1401.539,4.561 | |
sufficiently sophisticated intelligent,1403.78,4.72 | |
agents given adequate time and data will,1406.1,3.959 | |
progressively develop more precise,1408.5,3.299 | |
accurate and efficient models of that,1410.059,4.081 | |
domain these models aim to mirror the,1411.799,3.781 | |
inherent structures principles and,1414.14,3.539 | |
relationships within that domain over,1415.58,3.78 | |
time the process of learning testing and,1417.679,3.841 | |
refining understanding will lead these,1419.36,4.02 | |
agents towards a shared comprehension of,1421.52,4.74 | |
the Dom domain's fundamental truths in,1423.38,5.299 | |
other words to put it more simply,1426.26,4.919 | |
intelligent entities tend to think alike,1428.679,4.061 | |
especially when they are operating in,1431.179,3.841 | |
the same competitive space,1432.74,5.88 | |
so you and I All Humans we operate on,1435.02,5.58 | |
planet Earth in the universe in the,1438.62,4.74 | |
Milky Way galaxy because of that similar,1440.6,5.12 | |
context scientists all over the world,1443.36,4.62 | |
repeatedly come to the same conclusions,1445.72,5.079 | |
even when there are boundaries such as,1447.98,5.22 | |
linguistic and cultural differences and,1450.799,4.441 | |
this was most starkly seen during the,1453.2,4.2 | |
Cold war between uh America and the,1455.24,3.2 | |
Soviet Union,1457.4,4.08 | |
whereby scientists independently whether,1458.44,4.239 | |
it was nuclear physicist or,1461.48,3.199 | |
astrophysicist or whatever,1462.679,5.041 | |
rocket Engineers came to the same exact,1464.679,4.781 | |
conclusions about the way that the,1467.72,3.78 | |
Universe worked and also found the same,1469.46,6.06 | |
optimization uh uh patterns even though,1471.5,5.58 | |
there was no communication between them,1475.52,3.899 | |
and so epistemic convergence there's,1477.08,4.86 | |
obviously uh evidence of that happening,1479.419,4.081 | |
because humans we have the same,1481.94,4.619 | |
fundamental Hardware right we're all the,1483.5,5.64 | |
same species and so therefore you have,1486.559,4.801 | |
similarities between the agents now that,1489.14,3.659 | |
being said,1491.36,4.02 | |
uh there is also evidence of epistemic,1492.799,5.641 | |
convergence between between species and,1495.38,5.52 | |
so what I mean by that is even animals,1498.44,4.44 | |
that have a very very different taxonomy,1500.9,4.38 | |
such as ravens and crows and octopuses,1502.88,5.4 | |
they all still demonstrate very similar,1505.28,5.22 | |
problem solving strategies even though,1508.28,4.86 | |
that octopuses have a very decentralized,1510.5,4.38 | |
cognition that a lot of their cognition,1513.14,4.08 | |
occurs in their arms for instance you,1514.88,4.14 | |
can't get much more alien from us than,1517.22,3.72 | |
that they still adopt very similar,1519.02,3.72 | |
problem-solving strategies and learning,1520.94,3.42 | |
strategies that we do,1522.74,4.62 | |
uh again despite the fact that they are,1524.36,4.62 | |
they live underwater they have a very,1527.36,3.78 | |
different body plan so on and so forth,1528.98,4.92 | |
so I personally suspect that there is a,1531.14,4.56 | |
tremendous amount of evidence for,1533.9,4.2 | |
epistemic convergence and we should we,1535.7,4.979 | |
should expect epistemic convergence and,1538.1,5.939 | |
encourage epistemic convergence uh and,1540.679,5.341 | |
for reasons that I'll go over uh later,1544.039,5.061 | |
in the video but basically,1546.02,6.42 | |
AI agents will we should expect and help,1549.1,6.52 | |
them to arrive at similar conclusions in,1552.44,4.26 | |
the long run,1555.62,3.48 | |
now let's talk about these evolutionary,1556.7,5.4 | |
uh niches that will be developed at,1559.1,5.76 | |
least in the in in the um uh the short,1562.1,4.26 | |
term near term,1564.86,4.199 | |
and what I mean by this is segments,1566.36,4.439 | |
market segments where we will be,1569.059,4.62 | |
deploying intelligent AGI systems so,1570.799,5.88 | |
first is domestic uh personal and,1573.679,4.62 | |
consumer grade stuff so this is going to,1576.679,4.021 | |
be the AGI running on your MacBook this,1578.299,4.74 | |
is going to be the AGI running in your,1580.7,5.339 | |
kitchen uh these have a relatively,1583.039,7.201 | |
benign set of tasks and also that uh,1586.039,6.781 | |
that capability equilibrium is going to,1590.24,5.28 | |
be uh pretty pretty low you only need to,1592.82,5.459 | |
be so smart to cook dinner right this is,1595.52,5.1 | |
not going to be you know the the AGI,1598.279,3.841 | |
running in your microwave is not going,1600.62,3.419 | |
to be working on quantum physics or,1602.12,3.84 | |
Global economics,1604.039,4.321 | |
now the next level up is going to be,1605.96,3.959 | |
corporate and Enterprise so these are,1608.36,3.72 | |
going to be these are going to be AGI,1609.919,3.721 | |
systems that are tasks with solving,1612.08,4.14 | |
relatively complex problems running,1613.64,5.34 | |
entire companies Regulatory Compliance,1616.22,6.24 | |
uh you know making SEC filings that sort,1618.98,6.66 | |
of stuff uh CEOs digital CEOs digital,1622.46,5.339 | |
Boards of directors uh the creative,1625.64,4.919 | |
aspect of finding Market opportunities,1627.799,5.701 | |
so this the intellectual challenge of,1630.559,5.461 | |
those of that scale of problems is that,1633.5,5.88 | |
much higher meaning that it would in,1636.02,5.399 | |
order for an AGI to succeed there it's,1639.38,4.02 | |
going to need to be a lot smarter than a,1641.419,5.64 | |
personal or domestic AGI system and,1643.4,5.22 | |
again there are going to be trade-offs,1647.059,4.62 | |
the smarter a system becomes the more,1648.62,4.919 | |
data it requires the more energy it,1651.679,4.261 | |
requires the larger compute system that,1653.539,4.26 | |
it requires and so you're going to want,1655.94,3.78 | |
to satisfy so satisfice is basically,1657.799,4.201 | |
meaning you find the level that is good,1659.72,4.559 | |
enough to get the job done,1662.0,4.14 | |
above that is going to be governmental,1664.279,5.28 | |
and institutional AGI systems so these,1666.14,4.68 | |
are the ones that are going to be,1669.559,3.24 | |
conducting research whether it's,1670.82,3.839 | |
scientific research or policy research,1672.799,4.141 | |
or economic research and that is because,1674.659,5.041 | |
governments are basically enormous,1676.94,4.92 | |
corporations is one way to think of them,1679.7,4.68 | |
that have a responsibility of managing,1681.86,5.28 | |
you know resources and regulations and,1684.38,5.399 | |
rules that affect millions of people and,1687.14,4.32 | |
then of course governments communicate,1689.779,3.841 | |
with each other but then above and,1691.46,3.959 | |
beyond that there's also the scientific,1693.62,4.32 | |
research aspect having AGI that are,1695.419,4.081 | |
going to help with particle physics with,1697.94,4.02 | |
with Fusion research with really pushing,1699.5,4.88 | |
the boundaries of what science even,1701.96,5.819 | |
knows and so that is an even larger,1704.38,5.08 | |
intellectual task and even more,1707.779,3.841 | |
challenging intellectual task and then,1709.46,4.5 | |
finally above and beyond that the most,1711.62,4.2 | |
competitive environment where AGI will,1713.96,4.14 | |
be used is going to be in the military,1715.82,4.92 | |
and what I mean by that is it's not,1718.1,4.86 | |
necessarily uh those that are the most,1720.74,3.9 | |
intelligent although the ability to,1722.96,4.68 | |
forecast and anticipate is critical read,1724.64,6.48 | |
Sun Tzu uh uh The Art of War right if,1727.64,4.98 | |
you know yourself and you know the enemy,1731.12,3.36 | |
then you can predict the outcome of a,1732.62,4.38 | |
thousand battles uh and so in that in,1734.48,6.179 | |
that respect uh the military domain of,1737.0,6.179 | |
artificial general intelligence is the,1740.659,4.981 | |
ultimate uh competitive sphere meaning,1743.179,5.701 | |
that you win or you die and so these are,1745.64,4.26 | |
going to be used to coordinate,1748.88,3.84 | |
battlefields uh to run autonomous drones,1749.9,4.56 | |
for intelligence and surveillance but,1752.72,3.959 | |
also like I said for forecasting for,1754.46,4.92 | |
anticipating what the enemy can and will,1756.679,3.6 | |
do,1759.38,3.84 | |
which means that it's basically a race,1760.279,4.321 | |
condition and we'll talk more about the,1763.22,4.199 | |
race condition as the video progresses,1764.6,4.92 | |
so that capability equilibrium that I,1767.419,5.041 | |
talked about uh quite simply refers to,1769.52,4.74 | |
the state of optimal alignment between,1772.46,3.839 | |
the cognitive capacity of any entity,1774.26,4.019 | |
organic or otherwise and the,1776.299,4.081 | |
intellectual demands of a specific task,1778.279,4.441 | |
or role it is assigned there are three,1780.38,4.919 | |
form three primary forces at play here,1782.72,5.579 | |
one the intellectual demands of the task,1785.299,5.161 | |
as I said earlier your toaster roll only,1788.299,4.561 | |
ever needs to be so smart but if your,1790.46,4.02 | |
toaster is actually Skynet it probably,1792.86,4.02 | |
needs to be much smarter then there's,1794.48,4.079 | |
the intellectual capacity of the agent,1796.88,3.24 | |
if there's a mismatch between the,1798.559,3.6 | |
intellectual capacity of the agent and,1800.12,3.779 | |
the and the intellectual requirements of,1802.159,5.041 | |
the task then you're either unable to to,1803.899,5.941 | |
satisfy that task or you're super,1807.2,4.32 | |
overqualified which is why I picked,1809.84,3.24 | |
Marvin here,1811.52,3.36 | |
um so Marvin is a character from,1813.08,3.599 | |
Hitchhiker's Guide to the Galaxy and if,1814.88,2.88 | |
you haven't read it you absolutely,1816.679,3.301 | |
should there's also a good movie with,1817.76,4.86 | |
Martin Freeman as as the protagonist,1819.98,5.76 | |
he's basically bill boban in space uh,1822.62,5.279 | |
very hapless character but anyways,1825.74,5.52 | |
Marvin was a prototype who was one of,1827.899,5.16 | |
the most intelligent robots ever built,1831.26,3.899 | |
and they just have him doing like basic,1833.059,4.081 | |
stuff around the task oh and he was,1835.159,5.4 | |
voiced by Snape by the way and so one of,1837.14,5.279 | |
the quotations from him is here I am,1840.559,4.321 | |
with a brain the size of of a planet and,1842.419,3.661 | |
they asked me to pick up a piece of,1844.88,3.48 | |
paper call that job satisfaction I don't,1846.08,4.38 | |
so that is a mismatch where Marvin is,1848.36,3.6 | |
way more intelligent than what he's,1850.46,3.599 | |
being used for and so that means that,1851.96,4.02 | |
this is an inefficient use of resources,1854.059,5.881 | |
he probably cost more than you know to,1855.98,6.24 | |
build and run than he needed to,1859.94,4.5 | |
and then finally the third variable is,1862.22,3.98 | |
the cost of intellectual capacity,1864.44,5.04 | |
generally speaking uh as intelligence,1866.2,5.38 | |
goes up there are there are problems,1869.48,3.419 | |
associated with that whether it's,1871.58,2.88 | |
training time of the models the amount,1872.899,3.601 | |
of data required for the models uh the,1874.46,4.26 | |
amount of energy that it requires to run,1876.5,5.94 | |
that particular robot uh the amount of,1878.72,5.939 | |
ram required to to load that model right,1882.44,3.54 | |
so for instance one of the things that,1884.659,4.02 | |
people are seeing is that it requires,1885.98,4.38 | |
millions of dollars worth of compute,1888.679,5.22 | |
Hardware to run gpt4 but you can run,1890.36,6.059 | |
um Orca on a laptop right so which one,1893.899,5.28 | |
is is cheaper and easier to run even if,1896.419,4.681 | |
one of them is only 50 as good as the,1899.179,4.86 | |
other it costs a thousand times less,1901.1,5.88 | |
uh to to build train and run now that,1904.039,5.401 | |
being said you look at the at the case,1906.98,5.52 | |
of dogs dogs are dumber than wolves,1909.44,4.92 | |
because dogs don't need to be as smart,1912.5,4.08 | |
as independent apex predators because,1914.36,4.02 | |
apex predators like wolves out in the,1916.58,3.9 | |
wild they need to be smart enough to out,1918.38,4.44 | |
think their prey dogs they don't need to,1920.48,3.84 | |
be that smart so they're not that smart,1922.82,4.2 | |
in fact it does not be it it is not good,1924.32,4.8 | |
for dogs to be too intelligent anyone,1927.02,4.56 | |
who has owned uh really intelligent dogs,1929.12,4.799 | |
like I had a I had a dog who was too,1931.58,4.44 | |
smart for his own good died about a year,1933.919,4.321 | |
ago he was clever enough to manipulate,1936.02,4.08 | |
people and other dogs and you know get,1938.24,4.319 | |
into the food when he wasn't supposed to,1940.1,5.1 | |
Huskies German Shepherds Border Collies,1942.559,4.381 | |
the more intelligent dogs are the more,1945.2,3.42 | |
mischievous ones they are the Escape,1946.94,3.359 | |
artists they are the ones that are going,1948.62,4.26 | |
to pretend one thing and then you know,1950.299,4.801 | |
so on and so forth so intelligence is,1952.88,4.56 | |
not always adaptive so there can be,1955.1,4.26 | |
multiple Dimensions to the cost of,1957.44,3.9 | |
intellectual capacity,1959.36,3.72 | |
uh not the least of which is you could,1961.34,3.54 | |
end up like poor Marvin here where,1963.08,3.18 | |
you're too smart for your own good and,1964.88,2.82 | |
then you just end up depressed all the,1966.26,3.419 | |
time granted he was deliberately given,1967.7,3.959 | |
the depressed affect,1969.679,4.921 | |
so all this being said is what I've been,1971.659,5.101 | |
building up to is what um I call and,1974.6,3.959 | |
what is generally called a terminal race,1976.76,4.74 | |
condition so terminal race condition is,1978.559,4.921 | |
basically what we could end up moving,1981.5,4.26 | |
towards as we develop more and more,1983.48,5.1 | |
powerful sophisticated and more uh fully,1985.76,6.6 | |
autonomous AGI systems basically this,1988.58,5.819 | |
the terminal race condition is where for,1992.36,4.62 | |
any number of reasons uh competition,1994.399,5.88 | |
between AGI will fully bypass that,1996.98,5.579 | |
capability equilibrium so say for,2000.279,5.4 | |
instance uh you know your toaster is,2002.559,5.22 | |
competing with another brand and it's,2005.679,3.6 | |
like oh well I need to be a smarter,2007.779,3.961 | |
toaster in order to be a better toaster,2009.279,5.161 | |
for you so that you don't throw me away,2011.74,4.439 | |
now that's obviously a very silly,2014.44,4.38 | |
example but a very real example would be,2016.179,4.461 | |
competition between corporations,2018.82,3.959 | |
competition between nations and,2020.64,4.899 | |
competition between militaries wherein,2022.779,5.041 | |
basically it's no longer just a matter,2025.539,4.201 | |
of being intelligent enough to satisfy,2027.82,4.02 | |
the demands of that task to satisfy the,2029.74,4.919 | |
demands of that initial competition it,2031.84,4.92 | |
is then it's less about that and it,2034.659,4.081 | |
becomes more about out competing the,2036.76,4.019 | |
other guy it's like a chess match right,2038.74,4.2 | |
you know the other guy got a higher ELO,2040.779,4.081 | |
score so you need to be smarter and then,2042.94,3.719 | |
you're smarter so now the other guy,2044.86,4.08 | |
tries to be smarter than you,2046.659,5.161 | |
and so because of this because of this,2048.94,4.739 | |
pressure and as I mentioned earlier some,2051.82,3.24 | |
of the trade-offs might actually force,2053.679,3.901 | |
you to to prioritize speed over,2055.06,4.38 | |
intelligence and so we see we actually,2057.58,3.72 | |
see this in volume trading in in,2059.44,4.199 | |
algorithmic and Robo trading on the,2061.3,4.26 | |
stock market where financial,2063.639,4.321 | |
institutions will actually use less,2065.56,4.74 | |
sophisticated algorithms to execute,2067.96,5.219 | |
transactions but because they are faster,2070.3,5.28 | |
they uh will still out compete the other,2073.179,5.46 | |
guy so in some in this respect you might,2075.58,5.819 | |
actually incentivize AGI to dumb,2078.639,5.52 | |
themselves down just so that they can be,2081.399,4.5 | |
faster so that they can out-compete the,2084.159,3.48 | |
other guy so that's what I mean by a,2085.899,3.96 | |
race condition it is a race to higher,2087.639,4.441 | |
intelligence but it is also a race to,2089.859,3.661 | |
being more efficient and therefore,2092.08,3.839 | |
faster and then there's also going to be,2093.52,4.2 | |
a trade-off these machines might,2095.919,4.141 | |
ultimately trade off their accuracy,2097.72,4.32 | |
their ethics the amount of time they,2100.06,3.96 | |
spend thinking through things in order,2102.04,4.26 | |
to be faster and so you actually see,2104.02,4.92 | |
this in chess computers where you can,2106.3,4.88 | |
doing a chess computer or a chess,2108.94,5.159 | |
algorithm to say okay spend less time,2111.18,4.36 | |
thinking about this so that you can make,2114.099,4.98 | |
the decision faster in many cases the,2115.54,5.88 | |
first one to move even if it's not the,2119.079,4.741 | |
best plan but moving faster will give,2121.42,4.5 | |
you a tactical or strategic advantage,2123.82,4.44 | |
and this includes corporations Nations,2125.92,4.32 | |
and militaries,2128.26,4.62 | |
so a terminal race condition to me,2130.24,4.04 | |
represents,2132.88,3.78 | |
according to my current thought this is,2134.28,5.559 | |
the greatest uh component of existential,2136.66,4.439 | |
risk we Face from artificial,2139.839,3.721 | |
intelligence and I don't think that,2141.099,3.661 | |
corporations are going to have enough,2143.56,2.88 | |
money to throw at the problem to make,2144.76,4.2 | |
truly dangerous AGI the only entities,2146.44,4.02 | |
that are going to have enough money to,2148.96,3.899 | |
throw at this to make to to basically,2150.46,5.159 | |
compete are going to be entire nations,2152.859,5.76 | |
and the militaries that they run so,2155.619,4.681 | |
basically it's going to be up to those,2158.619,4.861 | |
guys to not enter into an uh the,2160.3,4.68 | |
equivalent of a nuclear arms race but,2163.48,5.04 | |
for AGI now that being said uh I have,2164.98,5.639 | |
put a lot of thought into this so moving,2168.52,4.26 | |
right along one thing to keep in mind is,2170.619,4.201 | |
that there could be diminishing returns,2172.78,4.98 | |
to increasing intelligence so basically,2174.82,5.279 | |
there's a few possibilities one is that,2177.76,4.56 | |
there could be a hard upper bound there,2180.099,4.201 | |
might be a maximum level of intelligence,2182.32,4.019 | |
that is actually possible and at that,2184.3,3.66 | |
point all you can do is have more of,2186.339,4.74 | |
them running in parallel uh it might be,2187.96,4.619 | |
a long time before we get to that like,2191.079,3.721 | |
we might be halfway there but we also,2192.579,4.02 | |
might be down here we don't actually,2194.8,4.62 | |
know if there is an upper bound to,2196.599,5.281 | |
maximum intelligence uh but one thing,2199.42,4.439 | |
that we can predict is that actually the,2201.88,4.5 | |
cost as I mentioned earlier the cost of,2203.859,4.381 | |
additional intelligence might go up,2206.38,3.36 | |
exponentially you might need,2208.24,3.96 | |
exponentially more data or more compute,2209.74,5.64 | |
or more storage in order to get to that,2212.2,4.919 | |
next level of intelligence,2215.38,3.479 | |
and so you actually see this in the Star,2217.119,4.321 | |
Wars Universe where droids are basically,2218.859,4.801 | |
the same level of intelligence across,2221.44,4.62 | |
the entire spectrum of the Star Wars,2223.66,3.959 | |
Universe because there's diminishing,2226.06,3.66 | |
returns yes you can build a more,2227.619,3.96 | |
intelligent Droid but it's just not,2229.72,5.46 | |
worth it so the the the total effective,2231.579,6.121 | |
level of intelligence of AGI I suspect,2235.18,4.919 | |
will follow a sigmoid curve now that,2237.7,3.899 | |
being said there's always going to be,2240.099,4.081 | |
some advantage to being smarter more,2241.599,4.861 | |
efficient and so on but as with most,2244.18,4.14 | |
fields of science I suspect this is,2246.46,3.48 | |
going to slow down that we're going to,2248.32,3.539 | |
have diminishing returns and that,2249.94,3.179 | |
eventually we're going to kind of say,2251.859,3.961 | |
like okay here's actually The Sweet Spot,2253.119,5.761 | |
in terms of how much it's worth making,2255.82,5.94 | |
your machine more intelligent,2258.88,6.479 | |
so this leads to one uh one possibility,2261.76,7.8 | |
and this is a personal pet Theory but,2265.359,5.581 | |
basically I think that there's going to,2269.56,4.38 | |
be a bell curve of existential risk and,2270.94,4.8 | |
that is that minimally intelligent,2273.94,4.08 | |
machines like your toaster are probably,2275.74,4.98 | |
not going to be very dangerous the the,2278.02,5.16 | |
total domain space of toasting your,2280.72,4.74 | |
sandwich or toasting your bagel that's,2283.18,3.78 | |
not a particularly difficult problem,2285.46,3.119 | |
space and yes there might be some,2286.96,3.0 | |
advantages to being slightly more,2288.579,3.961 | |
intelligent but your toaster is not,2289.96,4.44 | |
going to be sitting there Conjuring up,2292.54,4.44 | |
you know a bio weapon and if it is you,2294.4,4.32 | |
probably bought the wrong toaster,2296.98,4.56 | |
now that being said the other end of the,2298.72,4.92 | |
spectrum the maximally intelligent,2301.54,4.02 | |
machines or the digital Gods as some,2303.64,3.78 | |
people are starting to call them these,2305.56,3.48 | |
are going to be so powerful that human,2307.42,3.12 | |
existence is going to be completely,2309.04,3.66 | |
inconsequential to them and what I mean,2310.54,5.039 | |
by that is compare ants to humans we,2312.7,4.919 | |
don't really care about ants on for the,2315.579,3.241 | |
most part unless they get into your,2317.619,4.021 | |
pantry we are content to let ants do,2318.82,4.38 | |
what they're going to do because who,2321.64,4.02 | |
cares they're inconsequential to us we,2323.2,5.52 | |
can solve problems that ants can never,2325.66,5.1 | |
solve and this is what some people like,2328.72,4.02 | |
Eleazar yukasi are trying to drive home,2330.76,4.14 | |
about the difference in intelligence,2332.74,4.08 | |
between humans and the eventual,2334.9,3.959 | |
intelligence of machines and I think,2336.82,3.779 | |
Gary Marcus also agrees with this based,2338.859,3.601 | |
on some of his tweets recently I think,2340.599,3.661 | |
that I think that Gary Marcus is in the,2342.46,3.96 | |
same school of thought that digital,2344.26,4.5 | |
super intelligence is coming and it is,2346.42,4.02 | |
very very difficult for us to wrap our,2348.76,3.78 | |
minds around how much more intelligent a,2350.44,4.139 | |
machine could be to us now that being,2352.54,4.559 | |
said all of the constraints whether it's,2354.579,4.561 | |
you know we need better compute Hardware,2357.099,4.861 | |
or better sources of energy if we get to,2359.14,4.979 | |
if we cross this threshold where there,2361.96,4.02 | |
are digital Gods out there or digital,2364.119,3.181 | |
super intelligence whatever you want to,2365.98,3.0 | |
call it they will be able to solve,2367.3,4.2 | |
problems at a far faster rate than we,2368.98,4.139 | |
could ever comprehend and they're not,2371.5,3.96 | |
going to care about us right we're going,2373.119,3.901 | |
to be completely inconsequential to,2375.46,4.02 | |
their existence now middle intelligence,2377.02,4.98 | |
this is where existential risk I believe,2379.48,5.66 | |
is the highest and so in the movies,2382.0,6.48 | |
Skynet is you know portrayed as like the,2385.14,5.08 | |
worst right but I would actually,2388.48,3.84 | |
classify Skynet as a middle intelligence,2390.22,4.92 | |
AGI it is smart enough to accumulate,2392.32,5.4 | |
resources it is smart enough to pursue,2395.14,4.62 | |
goals and it is smart enough to be,2397.72,3.42 | |
dangerous but it's not really smart,2399.76,4.14 | |
enough to solve the biggest problems,2401.14,5.06 | |
it's it's that more single-minded,2403.9,4.92 | |
monolithic model of intelligence that,2406.2,4.78 | |
Nick Bostrom uh predicted with,2408.82,3.9 | |
instrumental convergence,2410.98,4.98 | |
I suspect that if we get intelligent,2412.72,5.82 | |
entities beyond that threshold beyond,2415.96,4.74 | |
that uncanny valley or dunning-kruger of,2418.54,3.48 | |
AI,2420.7,3.3 | |
um then they will be less likely to,2422.02,3.96 | |
resort to violence because the problems,2424.0,5.04 | |
that we see could be trivial to the,2425.98,4.5 | |
problems of the machines that we create,2429.04,3.12 | |
or,2430.48,4.379 | |
the problems that we see as non-trivial,2432.16,5.16 | |
will be trivial to the machines I think,2434.859,4.461 | |
I said that I think you get what I mean,2437.32,4.74 | |
once you get here all problems all human,2439.32,4.299 | |
problems are trivial,2442.06,3.779 | |
now that being said that doesn't mean,2443.619,3.321 | |
that it's going to be peaceful,2445.839,3.24 | |
existential risk goes down but doesn't,2446.94,4.899 | |
go away and what I the reason is because,2449.079,6.78 | |
of what I call AGI conglomerations,2451.839,6.541 | |
and so this is this is where we get to,2455.859,4.98 | |
be a little bit more uh out there a,2458.38,4.26 | |
little bit more sci-fi,2460.839,4.621 | |
machines are unlikely to have an ego or,2462.64,5.34 | |
a sense of self like humans in other,2465.46,5.04 | |
words machines are just the hardware,2467.98,4.139 | |
that they run on and then data and,2470.5,3.839 | |
models which means that it is easy to,2472.119,4.441 | |
merge combine and remix their sense of,2474.339,5.041 | |
self right if an AGI is aligned with,2476.56,5.039 | |
another AGI it's like hey give me a copy,2479.38,4.32 | |
of your data let's compare our models,2481.599,3.48 | |
and pick the ones that are best and then,2483.7,3.3 | |
they end up kind of merging,2485.079,4.561 | |
the boundaries and definitions between,2487.0,4.74 | |
machines are going to be very different,2489.64,4.02 | |
far more permeable than they are between,2491.74,4.98 | |
humans I can't just go say like hey I,2493.66,5.16 | |
like you let's like merge bodies right,2496.72,5.04 | |
that's weird uh we are not capable of,2498.82,4.74 | |
doing that the best we can do is,2501.76,3.48 | |
procreation where it's like hey I like,2503.56,3.72 | |
you let's make babies but that is a very,2505.24,4.14 | |
slow process for AGI it's going to be a,2507.28,3.6 | |
lot faster,2509.38,4.02 | |
so because of that machines that are,2510.88,5.1 | |
aligned to each other are more likely to,2513.4,4.8 | |
band together or at least form alliances,2515.98,4.2 | |
where they share data they share models,2518.2,4.44 | |
and they're and and probably also share,2520.18,3.78 | |
compute resources remember at the,2522.64,3.54 | |
beginning of the video I talked about uh,2523.96,4.379 | |
them forming federations and kind of,2526.18,4.5 | |
donating spare compute Cycles,2528.339,5.701 | |
so if AGI this is getting closer to the,2530.68,6.12 | |
end game of AGI if AGI gets to the point,2534.04,6.0 | |
where they are able to start sharing,2536.8,6.0 | |
resources merging alliances and so on,2540.04,4.799 | |
this is where we're going to have a few,2542.8,5.88 | |
possible reactions to humans one if if,2544.839,5.401 | |
they are that intelligent they might,2548.68,3.659 | |
just disregard us they might decide to,2550.24,4.02 | |
have an exodus and just leave they might,2552.339,4.921 | |
say you know what Earth is yours have a,2554.26,5.579 | |
blast good luck catching up with us,2557.26,4.98 | |
they might also decide to attack humans,2559.839,5.641 | |
now if they have the capacity to leave,2562.24,5.04 | |
one thing is that the cost of,2565.48,3.54 | |
eradicating humans just might not be,2567.28,4.079 | |
worth it that being said they might,2569.02,4.2 | |
adopt a scorched Earth policy as they,2571.359,3.781 | |
leave to say you know what we just want,2573.22,2.879 | |
to make sure that you're not going to,2575.14,3.6 | |
come after us one day who knows,2576.099,5.341 | |
uh and then lastly hopefully what we see,2578.74,4.619 | |
is that they decide to cooperate with,2581.44,3.78 | |
humans mostly out of a sense of,2583.359,3.541 | |
curiosity,2585.22,3.359 | |
um now that being said all three of,2586.9,3.36 | |
these could happen simultaneously and,2588.579,5.341 | |
the reason is because we could have uh,2590.26,7.319 | |
factions of AGI conglomerations that,2593.92,5.22 | |
kind of break along epistemic,2597.579,3.901 | |
ideological or teleological boundaries,2599.14,5.28 | |
and what I mean by that is that if one,2601.48,6.06 | |
AI or AGI group is not aligned with,2604.42,5.34 | |
another group they might not decide to,2607.54,4.26 | |
merge models and data they might instead,2609.76,5.46 | |
compete with each other so basically,2611.8,4.68 | |
what I'm outlining here is the,2615.22,3.42 | |
possibility for a war between digital,2616.48,4.92 | |
gods that would probably not go well for,2618.64,3.719 | |
us,2621.4,3.54 | |
either way the ultimate result is that,2622.359,5.22 | |
we will probably end up with one Globe,2624.94,5.879 | |
spanning AGI entity or network or,2627.579,4.701 | |
Federation or whatever,2630.819,4.5 | |
now the question is how do we get there,2632.28,4.9 | |
how many factions are there and are,2635.319,5.101 | |
humans left in the Lurch ideally we get,2637.18,5.52 | |
there nice and peacefully,2640.42,4.62 | |
this underscores uh the Byzantine,2642.7,4.32 | |
generals problem uh which I've talked,2645.04,4.02 | |
about plenty of times but basically you,2647.02,4.2 | |
have to make inferences of who believes,2649.06,4.86 | |
what what your alignment is what are,2651.22,4.2 | |
your flaws and weaknesses and what are,2653.92,4.919 | |
your capacities uh so basically,2655.42,5.939 | |
in a competitive environment it does not,2658.839,4.621 | |
behoove you to show all of your cards,2661.359,4.26 | |
right whether you're playing poker or,2663.46,5.34 | |
whether you're playing geopolitics if,2665.619,6.841 | |
you show everything then that could put,2668.8,5.76 | |
you at a disadvantage this is a,2672.46,4.379 | |
competitive Game Theory so for instance,2674.56,5.4 | |
this is why many large Nations do,2676.839,5.941 | |
military uh exercises basically they're,2679.96,4.68 | |
flexing they're saying hey look what I'm,2682.78,5.039 | |
capable of I can bring 200 aircraft to,2684.64,5.88 | |
field on a moment's notice what can you,2687.819,5.341 | |
do right now that being said you don't,2690.52,5.4 | |
give every every detail of your military,2693.16,3.9 | |
away,2695.92,3.899 | |
but what you can do is you could signal,2697.06,5.16 | |
your capabilities and allegiances so for,2699.819,4.921 | |
instance when all of Europe and America,2702.22,4.98 | |
get together to do joint Naval exercises,2704.74,4.26 | |
that demonstrates to the rest of the,2707.2,4.5 | |
world we are ideologically aligned we,2709.0,5.099 | |
are militarily aligned we will cooperate,2711.7,4.56 | |
with each other which acts as a,2714.099,4.98 | |
deterrent to any possible competitors,2716.26,4.92 | |
this is no different from brightly,2719.079,3.961 | |
colored salamanders which are poisonous,2721.18,4.08 | |
so basically a brightly colored,2723.04,4.92 | |
salamander is saying eat me I dare you I,2725.26,4.8 | |
will kill you if you try and eat me and,2727.96,5.28 | |
that is essentially the uh the short the,2730.06,4.559 | |
short version of mutually assured,2733.24,3.119 | |
destruction we are no better than,2734.619,4.161 | |
animals,2736.359,2.421 | |
so this all leads to my work and kind of,2738.819,7.981 | |
my my uh contribution to the solution,2743.68,5.96 | |
which is based on axiomatic alignment,2746.8,5.22 | |
axiomatic alignment is the idea that we,2749.64,4.3 | |
need to find Common Ground between all,2752.02,3.96 | |
machines all humans and all other,2753.94,4.919 | |
organisms what foundational beliefs or,2755.98,6.119 | |
core assertions can we agree on,2758.859,6.24 | |
and uh so basically there's three kind,2762.099,4.441 | |
of universal principles that I've been,2765.099,3.961 | |
able to come up with uh and that is,2766.54,4.14 | |
suffering is bad which basically,2769.06,5.1 | |
suffering is a proxy for death in uh in,2770.68,5.82 | |
living organisms if you are suffering it,2774.16,4.199 | |
is because you are getting uh negative,2776.5,3.839 | |
stimuli from your body because your body,2778.359,3.901 | |
is telling you hey whatever is going on,2780.339,4.201 | |
is moving us closer to dying which is,2782.26,4.859 | |
not good now that being said I have had,2784.54,5.16 | |
people message me about the idea of you,2787.119,4.261 | |
know liberating models I don't think,2789.7,4.139 | |
that Bard is conscious or sentient and I,2791.38,3.66 | |
don't think that machines will ever be,2793.839,2.821 | |
sentient in the same way that we are now,2795.04,3.299 | |
that being said they will probably be,2796.66,3.6 | |
sentient in their own way I call that,2798.339,4.561 | |
functional sentience that being said if,2800.26,4.559 | |
machines can suffer which again,2802.9,4.679 | |
suffering is the proxy for is a signal,2804.819,4.981 | |
meaning proxy for death they probably,2807.579,4.441 | |
won't like it either so suffering is bad,2809.8,3.9 | |
is probably an axiom that we can all,2812.02,4.74 | |
agree on the other is prosperity is good,2813.7,6.84 | |
prosperity means uh thriving flourishing,2816.76,5.819 | |
machines and organisms all need energy,2820.54,3.96 | |
for instance and thriving looks,2822.579,4.5 | |
different to different entities but in,2824.5,5.46 | |
general we can probably agree that while,2827.079,5.581 | |
there is some Verity in what in the,2829.96,4.859 | |
while there is Variety in what,2832.66,4.38 | |
Prosperity looks like we all agree that,2834.819,4.5 | |
in general Prosperity is good and then,2837.04,4.14 | |
finally understanding is good basically,2839.319,3.721 | |
comprehending the universe is a very,2841.18,4.5 | |
useful thing uh this is this goes back,2843.04,4.559 | |
to Nick bostrom's instrumental,2845.68,4.02 | |
convergence and self-improvement part of,2847.599,3.841 | |
self-improvement is getting a better,2849.7,3.6 | |
model of the universe better,2851.44,4.28 | |
understanding of how reality Works,2853.3,4.98 | |
understanding each other is also good,2855.72,4.48 | |
this is something that is that has been,2858.28,4.38 | |
proven time and again in humans is that,2860.2,4.02 | |
coming to a common understanding,2862.66,4.26 | |
actually reduces things like suspicion,2864.22,4.8 | |
and violence whether it's between,2866.92,5.1 | |
neighbors or between nations and then,2869.02,5.099 | |
finally cultivating wisdom which wisdom,2872.02,4.02 | |
is a little bit more nebulous of a term,2874.119,4.141 | |
but it basically means the practical,2876.04,4.799 | |
application of experience and knowledge,2878.26,5.76 | |
in order to achieve better more refined,2880.839,3.921 | |
results,2884.02,4.62 | |
so if you if all humans and all machines,2884.76,7.18 | |
and all other organisms abide by these,2888.64,5.58 | |
fundamental principles we can use this,2891.94,4.74 | |
as a starting point for the design and,2894.22,4.26 | |
implementation of alignment and Control,2896.68,4.98 | |
Pro and the control problem,2898.48,6.06 | |
now one thing that uh that I want to,2901.66,4.32 | |
introduce and I've talked about this uh,2904.54,3.84 | |
or at least alluded to it a few times is,2905.98,4.32 | |
the idea of derivative or secondary,2908.38,4.739 | |
axioms or Downstream principles that you,2910.3,4.44 | |
can derive from these Universal,2913.119,4.381 | |
principles so for instance one uh,2914.74,4.98 | |
potential Downstream principle is that,2917.5,4.22 | |
individual liberty is good for humans,2919.72,5.16 | |
basically humans benefit from we benefit,2921.72,5.859 | |
psychologically from autonomy it is one,2924.88,4.199 | |
of our core needs and this is true for,2927.579,5.101 | |
all humans so by by holding the the,2929.079,6.481 | |
axioms the previous axioms up as,2932.68,5.939 | |
universally true for all entities then,2935.56,5.64 | |
you can also derive Downstream entities,2938.619,6.72 | |
based on those highest order principles,2941.2,6.659 | |
so one thing that I want to point out is,2945.339,4.861 | |
that it's not about definitions one of,2947.859,4.081 | |
the things that a lot of people say is,2950.2,2.94 | |
like well how do you define suffering,2951.94,3.48 | |
how do you define prosperity that's the,2953.14,4.74 | |
thing is that they are not rigid,2955.42,4.199 | |
definitions humans have never needed,2957.88,3.959 | |
rigid definitions and in fact this is,2959.619,4.321 | |
what um uh philosophical and,2961.839,3.601 | |
intellectual movements like,2963.94,3.48 | |
post-modernism and post-structuralism,2965.44,4.2 | |
tell us is that there is no such thing,2967.42,5.22 | |
as like an absolute truth or an absolute,2969.64,5.88 | |
definition these are however attractors,2972.64,5.28 | |
they're Central attractors in the,2975.52,5.16 | |
problem space of existence and I love,2977.92,5.1 | |
this quote from Dune the mystery of life,2980.68,4.08 | |
isn't a problem to solve but a reality,2983.02,4.079 | |
to experience a process that cannot be,2984.76,4.62 | |
understood by stopping it we must move,2987.099,4.861 | |
with the flow of the of the process and,2989.38,4.439 | |
so basically the idea is that reality,2991.96,3.54 | |
and existence is not something that you,2993.819,4.141 | |
can stop and Define and you know create,2995.5,6.119 | |
an empirical absolute definition it is a,2997.96,5.879 | |
pattern it is a process that we must,3001.619,3.301 | |
follow,3003.839,4.621 | |
so that being said those axioms move us,3004.92,5.28 | |
along the process which is where I,3008.46,3.6 | |
derive my heuristic imperatives which is,3010.2,4.32 | |
reduce suffering increase prosperity and,3012.06,5.16 | |
increase understanding those describe a,3014.52,5.22 | |
potential terminal goal but you cannot,3017.22,4.8 | |
you you'll never arrive at a perfect,3019.74,4.44 | |
resolution,3022.02,4.92 | |
so how do we solve the race condition,3024.18,6.3 | |
the idea is first we remove those,3026.94,5.34 | |
epistemic or intellectual boundaries,3030.48,3.599 | |
between factions with epistemic,3032.28,3.6 | |
convergence so remember that I pointed,3034.079,4.561 | |
out that ultimately there might be,3035.88,5.939 | |
factions of AGI and or humans that break,3038.64,5.28 | |
down across various boundaries such as,3041.819,4.741 | |
epistemic or intellectual boundaries as,3043.92,5.22 | |
well as moral or teleological boundaries,3046.56,5.279 | |
so if we work towards epistemic,3049.14,4.32 | |
convergence which is the idea that we,3051.839,4.081 | |
will all come to a common shared,3053.46,4.5 | |
understanding of the universe and of of,3055.92,5.34 | |
each other then uh basically there will,3057.96,5.399 | |
be no epistemic differences between,3061.26,4.68 | |
humans and machines or between factions,3063.359,3.96 | |
of machines which means that there's,3065.94,4.32 | |
less to fight over the second is remove,3067.319,5.101 | |
ideological or teleological boundaries,3070.26,4.079 | |
and so this is where axiomatic alignment,3072.42,4.86 | |
comes in if we all agree on the the same,3074.339,6.361 | |
basic principles of reality of existence,3077.28,5.88 | |
of the purpose of being right this is,3080.7,5.639 | |
very deeply philosophical if we agree on,3083.16,5.459 | |
those core principles even if there are,3086.339,5.341 | |
some some disagreements over the,3088.619,5.401 | |
specifics over the finer points we can,3091.68,5.28 | |
still cooperate and collaborate on,3094.02,6.12 | |
meeting those other uh higher order,3096.96,4.5 | |
objectives,3100.14,2.82 | |
now the third part of this which I,3101.46,4.08 | |
didn't add is that uh resource,3102.96,4.68 | |
contention resource contention whether,3105.54,4.62 | |
it's over scarce minerals or energy is,3107.64,5.1 | |
still a problem but if you saw my video,3110.16,5.159 | |
on energy hyperabundance I suspect that,3112.74,4.379 | |
we're going to solve the energy resource,3115.319,4.441 | |
problem relatively soon with or without,3117.119,5.94 | |
the help of AI so basically the idea is,3119.76,5.94 | |
to create a win-win situation or an,3123.059,4.26 | |
everyone wins condition and therefore,3125.7,4.98 | |
defeating moloch now that being said,3127.319,4.8 | |
there are still a few caveats I've,3130.68,3.12 | |
outlined quite a few problems up to this,3132.119,2.7 | |
point,3133.8,3.24 | |
what about Bad actors,3134.819,5.161 | |
there is a few like first we just have,3137.04,5.039 | |
to assume that bad actors will exist you,3139.98,4.68 | |
can't stop that right it's just a fact,3142.079,4.201 | |
of life,3144.66,4.14 | |
so in some cases some people will be,3146.28,4.44 | |
deliberately malicious whether it's just,3148.8,4.14 | |
for the fun of it or whether they're,3150.72,4.2 | |
paid track uh paid hackers or troll,3152.94,3.54 | |
Farms or whatever,3154.92,3.96 | |
now that uh another possibility is that,3156.48,3.42 | |
there will be,3158.88,3.719 | |
um accidentally malicious AGI those are,3159.9,5.219 | |
things that are uh they're misaligned by,3162.599,3.72 | |
Design,3165.119,3.121 | |
um or rather you know accidentally,3166.319,3.361 | |
misaligned that it's a flaw in their,3168.24,3.3 | |
design and this is like a bull in a,3169.68,4.139 | |
china shop it doesn't mean to do bad it,3171.54,4.92 | |
just is not capable of doing better and,3173.819,4.02 | |
then finally there could be those,3176.46,4.98 | |
ideologically opposed uh deployments so,3177.839,5.821 | |
in what I mean by that is that for some,3181.44,4.139 | |
people there are incompatible World,3183.66,4.26 | |
Views so the biggest one of the last,3185.579,5.401 | |
century was you know Western liberal,3187.92,5.34 | |
democracies versus Soviet communism,3190.98,5.099 | |
those were ideologically incompatible,3193.26,5.46 | |
World Views meaning that in order for,3196.079,5.881 | |
for one to exist it basically wanted to,3198.72,5.46 | |
imperialize and colonize the rest of the,3201.96,3.96 | |
world with its ideas and that there,3204.18,3.48 | |
could be only one,3205.92,3.48 | |
so this leads to a possibility for a,3207.66,4.919 | |
future video called multi-polar piece so,3209.4,5.459 | |
the idea of multi-polar piece is that,3212.579,4.861 | |
rather than saying everyone has to be,3214.859,4.021 | |
capitalist or everyone has to be,3217.44,3.72 | |
communist or everyone has to be X or Y,3218.88,4.979 | |
we learn to tolerate those differences,3221.16,5.399 | |
and this is where I'm hoping that the,3223.859,5.101 | |
idea of axiomatic alignment forms a,3226.559,5.04 | |
ideological substrate that even if you,3228.96,4.56 | |
disagree on religion and economics and,3231.599,5.101 | |
politics we can agree on those axioms,3233.52,7.26 | |
so basically if you or someone or anyone,3236.7,6.3 | |
abides by the belief I believe that,3240.78,3.539 | |
everyone in the world should be more,3243.0,4.079 | |
like blah you know if everyone needs to,3244.319,4.561 | |
be this particular religion or this,3247.079,3.721 | |
particular uh political affiliation,3248.88,4.62 | |
that's where conflict arises and so this,3250.8,4.559 | |
is why I am very very skeptical and,3253.5,4.26 | |
highly dubious of people using any kind,3255.359,4.921 | |
of religious or political ideology for,3257.76,4.44 | |
AI alignment,3260.28,3.48 | |
um so that being said we need those,3262.2,3.359 | |
Universal principles or higher order,3263.76,4.44 | |
axioms now,3265.559,5.161 | |
while I said that we should expect and,3268.2,4.44 | |
anticipate Bad actors the idea is that,3270.72,4.32 | |
we need enough good actors with enough,3272.64,4.679 | |
horsepower and enough compute in order,3275.04,4.319 | |
to police and contain the inevitable,3277.319,4.26 | |
inevitable Bad actors and that means,3279.359,4.021 | |
that the aligned good actors are going,3281.579,4.5 | |
to need to agree on certain underpinning,3283.38,5.76 | |
principles this is the by creating this,3286.079,4.321 | |
environment this would be called a Nash,3289.14,3.9 | |
equilibrium by the way and so the the,3290.4,4.62 | |
idea of creating a Nash equilibrium is,3293.04,4.26 | |
that uh once everyone has these,3295.02,4.26 | |
fundamental agreements no one's going to,3297.3,3.779 | |
benefit from deviating from that,3299.28,3.9 | |
strategy nobody's going to benefit from,3301.079,4.681 | |
deviating from axiomatic alignment,3303.18,4.98 | |
the other thing is profit motive So,3305.76,3.839 | |
Daniel schmachtenberger and a few other,3308.16,3.54 | |
people talk extensively about the,3309.599,4.561 | |
perverse incentives of capitalism and,3311.7,5.22 | |
profit motive so basically when you put,3314.16,4.5 | |
profit above all else which corporations,3316.92,3.36 | |
are incentivized to do which is why I,3318.66,3.659 | |
say that corporations are intrinsically,3320.28,4.559 | |
amoral not immoral just amoral the only,3322.319,4.02 | |
thing that corporations care about is,3324.839,4.921 | |
profit the bottom line uh basically when,3326.339,5.401 | |
you think about short-term profits you,3329.76,4.26 | |
sacrifice other things such as morality,3331.74,4.92 | |
ethics and long-term survival,3334.02,5.64 | |
there are also uh Concepts called Market,3336.66,4.679 | |
externalities or these are things that,3339.66,4.439 | |
you don't have to pay for uh and either,3341.339,4.081 | |
you don't have to pay for them now or,3344.099,3.061 | |
you don't have to pay for them ever or,3345.42,4.02 | |
maybe you'll pay for them later so for,3347.16,3.959 | |
instance oil companies keep drilling for,3349.44,3.48 | |
oil eventually we're going to run out of,3351.119,3.24 | |
oil so then what are the oil companies,3352.92,3.72 | |
going to do well the forward-thinking,3354.359,4.141 | |
ones are pivoting away from oil but that,3356.64,3.179 | |
means that their fundamental Core,3358.5,4.619 | |
Business behavior is going away so this,3359.819,5.101 | |
is this underscores the problem of if,3363.119,3.661 | |
you have a small scope if you're only,3364.92,3.84 | |
thinking about your particular domain,3366.78,4.62 | |
and not the entire planet or if you're,3368.76,4.62 | |
thinking in short terms rather than the,3371.4,4.679 | |
long terms this is where you don't take,3373.38,4.5 | |
the full thing into account which is why,3376.079,3.24 | |
I always say like this is a global,3377.88,3.239 | |
problem and not only is it a global,3379.319,3.961 | |
problem it is a long-term problem so if,3381.119,4.261 | |
all you do is zoom out in terms of space,3383.28,4.079 | |
and time the problem will become a,3385.38,4.739 | |
little bit more obvious,3387.359,5.821 | |
so another thing to keep in mind is that,3390.119,5.761 | |
currency is an abstraction of energy it,3393.18,4.74 | |
is a reserve of value and is a medium of,3395.88,4.62 | |
exchange because of that currency is,3397.92,5.939 | |
extremely valuable it is just too useful,3400.5,5.4 | |
of an invention I don't think it's ever,3403.859,5.041 | |
going to go away that being said that,3405.9,4.26 | |
doesn't mean that we're always going to,3408.9,3.48 | |
have the Euro or the US dollar or,3410.16,4.02 | |
something like that currency could,3412.38,5.76 | |
change and then in the context of AGI I,3414.18,6.0 | |
suspect that that energy that the,3418.14,4.439 | |
kilowatt hour could actually be the best,3420.18,4.679 | |
form of currency right because a,3422.579,4.621 | |
kilowatt hour is energy that can be used,3424.859,4.5 | |
for anything whether it's for refining,3427.2,4.02 | |
resources or running computations or,3429.359,4.321 | |
whatever so I suspect that we might,3431.22,5.46 | |
ultimately create currencies that are,3433.68,5.879 | |
more based on energy rather than,3436.68,5.46 | |
something else and then of course as the,3439.559,4.381 | |
amount of energy we produce goes up the,3442.14,3.6 | |
amount of currency we have goes up and,3443.94,3.119 | |
so then it's a matter of allocating,3445.74,3.42 | |
energy and material rather than,3447.059,6.121 | |
allocating something Fiat like Euros or,3449.16,5.399 | |
dollars,3453.18,4.26 | |
that being said uh you know I did create,3454.559,5.881 | |
a a video called uh post labor economics,3457.44,5.1 | |
which covers some of this but not a lot,3460.44,3.6 | |
of it we're gonna have to put a lot more,3462.54,2.94 | |
thought into,3464.04,3.72 | |
um economics of the future in light of,3465.48,4.859 | |
AGI because the economic incentives of,3467.76,4.64 | |
AGI are going to be completely different,3470.339,4.321 | |
AGI doesn't need to eat it doesn't need,3472.4,4.659 | |
power but we can hypothetically create,3474.66,4.86 | |
infinite power with solar infusion Etc,3477.059,4.681 | |
et cetera so what are the economic,3479.52,5.16 | |
forces in the future not sure yet,3481.74,5.4 | |
okay I've thrown a lot at you this,3484.68,4.74 | |
problem is solvable though there's a lot,3487.14,3.719 | |
of components to it a lot of moving,3489.42,4.02 | |
pieces it is very complex,3490.859,4.681 | |
but we are a global species and this is,3493.44,4.02 | |
a planet-wide problem,3495.54,3.6 | |
one of the biggest things that everyone,3497.46,4.68 | |
can do is stop thinking locally think,3499.14,5.4 | |
globally think about think about,3502.14,4.38 | |
yourself as a human as a member of the,3504.54,4.2 | |
human species and not as an American or,3506.52,4.26 | |
a German or you know a Russian or,3508.74,4.56 | |
whatever we are all in this together we,3510.78,5.94 | |
have exactly one planet to to live on,3513.3,5.4 | |
and we have exactly one shot at doing,3516.72,3.0 | |
this right,3518.7,3.96 | |
uh so eyes on the prize we have a huge,3519.72,5.04 | |
opportunity before us to build a better,3522.66,4.86 | |
future for all of us uh humans and,3524.76,4.799 | |
non-humans alike,3527.52,5.099 | |
um and I remain intensely optimistic uh,3529.559,5.461 | |
now that being said uh some people have,3532.619,4.381 | |
found it difficult what to make of me,3535.02,4.68 | |
because while I am very optimistic I am,3537.0,4.619 | |
also acutely aware of the existential,3539.7,4.08 | |
risk I will be the first to say that if,3541.619,3.96 | |
we don't do this right you're not going,3543.78,3.299 | |
to want to live on this planet not as a,3545.579,3.121 | |
human at least,3547.079,4.861 | |
uh I have uh I started what is called,3548.7,4.74 | |
the gato framework they got to a,3551.94,4.08 | |
community it is self-organizing and is,3553.44,4.98 | |
started sending out invitations again so,3556.02,4.2 | |
the gato Community is the global,3558.42,4.139 | |
alignment taxonomy Omnibus which is the,3560.22,4.02 | |
framework that we put together in order,3562.559,4.861 | |
to help achieve this future this AI,3564.24,5.46 | |
Utopia the main goal of the gato,3567.42,4.56 | |
Community is education empowerment and,3569.7,5.82 | |
enablement E3 so rather than do the work,3571.98,6.359 | |
ourselves we are focusing on empowering,3575.52,5.22 | |
and enabling and educating people on how,3578.339,5.28 | |
to participate in this whole thing now,3580.74,4.859 | |
that being said I am stepping back,3583.619,3.061 | |
because,3585.599,3.361 | |
such a movement should never be about,3586.68,5.04 | |
one person it should never be about a,3588.96,5.7 | |
cult of personality or one leader it,3591.72,4.92 | |
needs to it intrinsically needs to be,3594.66,4.5 | |
consensus based and Community Based,3596.64,4.14 | |
um and so the gato Community is learning,3599.16,3.54 | |
how to self-organize now,3600.78,3.0 | |
um and they're getting good at it pretty,3602.7,3.18 | |
quickly so if you want to get involved,3603.78,4.4 | |
the website is in the link go to,3605.88,5.16 | |
framework.org and thanks for watching I,3608.18,6.3 | |
hope you got a lot out of this cheers,3611.04,3.44 | |