davidshapiro_youtube_transcripts
/
Doomerism Denialism and Optimism Introducing the GATO Framework AI Avengers Assemble_transcript.csv
text,start,duration | |
hey everybody David Shapiro here with a,1.38,5.479 | |
video today's video is going to be about,4.38,7.08 | |
doomerism uh denialism uh an alternative,6.859,8.141 | |
perspective optimism as well as a very,11.46,5.22 | |
comprehensive framework that I'm putting,15.0,4.619 | |
together with a lot of folks so let's go,16.68,6.42 | |
ahead and take a look at some ideas and,19.619,4.681 | |
some data,23.1,6.419 | |
so we are all talking about exponential,24.3,7.86 | |
growth if you look at comments across,29.519,5.88 | |
the internet and even mainstream news,32.16,7.079 | |
today talking about the rise of AI one,35.399,6.301 | |
thing that happens is that a lot of,39.239,5.401 | |
people tend to think in terms of linear,41.7,6.3 | |
progress you say oh well 10 years ago we,44.64,5.759 | |
were here and now we're you know now,48.0,4.68 | |
we're there and so 10 years from now,50.399,3.84 | |
we'll basically continue with the same,52.68,3.66 | |
amount of progress that's not actually,54.239,5.701 | |
true when you shorten that time Horizon,56.34,5.76 | |
to say oh well we've made a lot of,59.94,4.14 | |
progress in the last few months maybe,62.1,4.32 | |
that's the new rate of progress that's,64.08,4.74 | |
still not actually true,66.42,4.8 | |
with exponential growth which is what we,68.82,4.86 | |
are seeing right now the actual uh,71.22,5.7 | |
correct assumption to make is that uh is,73.68,5.7 | |
that you know the X amount of time from,76.92,5.64 | |
from now will actually have continued to,79.38,4.44 | |
accelerate,82.56,4.62 | |
now this is a nice lovely handmade graph,83.82,6.659 | |
that is shown in perfect clear data but,87.18,5.34 | |
let me show you some actual some real,90.479,5.1 | |
data about parameter counts in neural,92.52,4.68 | |
networks,95.579,4.22 | |
so here you can see it growing,97.2,5.76 | |
exponentially and then the exponential,99.799,5.32 | |
curve accelerates and it starts growing,102.96,3.78 | |
logarithmically,105.119,5.221 | |
so we are at the knee of the curve,106.74,6.059 | |
already so the knee of the curve is this,110.34,4.02 | |
part right here,112.799,4.68 | |
where the acceleration really starts to,114.36,5.039 | |
take off but the thing is is when you're,117.479,3.541 | |
in the middle of it it's kind of like,119.399,3.241 | |
boiled frog syndrome which we'll talk,121.02,3.9 | |
about a little bit more in just a minute,122.64,6.599 | |
so with this data in mind let's jump,124.92,6.42 | |
into the rest of the video,129.239,5.341 | |
so I mentioned doomerism and denialism,131.34,5.64 | |
and then finally optimism these are kind,134.58,5.34 | |
of the three main categories that people,136.98,5.82 | |
by and large fall into there's also,139.92,5.52 | |
people that are apathetic uh which I,142.8,4.079 | |
didn't include that just because it's a,145.44,4.2 | |
waste of screen space but so doomerism,146.879,6.961 | |
is uh the is the the belief that uh,149.64,8.04 | |
decline collapse Calamity is inevitable,153.84,6.3 | |
that we are going to end up in some sort,157.68,4.62 | |
of Extinction scenario or dystopian,160.14,3.84 | |
outcome and that there's not really,162.3,4.5 | |
anything that we can do to change it so,163.98,4.2 | |
this is why there's been lots of,166.8,3.6 | |
comments around like Malik which the,168.18,4.08 | |
idea of Malik will get into that a,170.4,3.96 | |
little bit as well,172.26,5.4 | |
um so then there's denialism so the,174.36,5.22 | |
denialists basically say there's nothing,177.66,5.46 | |
to see here uh AGI is not even possible,179.58,6.84 | |
or it's still decades away hard takeoff,183.12,6.54 | |
is not possible or it's decades away and,186.42,5.7 | |
uh then finally optimism techno,189.66,5.159 | |
optimists is people like myself who are,192.12,4.8 | |
just like yeah like we can do this these,194.819,4.381 | |
problems are all solvable and it will,196.92,6.42 | |
ultimately end up in the better so what,199.2,6.78 | |
I want to say is is that this is I'm not,203.34,4.679 | |
talking about individuals don't take it,205.98,3.539 | |
personally if you identify with these,208.019,3.061 | |
what I'm talking about here is thought,209.519,4.201 | |
leaders uh people like content creators,211.08,5.34 | |
like myself leading scientists people on,213.72,6.48 | |
Twitter uh basically famous people or,216.42,6.78 | |
respected people in The Establishment in,220.2,7.02 | |
the industry who take these mindsets uh,223.2,6.959 | |
so again not not calling on any,227.22,6.0 | |
particular commenter or fan or people on,230.159,5.881 | |
Reddit or Twitter this is talking about,233.22,4.98 | |
basically like people at my level or,236.04,3.419 | |
above,238.2,2.459 | |
um,239.459,4.14 | |
and also this is like it obviously,240.659,4.621 | |
doesn't fall into this symbol of,243.599,3.241 | |
categories I'm just kind of talking,245.28,3.539 | |
about the the kind of extreme ends most,246.84,3.78 | |
people fall somewhere in the middle like,248.819,3.06 | |
if you were to draw this out on a,250.62,2.88 | |
triangle most people are somewhere in,251.879,3.06 | |
the middle there's a few people at the,253.5,3.54 | |
extreme points I'm an extreme optimist,254.939,3.14 | |
so,257.04,3.84 | |
in in the yellow corner is the extreme,258.079,4.481 | |
optimists in the red and green Corners,260.88,4.259 | |
are the dumerous and denialists,262.56,5.88 | |
um okay so but as promised by the,265.139,4.681 | |
opening title I want to take a,268.44,4.08 | |
sympathetic look at these other uh other,269.82,4.92 | |
uh uh dispositions,272.52,6.06 | |
so Sympathy for the Doomer one it is,274.74,5.76 | |
good to acknowledge the existential,278.58,4.98 | |
risks of AI this has been true of all,280.5,5.22 | |
new technologies whether it's Medical,283.56,4.5 | |
Technology nuclear technology pretty,285.72,4.68 | |
much every new technology today carries,288.06,4.38 | |
with it some level of existential risk,290.4,4.859 | |
right you know the whether it's the,292.44,5.1 | |
ability to do Gene engineering or,295.259,5.16 | |
engineer new uh strains of flu or,297.54,5.099 | |
coronavirus or whatever there's always,300.419,4.5 | |
risks,302.639,5.041 | |
um the doomers understand the potential,304.919,5.821 | |
risk of uncontrolled AGI right the sky,307.68,5.4 | |
is the limit right as people are,310.74,4.08 | |
learning more about what AI is capable,313.08,4.44 | |
of the idea of an Extinction scenario,314.82,5.58 | |
like Skynet is actually not entirely,317.52,4.92 | |
impossible and when you look at the fact,320.4,4.2 | |
that Congress right now is working on,322.44,4.5 | |
passing legislation so that AI will,324.6,4.439 | |
never have uh control over nuclear,326.94,4.02 | |
weapons like they're taking it seriously,329.039,4.38 | |
too right so like there's something here,330.96,5.76 | |
uh there's it's not nothing,333.419,5.101 | |
um so then there's also the recognition,336.72,4.8 | |
for safeguards and regulations and then,338.52,4.56 | |
finally when you just look at the,341.52,3.72 | |
current trends,343.08,3.839 | |
um like stagnant wages and wealth,345.24,3.42 | |
inequality and other evidence of the,346.919,5.701 | |
Malik problem like it doesn't take a you,348.66,6.599 | |
know a a great leap of faith or logic to,352.62,4.139 | |
say what if these Trends continue and,355.259,3.541 | |
get worse which there's no evidence of,356.759,4.201 | |
some of these Trends reversing,358.8,3.66 | |
um then it's like okay well then we are,360.96,3.9 | |
all going to end up in a cyberpunk Hell,362.46,4.98 | |
and then finally these problems are all,364.86,4.32 | |
very large and complex they are global,367.44,4.68 | |
scale problems so what I want to say is,369.18,4.799 | |
I want to acknowledge that these are the,372.12,3.96 | |
primary as far as I can tell the primary,373.979,4.201 | |
concerns of doomers,376.08,5.1 | |
um and uh like there is some legitimacy,378.18,4.44 | |
to this position I'm not saying oh,381.18,3.18 | |
doomers are just flat out wrong you know,382.62,3.359 | |
to ignore them like no these are real,384.36,3.6 | |
things I need to acknowledge that but,385.979,4.201 | |
what I'll get to is like why I'm still,387.96,5.239 | |
optimistic despite all this,390.18,6.959 | |
now to play Devil's Advocate there are,393.199,6.461 | |
some flaws with tumorism which is people,397.139,5.06 | |
that just stick in this corner,399.66,4.979 | |
one is over emphasis on worst case,402.199,3.821 | |
scenarios,404.639,3.301 | |
yes we can think about worst case,406.02,4.14 | |
scenarios but it does not help for us to,407.94,4.08 | |
dwell on worst case scenarios and only,410.16,3.9 | |
worst case scenarios we need to think,412.02,3.48 | |
about the entire spectrum of,414.06,2.699 | |
possibilities,415.5,3.18 | |
another thing that is that is common,416.759,4.261 | |
with some doomers is that they're very,418.68,4.38 | |
dogmatic in their thinking they have,421.02,3.959 | |
come to believe for their own reasons,423.06,3.24 | |
with their own logic and their own,424.979,3.261 | |
research and minds and whatever else,426.3,4.679 | |
that catastrophe is a foregone,428.24,5.079 | |
conclusion they think that it is totally,430.979,5.101 | |
inevitable which results in dogmatic and,433.319,4.1 | |
rigid thinking,436.08,4.92 | |
this mentality discourages Innovation,437.419,5.441 | |
and collaboration they're like ah,441.0,4.38 | |
we're doomed who cares give up just,442.86,4.08 | |
throw your hands up and just let it,445.38,3.24 | |
happen,446.94,3.18 | |
um which creates a distraction from,448.62,4.139 | |
finding real solutions and the ultimate,450.12,4.68 | |
result of this is from an emotional and,452.759,3.72 | |
psychological perspective is that it,454.8,4.14 | |
leads to a sense of nihilism or fatalism,456.479,4.381 | |
so nihilism is the belief that nothing,458.94,4.979 | |
matters anyways uh which this kind of,460.86,6.0 | |
forms a vicious cycle where if you,463.919,4.741 | |
already have a nihilistic attitude and,466.86,3.66 | |
then you believe that between climate,468.66,4.74 | |
change and geopolitics and economics and,470.52,4.92 | |
AI that we're all doomed anyways you,473.4,3.419 | |
might as well give up while you're ahead,475.44,4.379 | |
and that is fatalism so the fatalism and,476.819,5.761 | |
nihilism play off of each other really,479.819,4.32 | |
powerfully,482.58,4.2 | |
um and it just leads to giving up,484.139,4.381 | |
and that is the hopelessness and,486.78,3.3 | |
inaction,488.52,4.26 | |
um so again I do want to sympathize with,490.08,4.5 | |
the doomers and say yes these are really,492.78,5.52 | |
difficult problems and uh the our,494.58,6.0 | |
success and survival as a species is not,498.3,4.38 | |
guaranteed it is not a foregone,500.58,4.44 | |
conclusion even for us optimists that we,502.68,3.859 | |
will come out in a better place,505.02,3.959 | |
generally speaking over the last century,506.539,4.6 | |
we have come out in a better place in,508.979,4.44 | |
the long run it can get pretty awful in,511.139,3.601 | |
the short term,513.419,2.941 | |
um and then but it's also not evenly,514.74,3.359 | |
distributed life gets better for some,516.36,3.359 | |
people worse for others,518.099,4.261 | |
so you know it is important to raise the,519.719,4.141 | |
alarm but,522.36,3.479 | |
you know we can't we can't just dwell,523.86,2.94 | |
right,525.839,3.421 | |
all right so Sympathy for the denier so,526.8,4.2 | |
the deniers and again I'm not trying to,529.26,3.48 | |
call out anyone by name I'm not trying,531.0,3.54 | |
to start Twitter beefs and YouTube beefs,532.74,4.38 | |
I'm just giving my perspective so,534.54,4.979 | |
Sympathy for the denier these are the,537.12,3.899 | |
people that have said yeah we've been,539.519,4.681 | |
promised AGI for like 60 years right I,541.019,4.38 | |
remember what was it there was a,544.2,2.759 | |
Consortium that was launched in like,545.399,3.721 | |
Stanford or something back in the 60s or,546.959,3.901 | |
70s and they're like oh yeah with a,549.12,3.12 | |
summer of work we should be able to,550.86,4.14 | |
figure out you know uh artificial,552.24,4.38 | |
intelligence and then here we are like,555.0,5.459 | |
40 or 60 years later and no,556.62,5.58 | |
um so yeah you know it's just like,560.459,3.661 | |
nuclear fusion right it's always 10,562.2,3.78 | |
years away or 20 years away,564.12,4.62 | |
so progress up to this point has been,565.98,7.14 | |
slow that is true uh there's a um on on,568.74,7.98 | |
the deniers side there is an emphasis,573.12,5.82 | |
more on like yeah it's you know AI is,576.72,3.96 | |
helpful and it could have some potential,578.94,4.74 | |
benefits but we shouldn't rely on this,580.68,5.46 | |
it's not a Magic Bullet right and that's,583.68,4.92 | |
that's always true like AI will change,586.14,4.68 | |
everything just the same way that steel,588.6,4.14 | |
and coal and steam power and internal,590.82,3.72 | |
combustion engines changed everything,592.74,3.599 | |
but it didn't solve The World's problems,594.54,4.919 | |
it it solved a bunch of problems created,596.339,6.721 | |
new problems and changed a lot of stuff,599.459,7.681 | |
um another uh uh benefit for for the,603.06,6.779 | |
deniers is that they're like hang on you,607.14,4.86 | |
know tap the brakes like let's not,609.839,4.56 | |
overreact let's not over uh like,612.0,4.8 | |
regulate or you know fear monger and I I,614.399,3.901 | |
do appreciate some of those comments,616.8,3.42 | |
actually it's like some of the deniers,618.3,3.24 | |
out there are like enough with the,620.22,3.84 | |
fear-mongering like I don't care,621.54,5.76 | |
um and then you know just we we have,624.06,5.279 | |
survived 100 of everything that has come,627.3,4.62 | |
our way so far and so like nothing has,629.339,4.981 | |
exploded yet the where's the fire right,631.92,5.34 | |
so there is some validity to the,634.32,4.86 | |
perspective of deniers out there which,637.26,3.36 | |
you know one of the things that they say,639.18,2.64 | |
is there's nothing to see here right,640.62,3.779 | |
nobody panic which you always need that,641.82,4.62 | |
kind of energy too like you in any,644.399,3.721 | |
society you want people raising the,646.44,3.54 | |
alarm and other people tapping the,648.12,4.08 | |
brakes we all have our purpose just like,649.98,5.34 | |
us optimists also have our role to play,652.2,5.34 | |
now there are some flaws with the,655.32,3.36 | |
denialism,657.54,5.28 | |
so one is from my perspective deniers,658.68,6.659 | |
seem to underestimate the potential,662.82,4.8 | |
risks especially when they say AGI is,665.339,4.44 | |
not possible hard takeoff isn't possible,667.62,5.159 | |
or these things are decades away,669.779,5.941 | |
um another possibility is just like not,672.779,4.8 | |
not fully thinking through like okay,675.72,3.42 | |
even if there's a five percent chance,677.579,3.961 | |
that AGI is happening within the next,679.14,4.8 | |
five years only a five percent chance,681.54,4.5 | |
what like still think through the cost,683.94,4.079 | |
of that right like look at Bayes theorem,686.04,3.479 | |
like okay there's a five percent chance,688.019,3.601 | |
that AGI is going to happen and if we,689.519,3.781 | |
don't do it right there is a very high,691.62,3.48 | |
likelihood that like we're all gonna die,693.3,5.7 | |
or end up in worse uh situation so in,695.1,6.06 | |
action the cost the potential cost of,699.0,4.74 | |
inaction or under action under reaction,701.16,5.28 | |
is still pretty high,703.74,6.42 | |
um another thing is is two uh or two two,706.44,5.459 | |
things are exponential growth and,710.16,4.38 | |
saltatory leaps so exponential growth,711.899,4.081 | |
which I provided some evidence for at,714.54,2.52 | |
the beginning of this video that's,715.98,3.24 | |
happening that is a fact,717.06,3.959 | |
and then actually let me go back to this,719.22,4.619 | |
so this here where you have this Gap up,721.019,5.461 | |
this is actually a mathematical evidence,723.839,5.221 | |
of a what's called a saltatory leap so a,726.48,4.5 | |
saltatory leap is when some breakthrough,729.06,4.98 | |
or some uh compounding returns of,730.98,5.94 | |
incremental progress result in sudden,734.04,4.739 | |
breakthroughs that you could not have,736.92,3.659 | |
predicted because if you just look at,738.779,3.721 | |
this trend line you'd predict like okay,740.579,3.601 | |
we wouldn't be here for another couple,742.5,4.32 | |
years but we're here now right so that,744.18,4.92 | |
that's a saltatory leap so you have to,746.82,5.28 | |
acknowledge that saltatory leaps not,749.1,4.979 | |
just do happen sometimes have happened,752.1,4.679 | |
recently and if it's happened recently,754.079,4.981 | |
it might happen again,756.779,5.281 | |
um the lack of urgency by saying eh it's,759.06,5.7 | |
decades away again you know you got to,762.06,3.959 | |
think through it like okay but what if,764.76,2.699 | |
it's not,766.019,5.041 | |
um and the the the taking a big step,767.459,5.521 | |
back the nothing to see here messaging,771.06,5.16 | |
might lead to boiled frog syndrome the,772.98,5.039 | |
the temperature is rising quickly this,776.22,3.54 | |
year I think a lot of us agree on that,778.019,3.841 | |
and so well you get used to it right,779.76,5.16 | |
okay it's warmer than it was but it's,781.86,5.88 | |
not hot yet thing is the time between it,784.92,4.44 | |
gets warm and it gets hot and it starts,787.74,5.039 | |
boiling that time could be shortening,789.36,6.719 | |
so the social impacts of these of of,792.779,6.18 | |
when thought leaders adopt more extreme,796.079,6.06 | |
uh stances such as doomerism or deny uh,798.959,6.601 | |
denialism is uh basically just a quick,802.139,6.421 | |
recap the doomers create nihilism and,805.56,6.779 | |
fatalism which discourages uh proactive,808.56,5.219 | |
Solutions because they say that there is,812.339,3.421 | |
no solution right that is one of the,813.779,3.601 | |
underpinning assumptions of doomers is,815.76,4.079 | |
that it's inevitable it's unavoidable,817.38,5.459 | |
there is no solution don't even try,819.839,4.56 | |
um and this promotes fear and anxiety,822.839,4.261 | |
right which yes fear and anxiety are,824.399,5.041 | |
evolutionarily there for a reason to,827.1,3.96 | |
motivate us to do something about a,829.44,3.18 | |
problem that we perceive,831.06,4.26 | |
but too much of a of of an important,832.62,4.32 | |
thing can still be bad,835.32,5.04 | |
finally our next the impact of denialism,836.94,5.22 | |
or denialism,840.36,3.599 | |
um is that there's a false sense of,842.16,3.66 | |
security right and we don't have to,843.959,4.021 | |
worry about it eh it's not coming for a,845.82,4.74 | |
long time right that's complacency and,847.98,5.88 | |
act and inaction which undermines some,850.56,4.44 | |
of the rest of us who are working and,853.86,3.0 | |
saying actually this might be something,855.0,3.3 | |
that we need to think a little bit,856.86,3.36 | |
further ahead on,858.3,4.02 | |
um because think about last year right,860.22,5.46 | |
how um AI images exploded onto the scene,862.32,5.94 | |
and nobody was ready for it right what,865.68,5.58 | |
if the next thing that happens is not,868.26,5.22 | |
just AI images or AI music or something,871.26,4.44 | |
like that but something a little bit,873.48,3.9 | |
more profound a little bit more,875.7,3.18 | |
significant that we just weren't ready,877.38,3.3 | |
for which means that the time to start,878.88,3.84 | |
preparing for those things that we know,880.68,4.44 | |
are coming eventually and we don't know,882.72,4.26 | |
when that Jack-in-the-Box is gonna pop,885.12,4.079 | |
the time to prepare is now,886.98,4.44 | |
so some of the some of the consequences,889.199,4.981 | |
that occur because of these the the,891.42,5.46 | |
messaging is one polarization of of,894.18,5.279 | |
public opinion some people are bending,896.88,4.62 | |
over backwards to say more jobs are,899.459,4.38 | |
coming let's not even think about you,901.5,5.519 | |
know AI based war or the control problem,903.839,5.461 | |
or anything like that uh what meanwhile,907.019,5.341 | |
others are like ah no we're you know,909.3,4.92 | |
lead the leaders around the world are,912.36,4.14 | |
not even addressing these risks they're,914.22,4.679 | |
just sitting on their on their hands so,916.5,4.86 | |
we're doomed right because if the adults,918.899,3.901 | |
in the room don't care or don't think,921.36,3.719 | |
it's a problem but all of the kids are,922.8,3.779 | |
like hey do you see that the house is on,925.079,3.301 | |
fire like maybe we should put that out,926.579,4.26 | |
first right that leads to more nihilism,928.38,5.699 | |
more fatalism a lot of polarization,930.839,5.641 | |
and then of course uh the Overton window,934.079,4.2 | |
is still too narrow so the Overton,936.48,3.719 | |
window is the concept of what is allowed,938.279,3.841 | |
to be talked about in political,940.199,5.521 | |
discourse so you know if you if you,942.12,5.159 | |
follow my channel every now and then,945.72,3.179 | |
I'll post links to videos like hey look,947.279,3.661 | |
you know the conversation is Shifting,948.899,5.661 | |
right just yesterday I posted a a a a a,950.94,6.48 | |
video from DW news which is in Germany,954.56,5.74 | |
where they try to address like hey,957.42,5.64 | |
there's actual anxiety about like,960.3,5.58 | |
everyone's jobs are going away right and,963.06,5.76 | |
they bent over real God I listened to it,965.88,4.92 | |
again they bent over backwards to try,968.82,4.319 | |
and say well yeah a lot of low-paid jobs,970.8,3.96 | |
are going away but there's a few high,973.139,3.901 | |
paid jobs coming in and it's like okay,974.76,4.199 | |
but still the point is is that most of,977.04,3.479 | |
the medium and low-paid jobs are going,978.959,3.781 | |
away and being replaced by a few,980.519,4.62 | |
high-paying jobs that's not the promise,982.74,4.62 | |
of like techno Revolution where AI,985.139,4.5 | |
creates a whole bunch of new jobs,987.36,5.4 | |
um and then I think it was Amazon uh or,989.639,5.101 | |
Facebook one of them just an announced,992.76,4.68 | |
even more layoffs and they explicitly,994.74,4.8 | |
said that the reason for the layoffs is,997.44,3.12 | |
that they're going to replace as many,999.54,3.599 | |
people with AI as possible I called it,1000.56,6.12 | |
I've been saying it so it's happening,1003.139,5.64 | |
um so the Overton window is Shifting now,1006.68,5.599 | |
okay why is this a big problem why like,1008.779,6.3 | |
fundamentally why is it that some people,1012.279,5.92 | |
are dimerous and denialists and what,1015.079,5.401 | |
what is left in the wash what is missing,1018.199,4.981 | |
from the conversation so one thing,1020.48,4.68 | |
that's missing is there is not a,1023.18,4.44 | |
coherent Global strategy,1025.16,5.22 | |
and so what I mean by that is everyone's,1027.62,5.4 | |
busy arguing you know in this little,1030.38,5.039 | |
domain or this little domain uh you know,1033.02,5.279 | |
about corporate governance or academic,1035.419,5.701 | |
Integrity or should we have a moratorium,1038.299,5.581 | |
right there's not really a global,1041.12,4.439 | |
strategy no one has even proposed,1043.88,3.24 | |
anything,1045.559,5.761 | |
um then on top of that is as I mentioned,1047.12,5.64 | |
just a moment ago,1051.32,3.66 | |
uh calling for moratoriums is not a,1052.76,3.539 | |
solution,1054.98,2.699 | |
um that's not even that's not even a,1056.299,2.941 | |
stopgap measure,1057.679,3.24 | |
um and so when all the thought leaders,1059.24,3.6 | |
in the world when none of them are,1060.919,3.841 | |
really offering Solutions of course,1062.84,4.44 | |
you're going to end up with a lot of uh,1064.76,4.98 | |
bickering and arguing and also a lot of,1067.28,5.54 | |
anxiety right we are humans and we love,1069.74,6.48 | |
love when there are adults in the room,1072.82,5.979 | |
that we trust to help make good,1076.22,4.5 | |
decisions and to make sure that we're,1078.799,4.681 | |
going to be okay right and right now on,1080.72,5.579 | |
the topic of AI there's nobody really,1083.48,4.74 | |
out there saying we're gonna be okay,1086.299,4.321 | |
I've got a plan,1088.22,5.52 | |
um and then uh on top of that Global,1090.62,4.799 | |
strategy is a comprehensive roadmap,1093.74,4.819 | |
right kind of the same thing,1095.419,3.14 | |
said a lot of this stuff but really what,1098.78,4.259 | |
we need is a is that Global,1100.34,4.5 | |
comprehensive roadmap and a,1103.039,4.441 | |
multi-layered approach to solving all,1104.84,4.92 | |
these problems at all these different uh,1107.48,3.66 | |
levels,1109.76,3.24 | |
so I've already alluded to some of these,1111.14,3.36 | |
things there's quite a bunch of stuff,1113.0,3.299 | |
that doesn't work right calling for,1114.5,4.14 | |
moratoriums just simply does not work,1116.299,4.62 | |
we'll get into more detail about why,1118.64,6.06 | |
moratoriums don't work and and and uh,1120.919,5.401 | |
and all the incentives against it in,1124.7,3.359 | |
just a moment another thing that doesn't,1126.32,4.14 | |
work is bombing data centers sorry that,1128.059,4.921 | |
is a really bone-headed suggestion,1130.46,5.099 | |
uh complaining on Twitter writing op-eds,1132.98,4.98 | |
writing mean comments on YouTube none of,1135.559,4.98 | |
these things are actually helpful and,1137.96,3.839 | |
another thing that's not helpful is,1140.539,2.88 | |
actually just trusting corporations or,1141.799,3.061 | |
the establishment to figure it out on,1143.419,2.281 | |
their own,1144.86,4.02 | |
we are all all humans Global,1145.7,5.96 | |
stakeholders in AI,1148.88,5.82 | |
so all these these the this list of,1151.66,4.36 | |
stuff that I've just have that that,1154.7,2.88 | |
doesn't work they're all molecule,1156.02,3.779 | |
reactions and molecule Solutions which,1157.58,3.3 | |
basically means that they will,1159.799,3.181 | |
inevitably lead to those lose-lose,1160.88,4.02 | |
outcomes that the doomers are are,1162.98,4.079 | |
warning us against right again I'm not,1164.9,4.2 | |
saying that the doomers are wrong if,1167.059,3.721 | |
things keep going as they are the,1169.1,4.5 | |
doomers are right I just I personally,1170.78,6.18 | |
don't ascribe to constantly yelling fire,1173.6,6.12 | |
and then claiming you know we're all,1176.96,4.44 | |
gonna die,1179.72,5.579 | |
okay so I outlined the big problems now,1181.4,5.82 | |
what,1185.299,4.26 | |
this video the entire purpose is to,1187.22,4.92 | |
introduce kind of the crowning,1189.559,4.5 | |
achievement so far of What Not Just I'm,1192.14,4.62 | |
working on but the the rapidly growing,1194.059,5.401 | |
community that I'm building,1196.76,4.74 | |
um uh what started around the years to,1199.46,5.04 | |
comparatives my research on alignment,1201.5,6.059 | |
for individual models and agents it has,1204.5,6.78 | |
quickly expanded so this gato framework,1207.559,7.62 | |
Global alignment taxonomy Omnibus is,1211.28,6.42 | |
that comprehensive strategy that I just,1215.179,4.74 | |
mentioned that is missing it is not just,1217.7,4.56 | |
for responsible AI development but is a,1219.919,5.461 | |
coherent road map that everyone on the,1222.26,5.58 | |
planet can participate in at various,1225.38,4.86 | |
levels whatever level makes the most,1227.84,3.66 | |
sense to you,1230.24,5.4 | |
this framework has seven layers on ways,1231.5,8.16 | |
to implement uh models AI systems and,1235.64,6.539 | |
also alignment uh alignment-based,1239.66,4.8 | |
regulations and we'll get into all the,1242.179,4.201 | |
layers in just a moment,1244.46,5.579 | |
uh but basically the the whole point of,1246.38,5.22 | |
this gato framework that we're working,1250.039,4.861 | |
on is that it will unite all,1251.6,5.28 | |
stakeholders give us a common framework,1254.9,5.04 | |
with which to have these discussions to,1256.88,5.28 | |
broaden the Overton window to open the,1259.94,4.38 | |
Overton window a little bit more so,1262.16,4.5 | |
whatever part of the spectrum you're on,1264.32,4.859 | |
whether you're saying eh it's not really,1266.66,4.379 | |
an issue yet or we're all going to die,1269.179,4.261 | |
or you don't care or you're an optimist,1271.039,4.981 | |
whatever this is a framework that we can,1273.44,4.56 | |
all participate in,1276.02,4.86 | |
um just in a decentralized distributed,1278.0,5.64 | |
and open source manner,1280.88,5.52 | |
so as promised here are the seven layers,1283.64,5.1 | |
of the gato framework and in the,1286.4,4.08 | |
community we started saying that it's,1288.74,3.6 | |
like a seven layer burrito so we use,1290.48,5.1 | |
like taco cat as our little Avatar so,1292.34,5.88 | |
layer one the lowest layer is model,1295.58,5.64 | |
alignment so model alignment has to do,1298.22,6.0 | |
with individual neural networks so that,1301.22,8.52 | |
means gpt2 gpt3 gpt4 Bert vicuna uh,1304.22,8.28 | |
stable LM all of these right large,1309.74,5.34 | |
language models are proliferating like,1312.5,4.559 | |
well I don't know just like locusts,1315.08,3.66 | |
whatever,1317.059,4.441 | |
it's happening right data sets are,1318.74,5.16 | |
growing models are growing they're all,1321.5,4.799 | |
coming out uh the cat's out of the bag,1323.9,5.04 | |
right language technology multimodal,1326.299,4.801 | |
technology it's all coming you can't,1328.94,3.719 | |
stop it,1331.1,3.48 | |
um so rather than stop it rather than,1332.659,4.801 | |
call for moratoriums what we're doing is,1334.58,5.459 | |
we're focusing on okay let's ride this,1337.46,5.06 | |
wave I all have already proposed,1340.039,4.741 | |
reinforcement learning with heuristic,1342.52,3.519 | |
imperatives which is different from,1344.78,2.34 | |
reinforcement learning with human,1346.039,3.661 | |
feedback because human feedback aligns,1347.12,5.28 | |
models to what humans want which what,1349.7,4.56 | |
humans want and what humans need often,1352.4,4.259 | |
very very different here is to,1354.26,3.84 | |
comparatives is not just what humans,1356.659,4.621 | |
want but what all life needs we're also,1358.1,4.74 | |
talking about data set curation and,1361.28,3.06 | |
inner alignment problems like Mesa,1362.84,3.12 | |
optimization,1364.34,6.06 | |
Layer Two is sorry autonomous systems so,1365.96,6.18 | |
these are cognitive architectures and,1370.4,3.38 | |
autonomous agents,1372.14,4.919 | |
this is this is recently exploded on the,1373.78,4.48 | |
scene with,1377.059,4.381 | |
um you know Jarvis and baby AGI and,1378.26,5.76 | |
agent GPT and all that fun stuff so you,1381.44,4.38 | |
guys know what that is and it's coming,1384.02,3.12 | |
and it's only going to get more,1385.82,3.359 | |
sophisticated we're on the ground floor,1387.14,4.62 | |
of autonomous systems this is year zero,1389.179,6.061 | |
year two three four five like you can't,1391.76,5.58 | |
on you cannot imagine how powerful,1395.24,4.679 | |
autonomous systems are going to be in,1397.34,5.4 | |
the coming years so at the at the the,1399.919,4.981 | |
low level the engine level right the,1402.74,4.2 | |
components under the hood that's the,1404.9,4.74 | |
models the autonomous systems are the,1406.94,4.979 | |
software architectures that use those,1409.64,3.96 | |
systems including memory systems and,1411.919,4.021 | |
apis and other stuff to create those,1413.6,5.52 | |
autonomous cognitive entities right,1415.94,6.66 | |
layer 3 is the decentralized network so,1419.12,5.16 | |
you might have seen some of my recent,1422.6,2.6 | |
videos where I've talked about,1424.28,3.3 | |
blockchain decentralized autonomous,1425.2,4.959 | |
organizations and also another component,1427.58,4.5 | |
of that is what's called a federation so,1430.159,4.741 | |
a federation is where you have either,1432.08,4.44 | |
independent nodes or independent,1434.9,3.48 | |
networks that can communicate and,1436.52,5.039 | |
collaborate through Federated systems so,1438.38,8.039 | |
these are the the network layer is how,1441.559,7.86 | |
do we create networked intelligent,1446.419,6.961 | |
entities that are also aligned and this,1449.419,5.701 | |
is a tough nut to crack we've had lots,1453.38,3.419 | |
of discussions in the group talking,1455.12,3.72 | |
about can you implement Heroes to,1456.799,5.941 | |
comparatives as a consensus mechanism at,1458.84,5.459 | |
what level do you process it do you,1462.74,5.64 | |
process it at every llm inference or do,1464.299,6.301 | |
you wait for the decisions how do you,1468.38,4.679 | |
make decisions around this kind of thing,1470.6,5.52 | |
excuse me real tough nut to crack number,1473.059,6.0 | |
four is where we jump from the technical,1476.12,5.34 | |
implementation and research to more of,1479.059,5.221 | |
the social political and economic uh,1481.46,4.5 | |
layer of the stack,1484.28,3.36 | |
and for all of you technologists out,1485.96,4.32 | |
there you can probably see,1487.64,5.34 | |
um my influence as a as a technologist,1490.28,5.04 | |
because this is it's not modeled on the,1492.98,5.04 | |
osm OSI model it's actually more closely,1495.32,4.92 | |
modeled on the defense and depth model,1498.02,5.58 | |
but it is a layered hierarchical stack,1500.24,6.6 | |
or onion of uh of Concepts so corporate,1503.6,4.199 | |
adoption,1506.84,2.339 | |
here's the thing,1507.799,3.841 | |
you cannot just tell a corporation you,1509.179,4.201 | |
know what stop with the AI we don't we,1511.64,3.72 | |
don't like where AI is going sure you,1513.38,4.74 | |
can try to with regulation uh but you,1515.36,4.5 | |
know like Italy tried to do that and,1518.12,3.9 | |
then they reverse course right,1519.86,4.26 | |
there's just way too much economic,1522.02,4.86 | |
incentive the bottom line you know that,1524.12,4.76 | |
is if you're if you're a corporation,1526.88,4.5 | |
shareholders and the bottom line that's,1528.88,4.659 | |
where the power is so rather than fight,1531.38,4.38 | |
that part of what this framework does is,1533.539,4.321 | |
say let's how let's figure out how we,1535.76,4.56 | |
can align those heuristic imperatives,1537.86,4.26 | |
reduce suffering increase prosperity and,1540.32,4.38 | |
increase understanding how can we align,1542.12,5.22 | |
those fundamental human needs the,1544.7,4.26 | |
fundamental needs of all living things,1547.34,3.66 | |
with corporate interest,1548.96,4.68 | |
and so one story that I like to share is,1551.0,4.5 | |
that I've had a few patreon uh,1553.64,3.659 | |
supporters reach out to me and they're,1555.5,3.36 | |
like hey I've got this autonomous system,1557.299,3.0 | |
that I'm working on but it's like it's,1558.86,3.96 | |
getting stuck or I need help or whatever,1560.299,5.041 | |
um or even without asking for my help uh,1562.82,4.2 | |
they said like hey I implemented the,1565.34,3.24 | |
heroes to comparatives in my autonomous,1567.02,3.0 | |
Business Systems and they work better,1568.58,5.699 | |
and I'm like thanks share so like if you,1570.02,5.82 | |
have any of those examples please post,1574.279,2.941 | |
them on Reddit on the heroes to,1575.84,3.48 | |
comparative subreddit because we need we,1577.22,3.839 | |
need more of those stories about how,1579.32,3.78 | |
aligned AI systems are actually good for,1581.059,4.381 | |
business it's that simple the bottom,1583.1,4.86 | |
line like I I will always say that,1585.44,4.56 | |
corporations are intrinsically amoral,1587.96,5.219 | |
however what I will say is that is that,1590.0,4.919 | |
their profit motive their primary,1593.179,3.12 | |
incentive structure which is to make,1594.919,4.62 | |
more money will benefit from adopting,1596.299,5.101 | |
heuristic comparative aligned systems,1599.539,4.081 | |
services and products which I also we,1601.4,3.54 | |
also have some members of the community,1603.62,3.12 | |
who are working on spinning this out,1604.94,3.18 | |
into either for-profit or not,1606.74,4.679 | |
not-for-profit services and of course,1608.12,4.799 | |
we're going to be publishing open source,1611.419,3.601 | |
data sets reference architectures that,1612.919,3.661 | |
sort of stuff to make it as easy as,1615.02,3.899 | |
possible for corporations all over the,1616.58,5.459 | |
world to adopt aligned AI,1618.919,5.041 | |
uh and we're going to work on convincing,1622.039,3.24 | |
them that this is the way to go too,1623.96,3.86 | |
number five National regulations,1625.279,5.041 | |
obviously as I just mentioned you know,1627.82,5.08 | |
corporations can or sorry Nations can do,1630.32,4.28 | |
some stuff like people pointed out like,1632.9,5.159 | |
gdpr uh European unions like you know,1634.6,6.459 | |
big package about like a uh data privacy,1638.059,5.161 | |
and stuff and certainly as an I.T,1641.059,3.261 | |
professional,1643.22,3.9 | |
people on the technology side are,1644.32,5.14 | |
terrified of gdpr right that's got some,1647.12,4.02 | |
teeth right you know right to be,1649.46,3.78 | |
forgotten where the data is owned and,1651.14,4.44 | |
housed and data governance okay great,1653.24,6.179 | |
that's all fine but see the thing is is,1655.58,5.82 | |
Nations have their own incentive,1659.419,4.14 | |
structure where it comes to Ai and what,1661.4,4.56 | |
I mean by that is uh the the national,1663.559,4.441 | |
interests of companies has to do with,1665.96,4.92 | |
their own GDP as a whole so this is a,1668.0,6.539 | |
big difference gdpr was about uh like,1670.88,5.64 | |
data privacy for Citizens and social,1674.539,5.341 | |
media it wasn't as directly tied to like,1676.52,6.899 | |
the national growth of their G ADP it,1679.88,5.52 | |
wasn't necessarily directly tied to,1683.419,3.661 | |
their geopolitical influence or their,1685.4,4.68 | |
military or their National Security,1687.08,4.92 | |
AI today though,1690.08,4.199 | |
is all of those things and more,1692.0,5.46 | |
because GDP growth geopolitical,1694.279,5.101 | |
influence National Security border,1697.46,4.38 | |
security whatever all of that has to do,1699.38,4.44 | |
those are the national interests that we,1701.84,4.62 | |
are going to be working on aligning AI,1703.82,6.0 | |
with and basically the long story short,1706.46,6.36 | |
is at a national level We're not gonna,1709.82,4.92 | |
we're not going to say hey Nations maybe,1712.82,4.02 | |
you shouldn't adopt AI maybe you should,1714.74,3.179 | |
slow it down maybe you should just,1716.84,3.12 | |
regulate it we're going to be actually,1717.919,3.481 | |
more I'm not going to say that like,1719.96,3.42 | |
we're accelerationists because like you,1721.4,3.779 | |
don't need to push the to go any faster,1723.38,3.48 | |
right I'm not advocating for,1725.179,3.721 | |
accelerationism I'm just observing that,1726.86,4.26 | |
acceleration is happening so how do we,1728.9,4.639 | |
steer it right and the idea is,1731.12,5.34 | |
encouraging Nations to adopt Heroes,1733.539,5.921 | |
comparative aligned uh models services,1736.46,5.699 | |
and systems because at every level of,1739.46,4.92 | |
government that will help steer the,1742.159,4.081 | |
nation in a better Direction and their,1744.38,4.5 | |
implementations will be safer more,1746.24,4.799 | |
reliable more trustworthy so on and so,1748.88,4.26 | |
forth and of course stability is good,1751.039,3.24 | |
for business it's good for the account,1753.14,3.32 | |
economy it's good for National Security,1754.279,5.52 | |
and all that other fun stuff next up is,1756.46,5.5 | |
number six uh layer six International,1759.799,4.681 | |
treaties so I actually did wasn't the,1761.96,4.02 | |
first one to come up with this idea but,1764.48,3.48 | |
basically we're going to be advocating,1765.98,4.679 | |
for an international Consortium like,1767.96,4.92 | |
CERN but for AI because here's the other,1770.659,4.02 | |
thing and a lot of people pointed this,1772.88,4.14 | |
out is that a lot of Nations,1774.679,5.22 | |
cannot even afford to participate in AI,1777.02,5.7 | |
research right AI research is carried,1779.899,4.561 | |
out largely by the wealthiest companies,1782.72,3.0 | |
on the planet and the wealthiest,1784.46,2.76 | |
countries on the planet,1785.72,3.42 | |
that's going to intrinsically leave a,1787.22,4.439 | |
lot of other nations uh behind in the,1789.14,4.98 | |
dust right and that's just not fair that,1791.659,4.201 | |
is a malarchy outcome where there's a,1794.12,3.419 | |
few wealthy bastions and the rest are,1795.86,4.98 | |
poor and they end up basically like,1797.539,6.26 | |
tossed on the on the rough Seas of an AI,1800.84,5.64 | |
saturated world so what we're going to,1803.799,5.081 | |
do is we're going to advocate for a,1806.48,5.64 | |
global international Consortium where uh,1808.88,6.899 | |
people people Nations pool resources,1812.12,5.88 | |
share their scientists share their,1815.779,5.101 | |
research share their data so that we can,1818.0,5.46 | |
all benefit equally across the whole,1820.88,6.539 | |
globe which that also uh has uh knock-on,1823.46,6.079 | |
benefits with in terms of alliances,1827.419,4.681 | |
economic benefits because you look at,1829.539,4.721 | |
like everyone's going to benefits from,1832.1,4.319 | |
from like CERN and the collaborations,1834.26,4.44 | |
between like NASA and Esa and and that,1836.419,3.721 | |
sort of stuff so International,1838.7,3.959 | |
scientific treaties generally one,1840.14,3.72 | |
they've got a pretty good track record,1842.659,3.24 | |
and two we've got a good model for them,1843.86,3.6 | |
so we're just basically saying let's,1845.899,4.861 | |
copy the success of NASA Issa of CERN,1847.46,6.24 | |
and let's do it for AI again that's not,1850.76,5.46 | |
like you know we're not this is nothing,1853.7,3.839 | |
Earth shattering right it's been done,1856.22,2.88 | |
before we're just saying maybe it is,1857.539,3.781 | |
time to do this with AI and finally,1859.1,5.1 | |
layer 7 of the gato framework is global,1861.32,5.76 | |
consensus so Global consensus has to do,1864.2,5.219 | |
with messaging,1867.08,5.459 | |
um uh working with universities academic,1869.419,5.401 | |
institutions uh industrial sectors,1872.539,5.701 | |
National sectors uh social media right,1874.82,6.3 | |
or all media really because if we can,1878.24,5.819 | |
build consensus in every sector in every,1881.12,5.059 | |
domain and at every level of society,1884.059,6.961 | |
then consensus around how to uh align AI,1886.179,7.0 | |
so that we all end up in a more utopian,1891.02,3.06 | |
state,1893.179,3.48 | |
the utopian attractor State rather than,1894.08,5.219 | |
dystopia or Extinction then we're going,1896.659,4.561 | |
to have a lot more energy right that,1899.299,3.961 | |
Overton window is going to be aligned in,1901.22,4.14 | |
the correct direction rather than you,1903.26,3.06 | |
know because right now the Overton,1905.36,3.24 | |
window is highly highly centered over,1906.32,4.979 | |
we're all going to die or nothing is,1908.6,6.179 | |
happening but really the truth is well,1911.299,5.461 | |
those are possibilities but the Overton,1914.779,4.081 | |
window needs to be broadened and that is,1916.76,4.139 | |
one of the key components of global,1918.86,3.78 | |
consensus,1920.899,4.38 | |
so I just threw a lot at you and this,1922.64,5.82 | |
all sounds really good Pie in the Sky uh,1925.279,5.461 | |
you know blah blah right there's,1928.46,3.66 | |
probably some skepticism so let's,1930.74,4.08 | |
address that this all started as a very,1932.12,5.039 | |
small Discord Community where I just,1934.82,3.9 | |
wanted to bring some people together to,1937.159,3.061 | |
help me do Heroes to comparatives,1938.72,4.92 | |
research and it quickly very quickly,1940.22,5.22 | |
scaled up,1943.64,4.74 | |
um we to as of this recording we have I,1945.44,5.04 | |
think right around just shy of 70 people,1948.38,5.22 | |
involved and more people coming all the,1950.48,4.62 | |
time we're actually having to work on,1953.6,4.14 | |
figuring out ways of automating the,1955.1,4.74 | |
recruiting the applications and the,1957.74,3.419 | |
onboarding which we haven't figured out,1959.84,4.079 | |
yet but we need to,1961.159,4.681 | |
um we're organizing teams and projects,1963.919,4.38 | |
around each layer of gato that I just uh,1965.84,4.02 | |
outlined and so you can see those here,1968.299,4.5 | |
on the right hand side so if you're a,1969.86,4.62 | |
reinforcement learning researcher or an,1972.799,4.26 | |
ml researcher or a data scientist we,1974.48,5.1 | |
need your help with layer one if you're,1977.059,4.081 | |
a software architect or a cloud,1979.58,4.079 | |
architect or someone or devops someone,1981.14,5.34 | |
who understands Automation and complex,1983.659,4.681 | |
systems we need your help in Layer Two,1986.48,4.319 | |
autonomous systems we've got a whole,1988.34,5.4 | |
bunch of blockchain endow people working,1990.799,5.1 | |
with us on layer three which is such a,1993.74,3.72 | |
cool topic because where this is like,1995.899,3.541 | |
super Cutting Edge also we're going to,1997.46,3.42 | |
eat our own dog food we're already,1999.44,5.88 | |
working on using Dows to help voting and,2000.88,6.0 | |
decision making and allocation of,2005.32,4.32 | |
resources within this project obviously,2006.88,4.98 | |
as I've said in many of my videos a lot,2009.64,4.139 | |
of blockchain and DOW technology is not,2011.86,4.319 | |
ready but we are going to eat our own,2013.779,4.38 | |
dog food and make sure that we are,2016.179,4.021 | |
testing these things so that they'll do,2018.159,3.36 | |
the things that we say that they need to,2020.2,2.819 | |
do right we're going to figure it out as,2021.519,4.441 | |
as we go number four corporate adoption,2023.019,5.821 | |
we have a few entrepreneurs and Business,2025.96,5.819 | |
Leaders we've got uh several ctOS in the,2028.84,4.8 | |
group we need more connections to,2031.779,3.9 | |
business and industry this means,2033.64,4.2 | |
conferences this means,2035.679,4.38 | |
um meetups this means,2037.84,4.8 | |
um people on boards right A lot of my,2040.059,4.261 | |
patreon supporters are business people,2042.64,3.659 | |
and so like I work with them directly,2044.32,4.039 | |
but we need more of that we need people,2046.299,5.401 | |
uh working to evangelize,2048.359,5.081 | |
um not just not just like saying hey,2051.7,3.659 | |
Corporation you should adopt your,2053.44,3.239 | |
heuristic imperatives and then leaving,2055.359,3.54 | |
it at that we have startups that we're,2056.679,4.801 | |
working with because the the companies,2058.899,4.801 | |
offering aligned Services don't exist,2061.48,4.439 | |
yet so we're helping incubate those,2063.7,3.36 | |
things and I don't mean from a financial,2065.919,3.18 | |
perspective but from a consultation,2067.06,5.039 | |
perspective and so because if the if hi,2069.099,5.401 | |
aligned Services goods and services,2072.099,5.401 | |
exist companies can adopt them but until,2074.5,5.399 | |
they exist they can't be adopted really,2077.5,5.159 | |
number five National regulation we're,2079.899,5.7 | |
just starting to have this conversation,2082.659,4.561 | |
um actually just a conversation I had,2085.599,3.901 | |
just a little while ago had to do with,2087.22,5.34 | |
uh talking with some of the uh policy,2089.5,6.3 | |
makers and lawyers and legislators that,2092.56,5.039 | |
are concerned about this kind of stuff,2095.8,3.24 | |
so for instance,2097.599,3.541 | |
um the vice president uh I don't know if,2099.04,4.2 | |
it's today but soon we'll be talking,2101.14,4.86 | |
with all of the big Tech Giants right so,2103.24,3.66 | |
we need to have more of those,2106.0,3.3 | |
conversations and we need to add some of,2106.9,4.92 | |
uh some of our perspective from the gato,2109.3,3.9 | |
framework,2111.82,4.32 | |
um into those National conversations but,2113.2,4.7 | |
not just from it not just from a,2116.14,3.719 | |
regulatory standpoint of the nation,2117.9,4.36 | |
looking down into the nation the,2119.859,4.081 | |
nation's looking up and out to the rest,2122.26,3.24 | |
of the world because as I mentioned,2123.94,4.32 | |
National Security that is a huge thing,2125.5,4.8 | |
GDP growth that is a big thing in,2128.26,4.5 | |
geopolitical influence AI is going to,2130.3,5.039 | |
affect all of these domains number six,2132.76,5.7 | |
uh the international treaty again we,2135.339,5.821 | |
need we need people that are connected,2138.46,5.159 | |
to the UN,2141.16,3.24 | |
um,2143.619,4.321 | |
uh maybe NATO I don't know oecd all,2144.4,6.48 | |
kinds of stuff uh UNESCO there's all,2147.94,4.5 | |
kinds of international organizations,2150.88,4.02 | |
that we would like to be connected with,2152.44,5.34 | |
and work with and talk to in order to,2154.9,5.4 | |
have these conversations and,2157.78,4.02 | |
by and large just make the right,2160.3,3.299 | |
connections so that these conversations,2161.8,4.14 | |
are happening and we can articulate the,2163.599,5.161 | |
gato framework and get it published and,2165.94,5.58 | |
then finally layer 7 Global consensus we,2168.76,4.14 | |
have writers we have graphic,2171.52,3.0 | |
communicators we've got editors we've,2172.9,3.66 | |
got audio Engineers,2174.52,4.76 | |
um we're working with uh people all over,2176.56,6.12 | |
even more influencers have excuse me,2179.28,6.22 | |
reached out to me so I'm uh I'm going to,2182.68,4.74 | |
be having conversations with them so,2185.5,6.18 | |
that we can all align on this consensus,2187.42,6.6 | |
and then here's our uh here's our our,2191.68,4.38 | |
mascot it's our own version of taco cat,2194.02,5.16 | |
so again you know gato cat and then you,2196.06,7.26 | |
know seven layered Taco you get the idea,2199.18,6.48 | |
um okay so you're probably glazing over,2203.32,4.14 | |
at this point but you've got the meat of,2205.66,3.3 | |
it so if you're really really super,2207.46,4.02 | |
interested in the layers let's take a,2208.96,4.56 | |
look at the layers of Gato in a little,2211.48,5.22 | |
bit bigger depth so number one of uh,2213.52,5.22 | |
layer one model alignment fine tuning,2216.7,3.72 | |
the very first experiment that I,2218.74,4.02 | |
published was on fine tuning large,2220.42,4.14 | |
language models so that they are aligned,2222.76,4.38 | |
number two reinforcement learning again,2224.56,4.44 | |
that is the goal is how do you create,2227.14,4.08 | |
the data sets in the systems and the,2229.0,6.0 | |
signals in order to have uh models that,2231.22,6.06 | |
not only are initially aligned to,2235.0,4.32 | |
heuristic imperatives and human needs,2237.28,4.2 | |
and the needs of all life but how do you,2239.32,3.72 | |
make sure that they get better at that,2241.48,3.3 | |
over time right that is the entire,2243.04,4.2 | |
purpose of heuristics heuristics uh,2244.78,3.839 | |
heuristic imperatives and reinforcement,2247.24,4.56 | |
learning basically the same thing,2248.619,5.281 | |
um at least here's the comparatives are,2251.8,3.84 | |
reinforcement learning on a specific,2253.9,2.939 | |
trajectory,2255.64,4.74 | |
model bias so there's uh there's a lot,2256.839,5.821 | |
of intrinsic bias in models there's been,2260.38,4.56 | |
uh some really interesting studies even,2262.66,4.02 | |
chat GPT with reinforcement learning,2264.94,3.72 | |
with human feedback is still pretty,2266.68,4.439 | |
sexist it's also pretty racist depending,2268.66,3.9 | |
on the kinds of prompts that you use,2271.119,3.661 | |
there's a lot of implicit bias then,2272.56,4.68 | |
there's also um Mesa optimization which,2274.78,4.74 | |
I'm not sure I'm not entirely sure that,2277.24,4.74 | |
Mesa optimization is a problem for,2279.52,5.16 | |
language models but it could be,2281.98,4.68 | |
um so we'll see but we need to be aware,2284.68,3.659 | |
of that and we need to study it and if,2286.66,3.54 | |
it is there we need to address it but,2288.339,4.321 | |
Mesa optimization is like a tiny,2290.2,4.44 | |
component of this whole framework,2292.66,5.699 | |
open source data sets so one of the,2294.64,6.3 | |
things that I mentioned is open source,2298.359,4.98 | |
open source Open Source by by creating,2300.94,4.26 | |
and Publishing open source data sets,2303.339,5.28 | |
that can uh one they're transparent but,2305.2,6.06 | |
two that can foster collaboration and,2308.619,4.261 | |
ultimately one of the things that I hope,2311.26,4.26 | |
to achieve is what I call axiomatic,2312.88,5.219 | |
alignment so axiomatic alignment is what,2315.52,4.8 | |
happens when through conversation,2318.099,4.98 | |
through experimentation through repeated,2320.32,5.64 | |
augmentation of Open Source data sets,2323.079,5.341 | |
practically every data set out there,2325.96,5.76 | |
that AI is trained on intrinsically has,2328.42,5.88 | |
some alignment baked into it and if,2331.72,5.1 | |
every data set or all or if enough data,2334.3,5.94 | |
sets are aligned then you can end up,2336.82,6.18 | |
with a a virtuous cycle or a positive,2340.24,5.46 | |
feedback loop where every subsequent,2343.0,5.339 | |
data set is also more and more aligned,2345.7,5.46 | |
so from a model perspective the,2348.339,5.041 | |
overarching goal is to arrive at a place,2351.16,4.56 | |
of axiomatic alignment,2353.38,4.14 | |
so this will require us to solve,2355.72,4.2 | |
problems around training model,2357.52,5.04 | |
architecture and then finally the data,2359.92,5.1 | |
ecosystem that we build and when I say,2362.56,4.559 | |
we I don't mean just those of us in gato,2365.02,4.74 | |
the gato framework project but everyone,2367.119,4.74 | |
everyone participating this whether,2369.76,4.2 | |
they're academic researchers corporate,2371.859,4.321 | |
government military so on and so forth,2373.96,5.28 | |
now Layer Two autonomous systems I,2376.18,3.96 | |
already talked a little bit about,2379.24,3.66 | |
cognitive architecture we don't need to,2380.14,4.92 | |
um you know beat the dead horse there,2382.9,4.439 | |
but one of the things that we want to,2385.06,4.62 | |
talk about and and publish is an open,2387.339,4.26 | |
source reference architecture that's,2389.68,3.72 | |
really the primary one of the primary,2391.599,3.961 | |
goals here is what are the components,2393.4,3.84 | |
what are the system components that you,2395.56,4.559 | |
need in order to have a fully aligned,2397.24,5.52 | |
and fully autonomous system so this,2400.119,3.901 | |
includes some of these things like,2402.76,3.78 | |
self-evaluation and stability we are,2404.02,5.339 | |
working on how do you how do you design,2406.54,4.92 | |
tasks how do you evaluate past,2409.359,3.661 | |
performance how do you automatically,2411.46,3.78 | |
label data and how do you create modular,2413.02,4.98 | |
design patterns that allow for anyone,2415.24,6.18 | |
and everyone to create their own fully,2418.0,5.82 | |
autonomous systems that are also aligned,2421.42,5.22 | |
to the heuristic imperatives and,2423.82,6.0 | |
therefore should be benevolent so by,2426.64,5.1 | |
getting by having the ultimate goal of,2429.82,3.84 | |
publishing these open source reference,2431.74,3.72 | |
architectures that'll make it really,2433.66,4.02 | |
easy for all corporations out there and,2435.46,3.78 | |
all private individuals and all,2437.68,4.86 | |
governments to adopt these uh these,2439.24,4.859 | |
patterns these software architecture,2442.54,4.62 | |
patterns which again just by providing,2444.099,4.621 | |
that answer and making it as easy as,2447.16,5.3 | |
possible will be a one component in,2448.72,5.82 | |
solving alignment and the control,2452.46,4.24 | |
problem globally,2454.54,5.52 | |
so decentralized networks this is not,2456.7,5.52 | |
just blockchain not just Dows but also,2460.06,5.34 | |
federations so keep that in mind,2462.22,4.28 | |
um,2465.4,4.08 | |
there's two primary components here one,2466.5,5.02 | |
first we just have to figure out how to,2469.48,3.72 | |
do these Technologies because by and,2471.52,3.72 | |
large these are still highly,2473.2,4.32 | |
experimental Technologies,2475.24,3.839 | |
um and I will be the first to admit that,2477.52,3.18 | |
maybe blockchain and DOW is not the,2479.079,4.381 | |
correct way but in principle some kind,2480.7,4.8 | |
of Federated system or decentralized,2483.46,4.68 | |
network is probably the way to go in,2485.5,4.74 | |
order to have some of these things such,2488.14,4.68 | |
as algorithmic consensus when we're in a,2490.24,3.96 | |
world where we have billions upon,2492.82,3.6 | |
billions of autonomous agents all,2494.2,4.62 | |
working on their own we need a way for,2496.42,4.8 | |
them to work with each other and with us,2498.82,6.06 | |
to come up with a consensus mechanisms,2501.22,7.02 | |
that will slow slow the roll basically,2504.88,5.04 | |
so there's a couple components that can,2508.24,3.96 | |
go into that one is trust and reputation,2509.92,6.179 | |
mechanisms so if you have you know some,2512.2,6.659 | |
arbitrary AI agent operating out on the,2516.099,4.26 | |
net on its own,2518.859,4.561 | |
if it is an untrusted agent then maybe,2520.359,4.74 | |
you don't want to give it resources or,2523.42,3.199 | |
you don't want to give it any Credence,2525.099,3.541 | |
that's what I mean by trust and,2526.619,4.0 | |
reputation mechanisms resource control,2528.64,4.8 | |
and allocation is another aspect of,2530.619,4.861 | |
using blockchain or Dao or Federated,2533.44,4.86 | |
Technologies which basically means if an,2535.48,5.16 | |
agent is behaving in a way that is not,2538.3,5.4 | |
aligned if the consensus of all agents,2540.64,4.74 | |
says hey that's a little bit destructive,2543.7,3.84 | |
maybe you shouldn't do it you revoke its,2545.38,4.979 | |
access to computational resources data,2547.54,5.22 | |
that sort of thing which can be a way to,2550.359,6.301 | |
allow and Empower uh autonomous agents,2552.76,5.7 | |
to police each other,2556.66,5.16 | |
and then finally incentivizing alignment,2558.46,5.1 | |
um so one of the things uh that people,2561.82,3.48 | |
are concerned about is instrumental,2563.56,3.84 | |
convergence so instrumental convergence,2565.3,6.72 | |
is the idea that um AI uh no matter what,2567.4,6.78 | |
goals you give it will be incentivized,2572.02,4.92 | |
to pursue basic similar things like,2574.18,5.22 | |
control of power more data that sort of,2576.94,5.639 | |
stuff but so that's that's based on its,2579.4,5.699 | |
in intrinsic motivations right an AI,2582.579,5.401 | |
needs electricity to run so therefore it,2585.099,4.081 | |
will always have some intrinsic,2587.98,3.9 | |
motivation to do that now through these,2589.18,4.32 | |
Network systems whether it's Federated,2591.88,3.9 | |
decentralized however the network,2593.5,5.579 | |
architecture is ultimately designed if,2595.78,6.18 | |
you incentivize their behavior to get,2599.079,4.801 | |
the behavior that you want so that they,2601.96,4.139 | |
can get what they want then that is the,2603.88,5.52 | |
way to go so for instance if you use,2606.099,5.581 | |
resource tokens or cryptocurrency or,2609.4,3.78 | |
whatever to say hey,2611.68,3.72 | |
everything that you do that is aligned,2613.18,4.8 | |
the the the the the the rest of the,2615.4,4.38 | |
network says we agree with that behavior,2617.98,3.66 | |
we agree with that decision we'll give,2619.78,3.6 | |
you a little bit more data or a little,2621.64,4.679 | |
bit more computational horsepower that,2623.38,5.1 | |
sort of stuff so you incentivize the,2626.319,4.5 | |
behavior that you want to see,2628.48,5.22 | |
number four corporate adoption so again,2630.819,4.981 | |
like I said for everyone that's talked,2633.7,3.6 | |
to me about it implementing Heroes to,2635.8,3.779 | |
comparatives ultimately just creates,2637.3,4.799 | |
better Solutions so if the best AI,2639.579,4.981 | |
services and products are aligned,2642.099,4.861 | |
the solution sells itself that's that,2644.56,4.08 | |
can that could literally be the end of,2646.96,4.2 | |
the conversation is that working with,2648.64,3.9 | |
corporations whether it's the tech,2651.16,3.84 | |
Giants providing these services or,2652.54,4.22 | |
everyone else consuming those services,2655.0,5.579 | |
to realize and develop those services so,2656.76,6.4 | |
that all AI services are intrinsically,2660.579,4.74 | |
aligned and of course open AI has done,2663.16,4.26 | |
their best you know they have,2665.319,3.661 | |
um they have their own internal research,2667.42,3.24 | |
one problem though is that they're not,2668.98,3.42 | |
sharing that research,2670.66,3.959 | |
um so their their work on alignment is a,2672.4,3.959 | |
total black box which means nobody else,2674.619,3.061 | |
can,2676.359,3.781 | |
um Can can duplicate it so we need an,2677.68,4.8 | |
open source way so that everyone can,2680.14,4.679 | |
duplicate alignment research and make,2682.48,4.44 | |
sure that all their apis all their AIS,2684.819,4.921 | |
are aligned and then corporations don't,2686.92,4.08 | |
even need to think about it right,2689.74,3.06 | |
because again corporations are like okay,2691.0,3.0 | |
whatever whatever is going to make us,2692.8,3.36 | |
the most money will do that and that's,2694.0,5.579 | |
yeah so if we if we create a corporate,2696.16,6.54 | |
ecosystem if an economic ecosystem in,2699.579,5.221 | |
which the best option Finance actually,2702.7,4.68 | |
is also the most aligned option problem,2704.8,4.74 | |
solved now that's a big if,2707.38,4.5 | |
there's a few other reasons though that,2709.54,4.86 | |
adopting aligned AI services and systems,2711.88,4.5 | |
would be good for corporations one,2714.4,3.48 | |
public relations,2716.38,3.479 | |
you know whatever whatever is popular in,2717.88,4.86 | |
Vogue so for instance like LGBT rights,2719.859,4.98 | |
super popular right now all the rage so,2722.74,4.079 | |
guess what a lot of corporations are,2724.839,4.081 | |
jumping on that bandwagon bandwagon,2726.819,4.081 | |
mentality is good as long as it aligns,2728.92,4.74 | |
on also something that is good employee,2730.9,5.1 | |
satisfaction now obviously I think that,2733.66,4.14 | |
employment conventional employment is,2736.0,2.76 | |
going to be going the way of the,2737.8,2.819 | |
dinosaurs by and large but for the,2738.76,3.78 | |
employees that are there it really feels,2740.619,4.681 | |
good to know that your company as part,2742.54,4.86 | |
of a higher mission to make the world,2745.3,4.62 | |
better for everyone so just gonna throw,2747.4,3.9 | |
that out there and then finally,2749.92,3.179 | |
stakeholder capitalism,2751.3,4.559 | |
stakeholder capitalism is an is a a,2753.099,6.061 | |
paradigm whereby it's not just you you,2755.859,5.281 | |
the corporation and your customers it's,2759.16,3.78 | |
everyone as a stakeholder so that's,2761.14,3.959 | |
employees customer suppliers environment,2762.94,5.04 | |
the rest of society so by adopting,2765.099,4.921 | |
aligned AI that can also bring,2767.98,3.9 | |
corporations in a line with stakeholder,2770.02,4.68 | |
capitalism as that idea continues to,2771.88,4.14 | |
develop,2774.7,3.899 | |
oh this is a long video uh number five,2776.02,4.5 | |
National regulations I already mentioned,2778.599,6.0 | |
GDP GDP growth obviously AI is a gonna,2780.52,7.26 | |
be a huge powerful economic engine for,2784.599,5.401 | |
the foreseeable future so we need to,2787.78,4.44 | |
make sure that as Nations you know try,2790.0,4.5 | |
to maximize their GDP which they are all,2792.22,3.96 | |
incentivized to do so that's fine I'm,2794.5,2.64 | |
not going to tell them that they're,2796.18,1.919 | |
wrong,2797.14,2.28 | |
um I don't think that it's necessarily,2798.099,3.24 | |
the best thing to optimize for but,2799.42,4.14 | |
that's how the world works right now you,2801.339,4.201 | |
can wish in one hand and you know you,2803.56,3.18 | |
know what you can do on the other hand,2805.54,3.48 | |
guess which one fills up,2806.74,4.02 | |
um National Security so this is the,2809.02,4.02 | |
biggest thing right the US's uh chips,2810.76,4.319 | |
act where we you know did the the AI,2813.04,4.319 | |
chips embargo against China right that's,2815.079,5.341 | |
an example of the geopolitical game of,2817.359,4.381 | |
chess that is going to be playing out,2820.42,4.86 | |
for the foreseeable future around Ai and,2821.74,6.96 | |
adversarial uses of AI so by working,2825.28,6.539 | |
with nations in in line in alignment,2828.7,5.639 | |
with their their national interests we,2831.819,5.401 | |
can also work with them to adopt more,2834.339,5.421 | |
aligned AI solicit Solutions and systems,2837.22,5.76 | |
Democratic institutions so uh voter,2839.76,7.059 | |
rights electric transparency Judicial,2842.98,7.08 | |
Systems AI is going to impact every,2846.819,7.321 | |
element every aspect of uh liberal,2850.06,6.6 | |
Democratic societies including the,2854.14,4.38 | |
agencies that the that those governments,2856.66,4.14 | |
run on so by working with them to say,2858.52,5.04 | |
here's how you can Implement AI to both,2860.8,5.46 | |
save money and be a better Society to,2863.56,4.259 | |
strengthen your Democratic institutions,2866.26,3.839 | |
that will benefit everyone,2867.819,4.8 | |
geopolitical influence ditto there's,2870.099,5.881 | |
going to be things about trade for,2872.619,6.24 | |
instance alliances all of those things,2875.98,4.98 | |
are going to be impacted by AI which we,2878.859,4.621 | |
need to study and we need to become the,2880.96,4.26 | |
world experts on so that we can advise,2883.48,4.02 | |
and consult properly and then finally,2885.22,4.139 | |
sustainability which comes down to,2887.5,3.78 | |
environmental challenges in the grand,2889.359,4.021 | |
scheme of things I think that if we,2891.28,4.38 | |
solve these other problems then by,2893.38,4.68 | |
virtue of solving those problems around,2895.66,5.179 | |
consensus we'll probably also figure out,2898.06,5.1 | |
environmental control,2900.839,5.141 | |
layer 6 International treaty I already,2903.16,5.159 | |
mentioned um basically CERN but for AI,2905.98,4.56 | |
so just a really quick recap of the,2908.319,4.621 | |
benefits one membership and governance,2910.54,5.16 | |
where all uh all nations are,2912.94,4.98 | |
stakeholders and so they can join and,2915.7,3.48 | |
make decisions collectively,2917.92,4.38 | |
collaborative research again same exact,2919.18,4.2 | |
thing that we already see with CERN,2922.3,3.5 | |
shared resources and infrastructure,2923.38,4.5 | |
Education and Training so this is,2925.8,3.94 | |
another thing is there's probably going,2927.88,4.199 | |
to be a shortfall of qualified Ai,2929.74,4.619 | |
blockchain and and cognitive Architects,2932.079,4.921 | |
for a while so by working together to,2934.359,4.201 | |
make sure that we train up the people,2937.0,3.96 | |
that we need to solve this problem that,2938.56,3.779 | |
is something that International,2940.96,4.08 | |
cooperation could do a lot for open,2942.339,4.5 | |
science and knowledge sharing again that,2945.04,4.02 | |
has been well established with,2946.839,3.541 | |
um with some of these existing things,2949.06,4.98 | |
International cooperation ditto huh see,2950.38,6.479 | |
above statements and then finally uh,2954.04,4.74 | |
Global consensus I already mentioned uh,2956.859,3.48 | |
pretty much all of these,2958.78,3.42 | |
um academic institutions we've got we've,2960.339,3.78 | |
already got a few professors and,2962.2,4.08 | |
students in the group so we've got a few,2964.119,3.901 | |
lines in you know we've got feelers and,2966.28,4.92 | |
fingers into um into the academic,2968.02,4.5 | |
establishment,2971.2,2.879 | |
um I've actually personally had probably,2972.52,3.18 | |
a hundred different students reach out,2974.079,2.76 | |
to me,2975.7,4.02 | |
um either on patreon Discord or LinkedIn,2976.839,5.101 | |
or Twitter and every time they ask me,2979.72,3.599 | |
like Dave what should I what should I do,2981.94,4.08 | |
and I'm like AI man like it's going that,2983.319,5.04 | |
way if you care about the future like,2986.02,4.2 | |
take a look at like some of the stuff,2988.359,3.96 | |
that I've written and advocate for yours,2990.22,3.66 | |
to comparatives research and they're,2992.319,3.841 | |
like cool that's what I'll do,2993.88,5.64 | |
um so you know because education is the,2996.16,6.9 | |
future as as as many criticisms as I,2999.52,5.22 | |
have of particularly American,3003.06,4.14 | |
institutions universities are here,3004.74,4.079 | |
they're here to stay they're important,3007.2,4.58 | |
stakeholders in this entire conversation,3008.819,5.941 | |
media engagement so this is this has to,3011.78,5.44 | |
do with mainstream media this has to do,3014.76,5.16 | |
with social media uh all of the above,3017.22,4.2 | |
one of the things that we're working on,3019.92,3.24 | |
is we're working on producing materials,3021.42,4.08 | |
to make all this as accessible and,3023.16,4.74 | |
shareable as possible so we're creating,3025.5,3.96 | |
graphical slide decks we're creating,3027.9,4.62 | |
educational materials I've got my videos,3029.46,4.74 | |
um that sort of stuff because the more,3032.52,3.54 | |
information that we get out there the,3034.2,4.2 | |
easier it is to consume the more widely,3036.06,4.08 | |
it's shared the better off we're all,3038.4,3.02 | |
going to be,3040.14,4.26 | |
next up is industry Partnerships again,3041.42,5.199 | |
as I mentioned just a minute ago one of,3044.4,3.419 | |
the things that we're that we're working,3046.619,3.48 | |
on is publishing those open source,3047.819,5.28 | |
standards advocating for startups and,3050.099,4.861 | |
other companies to build and adopt,3053.099,5.161 | |
aligned AI services and Pro products and,3054.96,5.46 | |
just by working with them to say hey we,3058.26,4.559 | |
recognize that your bottom line is the,3060.42,4.139 | |
most important thing to companies let's,3062.819,3.78 | |
make sure that that that that you,3064.559,4.681 | |
implement and deploy these things in a,3066.599,3.72 | |
way that doesn't have unintended,3069.24,3.599 | |
negative consequences and then finally,3070.319,5.401 | |
policy advocacy so this has to do with,3072.839,5.041 | |
back going back every layer which is,3075.72,5.78 | |
working with legislators lawyers,3077.88,5.939 | |
and other groups you know whether it's,3081.5,4.9 | |
think tanks whoever in order to better,3083.819,4.981 | |
understand this stuff so an example of,3086.4,4.26 | |
this is I've got a few meetings coming,3088.8,4.62 | |
up later in May where I'll be meeting,3090.66,5.52 | |
with people to help bring them up to,3093.42,4.919 | |
speed with some of these ideas and help,3096.18,4.26 | |
guide them as to like okay this is,3098.339,4.02 | |
what's happening this is how it works,3100.44,4.08 | |
and here's a here's an approach that we,3102.359,4.681 | |
can take to make sure that it doesn't uh,3104.52,5.22 | |
go uh belly side up,3107.04,4.62 | |
now,3109.74,4.02 | |
um we all have a good story,3111.66,4.98 | |
for understanding this so in Avengers,3113.76,5.099 | |
which I talk about this probably more,3116.64,4.5 | |
than I should near the very end when,3118.859,4.861 | |
Thanos said I am inevitable,3121.14,6.54 | |
that is a fictional representation of,3123.72,7.92 | |
Malik so the the idea is that Thanos was,3127.68,6.24 | |
an Unstoppable destructive force that,3131.64,4.14 | |
nobody wanted he wanted an outcome that,3133.92,4.56 | |
nobody wanted but it seemed inevitable,3135.78,5.22 | |
and he even said I am inevitable,3138.48,3.839 | |
the snap,3141.0,3.78 | |
the idea that there could be a moment in,3142.319,5.641 | |
time that everything goes sideways,3144.78,5.94 | |
everything goes wrong that is what,3147.96,5.399 | |
Singularity or hard takeoff or whatever,3150.72,5.58 | |
could represent the Infinity Stones,3153.359,5.641 | |
think of those as the power of AI as as,3156.3,4.68 | |
we get more and more AI capabilities,3159.0,4.8 | |
it's like we're loading up our Gauntlet,3160.98,4.68 | |
um the sacrifice that various people,3163.8,3.84 | |
make like Tony Stark we have a lot of,3165.66,4.02 | |
hard choices to make including just the,3167.64,3.78 | |
investment that people like me and,3169.68,3.84 | |
everyone in the community are making in,3171.42,5.1 | |
terms of time and energy and the risks,3173.52,5.4 | |
that we're taking in order to say hey we,3176.52,4.26 | |
see this problem coming and we're going,3178.92,3.72 | |
to try and do something about it,3180.78,4.74 | |
in the story of undoing the snap the,3182.64,5.1 | |
idea is that there is always hope that,3185.52,4.68 | |
with the right people the right team and,3187.74,5.22 | |
the right effort you can either avert,3190.2,6.119 | |
disaster or undo disaster now obviously,3192.96,4.74 | |
a lot of doomers say we don't get a,3196.319,3.721 | |
do-over we don't get we we get one shot,3197.7,3.84 | |
at this I don't know whether or not,3200.04,4.68 | |
that's true but the idea is that we are,3201.54,6.6 | |
barreling towards our end game right we,3204.72,5.82 | |
have we must have the right people the,3208.14,3.9 | |
right team,3210.54,4.98 | |
um in a concerted Global effort in order,3212.04,6.18 | |
to solve this problem safely and not,3215.52,4.2 | |
just not just solve it like,3218.22,3.06 | |
satisfactorily,3219.72,3.48 | |
because again there's many possible,3221.28,3.6 | |
outcomes I don't want a dystopian,3223.2,4.02 | |
outcome any more than I want Extinction,3224.88,6.179 | |
or collapse there's one possible outcome,3227.22,6.839 | |
that is win-win that is Utopia and we,3231.059,4.621 | |
got to thread that needle and we'll be,3234.059,3.361 | |
working as hard as we can to make sure,3235.68,4.679 | |
that that happens so this is The,3237.42,6.24 | |
Avengers Assemble moment if you want to,3240.359,6.361 | |
join this effort the link to apply is in,3243.66,5.34 | |
the description of this video if you,3246.72,4.08 | |
don't want to participate directly you,3249.0,3.96 | |
can also support me on patreon I'm also,3250.8,3.66 | |
happy to support you if you support me,3252.96,4.32 | |
on patreon I have a private patreon,3254.46,5.159 | |
Discord where I answer questions we,3257.28,4.799 | |
actually just started having office,3259.619,5.881 | |
hours Town Hall Days where all my,3262.079,4.74 | |
patreon supporters can interact with,3265.5,4.14 | |
each other and with me in real time if,3266.819,4.141 | |
you've been laid off and you've got,3269.64,3.24 | |
technical skills or political skills or,3270.96,3.84 | |
communication skills or whatever,3272.88,3.9 | |
maybe now's the time for you to join the,3274.8,4.74 | |
effort if you're scared one of the one,3276.78,4.38 | |
of the most powerful things that people,3279.54,4.14 | |
have told me in in in the heuristics,3281.16,4.56 | |
imperatives Discord is that for the,3283.68,4.56 | |
first time in since forever they feel,3285.72,5.099 | |
empowered to make a difference in the,3288.24,4.26 | |
outcome that we're heading towards and,3290.819,3.421 | |
if you're optimistic like me we also,3292.5,6.0 | |
need that so Avengers assembled thank,3294.24,6.44 | |
you,3298.5,2.18 | |