text,start,duration hey everybody David Shapiro here with a,1.38,5.479 video today's video is going to be about,4.38,7.08 doomerism uh denialism uh an alternative,6.859,8.141 perspective optimism as well as a very,11.46,5.22 comprehensive framework that I'm putting,15.0,4.619 together with a lot of folks so let's go,16.68,6.42 ahead and take a look at some ideas and,19.619,4.681 some data,23.1,6.419 so we are all talking about exponential,24.3,7.86 growth if you look at comments across,29.519,5.88 the internet and even mainstream news,32.16,7.079 today talking about the rise of AI one,35.399,6.301 thing that happens is that a lot of,39.239,5.401 people tend to think in terms of linear,41.7,6.3 progress you say oh well 10 years ago we,44.64,5.759 were here and now we're you know now,48.0,4.68 we're there and so 10 years from now,50.399,3.84 we'll basically continue with the same,52.68,3.66 amount of progress that's not actually,54.239,5.701 true when you shorten that time Horizon,56.34,5.76 to say oh well we've made a lot of,59.94,4.14 progress in the last few months maybe,62.1,4.32 that's the new rate of progress that's,64.08,4.74 still not actually true,66.42,4.8 with exponential growth which is what we,68.82,4.86 are seeing right now the actual uh,71.22,5.7 correct assumption to make is that uh is,73.68,5.7 that you know the X amount of time from,76.92,5.64 from now will actually have continued to,79.38,4.44 accelerate,82.56,4.62 now this is a nice lovely handmade graph,83.82,6.659 that is shown in perfect clear data but,87.18,5.34 let me show you some actual some real,90.479,5.1 data about parameter counts in neural,92.52,4.68 networks,95.579,4.22 so here you can see it growing,97.2,5.76 exponentially and then the exponential,99.799,5.32 curve accelerates and it starts growing,102.96,3.78 logarithmically,105.119,5.221 so we are at the knee of the curve,106.74,6.059 already so the knee of the curve is this,110.34,4.02 part right here,112.799,4.68 where the acceleration really starts to,114.36,5.039 take off but the thing is is when you're,117.479,3.541 in the middle of it it's kind of like,119.399,3.241 boiled frog syndrome which we'll talk,121.02,3.9 about a little bit more in just a minute,122.64,6.599 so with this data in mind let's jump,124.92,6.42 into the rest of the video,129.239,5.341 so I mentioned doomerism and denialism,131.34,5.64 and then finally optimism these are kind,134.58,5.34 of the three main categories that people,136.98,5.82 by and large fall into there's also,139.92,5.52 people that are apathetic uh which I,142.8,4.079 didn't include that just because it's a,145.44,4.2 waste of screen space but so doomerism,146.879,6.961 is uh the is the the belief that uh,149.64,8.04 decline collapse Calamity is inevitable,153.84,6.3 that we are going to end up in some sort,157.68,4.62 of Extinction scenario or dystopian,160.14,3.84 outcome and that there's not really,162.3,4.5 anything that we can do to change it so,163.98,4.2 this is why there's been lots of,166.8,3.6 comments around like Malik which the,168.18,4.08 idea of Malik will get into that a,170.4,3.96 little bit as well,172.26,5.4 um so then there's denialism so the,174.36,5.22 denialists basically say there's nothing,177.66,5.46 to see here uh AGI is not even possible,179.58,6.84 or it's still decades away hard takeoff,183.12,6.54 is not possible or it's decades away and,186.42,5.7 uh then finally optimism techno,189.66,5.159 optimists is people like myself who are,192.12,4.8 just like yeah like we can do this these,194.819,4.381 problems are all solvable and it will,196.92,6.42 ultimately end up in the better so what,199.2,6.78 I want to say is is that this is I'm not,203.34,4.679 talking about individuals don't take it,205.98,3.539 personally if you identify with these,208.019,3.061 what I'm talking about here is thought,209.519,4.201 leaders uh people like content creators,211.08,5.34 like myself leading scientists people on,213.72,6.48 Twitter uh basically famous people or,216.42,6.78 respected people in The Establishment in,220.2,7.02 the industry who take these mindsets uh,223.2,6.959 so again not not calling on any,227.22,6.0 particular commenter or fan or people on,230.159,5.881 Reddit or Twitter this is talking about,233.22,4.98 basically like people at my level or,236.04,3.419 above,238.2,2.459 um,239.459,4.14 and also this is like it obviously,240.659,4.621 doesn't fall into this symbol of,243.599,3.241 categories I'm just kind of talking,245.28,3.539 about the the kind of extreme ends most,246.84,3.78 people fall somewhere in the middle like,248.819,3.06 if you were to draw this out on a,250.62,2.88 triangle most people are somewhere in,251.879,3.06 the middle there's a few people at the,253.5,3.54 extreme points I'm an extreme optimist,254.939,3.14 so,257.04,3.84 in in the yellow corner is the extreme,258.079,4.481 optimists in the red and green Corners,260.88,4.259 are the dumerous and denialists,262.56,5.88 um okay so but as promised by the,265.139,4.681 opening title I want to take a,268.44,4.08 sympathetic look at these other uh other,269.82,4.92 uh uh dispositions,272.52,6.06 so Sympathy for the Doomer one it is,274.74,5.76 good to acknowledge the existential,278.58,4.98 risks of AI this has been true of all,280.5,5.22 new technologies whether it's Medical,283.56,4.5 Technology nuclear technology pretty,285.72,4.68 much every new technology today carries,288.06,4.38 with it some level of existential risk,290.4,4.859 right you know the whether it's the,292.44,5.1 ability to do Gene engineering or,295.259,5.16 engineer new uh strains of flu or,297.54,5.099 coronavirus or whatever there's always,300.419,4.5 risks,302.639,5.041 um the doomers understand the potential,304.919,5.821 risk of uncontrolled AGI right the sky,307.68,5.4 is the limit right as people are,310.74,4.08 learning more about what AI is capable,313.08,4.44 of the idea of an Extinction scenario,314.82,5.58 like Skynet is actually not entirely,317.52,4.92 impossible and when you look at the fact,320.4,4.2 that Congress right now is working on,322.44,4.5 passing legislation so that AI will,324.6,4.439 never have uh control over nuclear,326.94,4.02 weapons like they're taking it seriously,329.039,4.38 too right so like there's something here,330.96,5.76 uh there's it's not nothing,333.419,5.101 um so then there's also the recognition,336.72,4.8 for safeguards and regulations and then,338.52,4.56 finally when you just look at the,341.52,3.72 current trends,343.08,3.839 um like stagnant wages and wealth,345.24,3.42 inequality and other evidence of the,346.919,5.701 Malik problem like it doesn't take a you,348.66,6.599 know a a great leap of faith or logic to,352.62,4.139 say what if these Trends continue and,355.259,3.541 get worse which there's no evidence of,356.759,4.201 some of these Trends reversing,358.8,3.66 um then it's like okay well then we are,360.96,3.9 all going to end up in a cyberpunk Hell,362.46,4.98 and then finally these problems are all,364.86,4.32 very large and complex they are global,367.44,4.68 scale problems so what I want to say is,369.18,4.799 I want to acknowledge that these are the,372.12,3.96 primary as far as I can tell the primary,373.979,4.201 concerns of doomers,376.08,5.1 um and uh like there is some legitimacy,378.18,4.44 to this position I'm not saying oh,381.18,3.18 doomers are just flat out wrong you know,382.62,3.359 to ignore them like no these are real,384.36,3.6 things I need to acknowledge that but,385.979,4.201 what I'll get to is like why I'm still,387.96,5.239 optimistic despite all this,390.18,6.959 now to play Devil's Advocate there are,393.199,6.461 some flaws with tumorism which is people,397.139,5.06 that just stick in this corner,399.66,4.979 one is over emphasis on worst case,402.199,3.821 scenarios,404.639,3.301 yes we can think about worst case,406.02,4.14 scenarios but it does not help for us to,407.94,4.08 dwell on worst case scenarios and only,410.16,3.9 worst case scenarios we need to think,412.02,3.48 about the entire spectrum of,414.06,2.699 possibilities,415.5,3.18 another thing that is that is common,416.759,4.261 with some doomers is that they're very,418.68,4.38 dogmatic in their thinking they have,421.02,3.959 come to believe for their own reasons,423.06,3.24 with their own logic and their own,424.979,3.261 research and minds and whatever else,426.3,4.679 that catastrophe is a foregone,428.24,5.079 conclusion they think that it is totally,430.979,5.101 inevitable which results in dogmatic and,433.319,4.1 rigid thinking,436.08,4.92 this mentality discourages Innovation,437.419,5.441 and collaboration they're like ah,441.0,4.38 we're doomed who cares give up just,442.86,4.08 throw your hands up and just let it,445.38,3.24 happen,446.94,3.18 um which creates a distraction from,448.62,4.139 finding real solutions and the ultimate,450.12,4.68 result of this is from an emotional and,452.759,3.72 psychological perspective is that it,454.8,4.14 leads to a sense of nihilism or fatalism,456.479,4.381 so nihilism is the belief that nothing,458.94,4.979 matters anyways uh which this kind of,460.86,6.0 forms a vicious cycle where if you,463.919,4.741 already have a nihilistic attitude and,466.86,3.66 then you believe that between climate,468.66,4.74 change and geopolitics and economics and,470.52,4.92 AI that we're all doomed anyways you,473.4,3.419 might as well give up while you're ahead,475.44,4.379 and that is fatalism so the fatalism and,476.819,5.761 nihilism play off of each other really,479.819,4.32 powerfully,482.58,4.2 um and it just leads to giving up,484.139,4.381 and that is the hopelessness and,486.78,3.3 inaction,488.52,4.26 um so again I do want to sympathize with,490.08,4.5 the doomers and say yes these are really,492.78,5.52 difficult problems and uh the our,494.58,6.0 success and survival as a species is not,498.3,4.38 guaranteed it is not a foregone,500.58,4.44 conclusion even for us optimists that we,502.68,3.859 will come out in a better place,505.02,3.959 generally speaking over the last century,506.539,4.6 we have come out in a better place in,508.979,4.44 the long run it can get pretty awful in,511.139,3.601 the short term,513.419,2.941 um and then but it's also not evenly,514.74,3.359 distributed life gets better for some,516.36,3.359 people worse for others,518.099,4.261 so you know it is important to raise the,519.719,4.141 alarm but,522.36,3.479 you know we can't we can't just dwell,523.86,2.94 right,525.839,3.421 all right so Sympathy for the denier so,526.8,4.2 the deniers and again I'm not trying to,529.26,3.48 call out anyone by name I'm not trying,531.0,3.54 to start Twitter beefs and YouTube beefs,532.74,4.38 I'm just giving my perspective so,534.54,4.979 Sympathy for the denier these are the,537.12,3.899 people that have said yeah we've been,539.519,4.681 promised AGI for like 60 years right I,541.019,4.38 remember what was it there was a,544.2,2.759 Consortium that was launched in like,545.399,3.721 Stanford or something back in the 60s or,546.959,3.901 70s and they're like oh yeah with a,549.12,3.12 summer of work we should be able to,550.86,4.14 figure out you know uh artificial,552.24,4.38 intelligence and then here we are like,555.0,5.459 40 or 60 years later and no,556.62,5.58 um so yeah you know it's just like,560.459,3.661 nuclear fusion right it's always 10,562.2,3.78 years away or 20 years away,564.12,4.62 so progress up to this point has been,565.98,7.14 slow that is true uh there's a um on on,568.74,7.98 the deniers side there is an emphasis,573.12,5.82 more on like yeah it's you know AI is,576.72,3.96 helpful and it could have some potential,578.94,4.74 benefits but we shouldn't rely on this,580.68,5.46 it's not a Magic Bullet right and that's,583.68,4.92 that's always true like AI will change,586.14,4.68 everything just the same way that steel,588.6,4.14 and coal and steam power and internal,590.82,3.72 combustion engines changed everything,592.74,3.599 but it didn't solve The World's problems,594.54,4.919 it it solved a bunch of problems created,596.339,6.721 new problems and changed a lot of stuff,599.459,7.681 um another uh uh benefit for for the,603.06,6.779 deniers is that they're like hang on you,607.14,4.86 know tap the brakes like let's not,609.839,4.56 overreact let's not over uh like,612.0,4.8 regulate or you know fear monger and I I,614.399,3.901 do appreciate some of those comments,616.8,3.42 actually it's like some of the deniers,618.3,3.24 out there are like enough with the,620.22,3.84 fear-mongering like I don't care,621.54,5.76 um and then you know just we we have,624.06,5.279 survived 100 of everything that has come,627.3,4.62 our way so far and so like nothing has,629.339,4.981 exploded yet the where's the fire right,631.92,5.34 so there is some validity to the,634.32,4.86 perspective of deniers out there which,637.26,3.36 you know one of the things that they say,639.18,2.64 is there's nothing to see here right,640.62,3.779 nobody panic which you always need that,641.82,4.62 kind of energy too like you in any,644.399,3.721 society you want people raising the,646.44,3.54 alarm and other people tapping the,648.12,4.08 brakes we all have our purpose just like,649.98,5.34 us optimists also have our role to play,652.2,5.34 now there are some flaws with the,655.32,3.36 denialism,657.54,5.28 so one is from my perspective deniers,658.68,6.659 seem to underestimate the potential,662.82,4.8 risks especially when they say AGI is,665.339,4.44 not possible hard takeoff isn't possible,667.62,5.159 or these things are decades away,669.779,5.941 um another possibility is just like not,672.779,4.8 not fully thinking through like okay,675.72,3.42 even if there's a five percent chance,677.579,3.961 that AGI is happening within the next,679.14,4.8 five years only a five percent chance,681.54,4.5 what like still think through the cost,683.94,4.079 of that right like look at Bayes theorem,686.04,3.479 like okay there's a five percent chance,688.019,3.601 that AGI is going to happen and if we,689.519,3.781 don't do it right there is a very high,691.62,3.48 likelihood that like we're all gonna die,693.3,5.7 or end up in worse uh situation so in,695.1,6.06 action the cost the potential cost of,699.0,4.74 inaction or under action under reaction,701.16,5.28 is still pretty high,703.74,6.42 um another thing is is two uh or two two,706.44,5.459 things are exponential growth and,710.16,4.38 saltatory leaps so exponential growth,711.899,4.081 which I provided some evidence for at,714.54,2.52 the beginning of this video that's,715.98,3.24 happening that is a fact,717.06,3.959 and then actually let me go back to this,719.22,4.619 so this here where you have this Gap up,721.019,5.461 this is actually a mathematical evidence,723.839,5.221 of a what's called a saltatory leap so a,726.48,4.5 saltatory leap is when some breakthrough,729.06,4.98 or some uh compounding returns of,730.98,5.94 incremental progress result in sudden,734.04,4.739 breakthroughs that you could not have,736.92,3.659 predicted because if you just look at,738.779,3.721 this trend line you'd predict like okay,740.579,3.601 we wouldn't be here for another couple,742.5,4.32 years but we're here now right so that,744.18,4.92 that's a saltatory leap so you have to,746.82,5.28 acknowledge that saltatory leaps not,749.1,4.979 just do happen sometimes have happened,752.1,4.679 recently and if it's happened recently,754.079,4.981 it might happen again,756.779,5.281 um the lack of urgency by saying eh it's,759.06,5.7 decades away again you know you got to,762.06,3.959 think through it like okay but what if,764.76,2.699 it's not,766.019,5.041 um and the the the taking a big step,767.459,5.521 back the nothing to see here messaging,771.06,5.16 might lead to boiled frog syndrome the,772.98,5.039 the temperature is rising quickly this,776.22,3.54 year I think a lot of us agree on that,778.019,3.841 and so well you get used to it right,779.76,5.16 okay it's warmer than it was but it's,781.86,5.88 not hot yet thing is the time between it,784.92,4.44 gets warm and it gets hot and it starts,787.74,5.039 boiling that time could be shortening,789.36,6.719 so the social impacts of these of of,792.779,6.18 when thought leaders adopt more extreme,796.079,6.06 uh stances such as doomerism or deny uh,798.959,6.601 denialism is uh basically just a quick,802.139,6.421 recap the doomers create nihilism and,805.56,6.779 fatalism which discourages uh proactive,808.56,5.219 Solutions because they say that there is,812.339,3.421 no solution right that is one of the,813.779,3.601 underpinning assumptions of doomers is,815.76,4.079 that it's inevitable it's unavoidable,817.38,5.459 there is no solution don't even try,819.839,4.56 um and this promotes fear and anxiety,822.839,4.261 right which yes fear and anxiety are,824.399,5.041 evolutionarily there for a reason to,827.1,3.96 motivate us to do something about a,829.44,3.18 problem that we perceive,831.06,4.26 but too much of a of of an important,832.62,4.32 thing can still be bad,835.32,5.04 finally our next the impact of denialism,836.94,5.22 or denialism,840.36,3.599 um is that there's a false sense of,842.16,3.66 security right and we don't have to,843.959,4.021 worry about it eh it's not coming for a,845.82,4.74 long time right that's complacency and,847.98,5.88 act and inaction which undermines some,850.56,4.44 of the rest of us who are working and,853.86,3.0 saying actually this might be something,855.0,3.3 that we need to think a little bit,856.86,3.36 further ahead on,858.3,4.02 um because think about last year right,860.22,5.46 how um AI images exploded onto the scene,862.32,5.94 and nobody was ready for it right what,865.68,5.58 if the next thing that happens is not,868.26,5.22 just AI images or AI music or something,871.26,4.44 like that but something a little bit,873.48,3.9 more profound a little bit more,875.7,3.18 significant that we just weren't ready,877.38,3.3 for which means that the time to start,878.88,3.84 preparing for those things that we know,880.68,4.44 are coming eventually and we don't know,882.72,4.26 when that Jack-in-the-Box is gonna pop,885.12,4.079 the time to prepare is now,886.98,4.44 so some of the some of the consequences,889.199,4.981 that occur because of these the the,891.42,5.46 messaging is one polarization of of,894.18,5.279 public opinion some people are bending,896.88,4.62 over backwards to say more jobs are,899.459,4.38 coming let's not even think about you,901.5,5.519 know AI based war or the control problem,903.839,5.461 or anything like that uh what meanwhile,907.019,5.341 others are like ah no we're you know,909.3,4.92 lead the leaders around the world are,912.36,4.14 not even addressing these risks they're,914.22,4.679 just sitting on their on their hands so,916.5,4.86 we're doomed right because if the adults,918.899,3.901 in the room don't care or don't think,921.36,3.719 it's a problem but all of the kids are,922.8,3.779 like hey do you see that the house is on,925.079,3.301 fire like maybe we should put that out,926.579,4.26 first right that leads to more nihilism,928.38,5.699 more fatalism a lot of polarization,930.839,5.641 and then of course uh the Overton window,934.079,4.2 is still too narrow so the Overton,936.48,3.719 window is the concept of what is allowed,938.279,3.841 to be talked about in political,940.199,5.521 discourse so you know if you if you,942.12,5.159 follow my channel every now and then,945.72,3.179 I'll post links to videos like hey look,947.279,3.661 you know the conversation is Shifting,948.899,5.661 right just yesterday I posted a a a a a,950.94,6.48 video from DW news which is in Germany,954.56,5.74 where they try to address like hey,957.42,5.64 there's actual anxiety about like,960.3,5.58 everyone's jobs are going away right and,963.06,5.76 they bent over real God I listened to it,965.88,4.92 again they bent over backwards to try,968.82,4.319 and say well yeah a lot of low-paid jobs,970.8,3.96 are going away but there's a few high,973.139,3.901 paid jobs coming in and it's like okay,974.76,4.199 but still the point is is that most of,977.04,3.479 the medium and low-paid jobs are going,978.959,3.781 away and being replaced by a few,980.519,4.62 high-paying jobs that's not the promise,982.74,4.62 of like techno Revolution where AI,985.139,4.5 creates a whole bunch of new jobs,987.36,5.4 um and then I think it was Amazon uh or,989.639,5.101 Facebook one of them just an announced,992.76,4.68 even more layoffs and they explicitly,994.74,4.8 said that the reason for the layoffs is,997.44,3.12 that they're going to replace as many,999.54,3.599 people with AI as possible I called it,1000.56,6.12 I've been saying it so it's happening,1003.139,5.64 um so the Overton window is Shifting now,1006.68,5.599 okay why is this a big problem why like,1008.779,6.3 fundamentally why is it that some people,1012.279,5.92 are dimerous and denialists and what,1015.079,5.401 what is left in the wash what is missing,1018.199,4.981 from the conversation so one thing,1020.48,4.68 that's missing is there is not a,1023.18,4.44 coherent Global strategy,1025.16,5.22 and so what I mean by that is everyone's,1027.62,5.4 busy arguing you know in this little,1030.38,5.039 domain or this little domain uh you know,1033.02,5.279 about corporate governance or academic,1035.419,5.701 Integrity or should we have a moratorium,1038.299,5.581 right there's not really a global,1041.12,4.439 strategy no one has even proposed,1043.88,3.24 anything,1045.559,5.761 um then on top of that is as I mentioned,1047.12,5.64 just a moment ago,1051.32,3.66 uh calling for moratoriums is not a,1052.76,3.539 solution,1054.98,2.699 um that's not even that's not even a,1056.299,2.941 stopgap measure,1057.679,3.24 um and so when all the thought leaders,1059.24,3.6 in the world when none of them are,1060.919,3.841 really offering Solutions of course,1062.84,4.44 you're going to end up with a lot of uh,1064.76,4.98 bickering and arguing and also a lot of,1067.28,5.54 anxiety right we are humans and we love,1069.74,6.48 love when there are adults in the room,1072.82,5.979 that we trust to help make good,1076.22,4.5 decisions and to make sure that we're,1078.799,4.681 going to be okay right and right now on,1080.72,5.579 the topic of AI there's nobody really,1083.48,4.74 out there saying we're gonna be okay,1086.299,4.321 I've got a plan,1088.22,5.52 um and then uh on top of that Global,1090.62,4.799 strategy is a comprehensive roadmap,1093.74,4.819 right kind of the same thing,1095.419,3.14 said a lot of this stuff but really what,1098.78,4.259 we need is a is that Global,1100.34,4.5 comprehensive roadmap and a,1103.039,4.441 multi-layered approach to solving all,1104.84,4.92 these problems at all these different uh,1107.48,3.66 levels,1109.76,3.24 so I've already alluded to some of these,1111.14,3.36 things there's quite a bunch of stuff,1113.0,3.299 that doesn't work right calling for,1114.5,4.14 moratoriums just simply does not work,1116.299,4.62 we'll get into more detail about why,1118.64,6.06 moratoriums don't work and and and uh,1120.919,5.401 and all the incentives against it in,1124.7,3.359 just a moment another thing that doesn't,1126.32,4.14 work is bombing data centers sorry that,1128.059,4.921 is a really bone-headed suggestion,1130.46,5.099 uh complaining on Twitter writing op-eds,1132.98,4.98 writing mean comments on YouTube none of,1135.559,4.98 these things are actually helpful and,1137.96,3.839 another thing that's not helpful is,1140.539,2.88 actually just trusting corporations or,1141.799,3.061 the establishment to figure it out on,1143.419,2.281 their own,1144.86,4.02 we are all all humans Global,1145.7,5.96 stakeholders in AI,1148.88,5.82 so all these these the this list of,1151.66,4.36 stuff that I've just have that that,1154.7,2.88 doesn't work they're all molecule,1156.02,3.779 reactions and molecule Solutions which,1157.58,3.3 basically means that they will,1159.799,3.181 inevitably lead to those lose-lose,1160.88,4.02 outcomes that the doomers are are,1162.98,4.079 warning us against right again I'm not,1164.9,4.2 saying that the doomers are wrong if,1167.059,3.721 things keep going as they are the,1169.1,4.5 doomers are right I just I personally,1170.78,6.18 don't ascribe to constantly yelling fire,1173.6,6.12 and then claiming you know we're all,1176.96,4.44 gonna die,1179.72,5.579 okay so I outlined the big problems now,1181.4,5.82 what,1185.299,4.26 this video the entire purpose is to,1187.22,4.92 introduce kind of the crowning,1189.559,4.5 achievement so far of What Not Just I'm,1192.14,4.62 working on but the the rapidly growing,1194.059,5.401 community that I'm building,1196.76,4.74 um uh what started around the years to,1199.46,5.04 comparatives my research on alignment,1201.5,6.059 for individual models and agents it has,1204.5,6.78 quickly expanded so this gato framework,1207.559,7.62 Global alignment taxonomy Omnibus is,1211.28,6.42 that comprehensive strategy that I just,1215.179,4.74 mentioned that is missing it is not just,1217.7,4.56 for responsible AI development but is a,1219.919,5.461 coherent road map that everyone on the,1222.26,5.58 planet can participate in at various,1225.38,4.86 levels whatever level makes the most,1227.84,3.66 sense to you,1230.24,5.4 this framework has seven layers on ways,1231.5,8.16 to implement uh models AI systems and,1235.64,6.539 also alignment uh alignment-based,1239.66,4.8 regulations and we'll get into all the,1242.179,4.201 layers in just a moment,1244.46,5.579 uh but basically the the whole point of,1246.38,5.22 this gato framework that we're working,1250.039,4.861 on is that it will unite all,1251.6,5.28 stakeholders give us a common framework,1254.9,5.04 with which to have these discussions to,1256.88,5.28 broaden the Overton window to open the,1259.94,4.38 Overton window a little bit more so,1262.16,4.5 whatever part of the spectrum you're on,1264.32,4.859 whether you're saying eh it's not really,1266.66,4.379 an issue yet or we're all going to die,1269.179,4.261 or you don't care or you're an optimist,1271.039,4.981 whatever this is a framework that we can,1273.44,4.56 all participate in,1276.02,4.86 um just in a decentralized distributed,1278.0,5.64 and open source manner,1280.88,5.52 so as promised here are the seven layers,1283.64,5.1 of the gato framework and in the,1286.4,4.08 community we started saying that it's,1288.74,3.6 like a seven layer burrito so we use,1290.48,5.1 like taco cat as our little Avatar so,1292.34,5.88 layer one the lowest layer is model,1295.58,5.64 alignment so model alignment has to do,1298.22,6.0 with individual neural networks so that,1301.22,8.52 means gpt2 gpt3 gpt4 Bert vicuna uh,1304.22,8.28 stable LM all of these right large,1309.74,5.34 language models are proliferating like,1312.5,4.559 well I don't know just like locusts,1315.08,3.66 whatever,1317.059,4.441 it's happening right data sets are,1318.74,5.16 growing models are growing they're all,1321.5,4.799 coming out uh the cat's out of the bag,1323.9,5.04 right language technology multimodal,1326.299,4.801 technology it's all coming you can't,1328.94,3.719 stop it,1331.1,3.48 um so rather than stop it rather than,1332.659,4.801 call for moratoriums what we're doing is,1334.58,5.459 we're focusing on okay let's ride this,1337.46,5.06 wave I all have already proposed,1340.039,4.741 reinforcement learning with heuristic,1342.52,3.519 imperatives which is different from,1344.78,2.34 reinforcement learning with human,1346.039,3.661 feedback because human feedback aligns,1347.12,5.28 models to what humans want which what,1349.7,4.56 humans want and what humans need often,1352.4,4.259 very very different here is to,1354.26,3.84 comparatives is not just what humans,1356.659,4.621 want but what all life needs we're also,1358.1,4.74 talking about data set curation and,1361.28,3.06 inner alignment problems like Mesa,1362.84,3.12 optimization,1364.34,6.06 Layer Two is sorry autonomous systems so,1365.96,6.18 these are cognitive architectures and,1370.4,3.38 autonomous agents,1372.14,4.919 this is this is recently exploded on the,1373.78,4.48 scene with,1377.059,4.381 um you know Jarvis and baby AGI and,1378.26,5.76 agent GPT and all that fun stuff so you,1381.44,4.38 guys know what that is and it's coming,1384.02,3.12 and it's only going to get more,1385.82,3.359 sophisticated we're on the ground floor,1387.14,4.62 of autonomous systems this is year zero,1389.179,6.061 year two three four five like you can't,1391.76,5.58 on you cannot imagine how powerful,1395.24,4.679 autonomous systems are going to be in,1397.34,5.4 the coming years so at the at the the,1399.919,4.981 low level the engine level right the,1402.74,4.2 components under the hood that's the,1404.9,4.74 models the autonomous systems are the,1406.94,4.979 software architectures that use those,1409.64,3.96 systems including memory systems and,1411.919,4.021 apis and other stuff to create those,1413.6,5.52 autonomous cognitive entities right,1415.94,6.66 layer 3 is the decentralized network so,1419.12,5.16 you might have seen some of my recent,1422.6,2.6 videos where I've talked about,1424.28,3.3 blockchain decentralized autonomous,1425.2,4.959 organizations and also another component,1427.58,4.5 of that is what's called a federation so,1430.159,4.741 a federation is where you have either,1432.08,4.44 independent nodes or independent,1434.9,3.48 networks that can communicate and,1436.52,5.039 collaborate through Federated systems so,1438.38,8.039 these are the the network layer is how,1441.559,7.86 do we create networked intelligent,1446.419,6.961 entities that are also aligned and this,1449.419,5.701 is a tough nut to crack we've had lots,1453.38,3.419 of discussions in the group talking,1455.12,3.72 about can you implement Heroes to,1456.799,5.941 comparatives as a consensus mechanism at,1458.84,5.459 what level do you process it do you,1462.74,5.64 process it at every llm inference or do,1464.299,6.301 you wait for the decisions how do you,1468.38,4.679 make decisions around this kind of thing,1470.6,5.52 excuse me real tough nut to crack number,1473.059,6.0 four is where we jump from the technical,1476.12,5.34 implementation and research to more of,1479.059,5.221 the social political and economic uh,1481.46,4.5 layer of the stack,1484.28,3.36 and for all of you technologists out,1485.96,4.32 there you can probably see,1487.64,5.34 um my influence as a as a technologist,1490.28,5.04 because this is it's not modeled on the,1492.98,5.04 osm OSI model it's actually more closely,1495.32,4.92 modeled on the defense and depth model,1498.02,5.58 but it is a layered hierarchical stack,1500.24,6.6 or onion of uh of Concepts so corporate,1503.6,4.199 adoption,1506.84,2.339 here's the thing,1507.799,3.841 you cannot just tell a corporation you,1509.179,4.201 know what stop with the AI we don't we,1511.64,3.72 don't like where AI is going sure you,1513.38,4.74 can try to with regulation uh but you,1515.36,4.5 know like Italy tried to do that and,1518.12,3.9 then they reverse course right,1519.86,4.26 there's just way too much economic,1522.02,4.86 incentive the bottom line you know that,1524.12,4.76 is if you're if you're a corporation,1526.88,4.5 shareholders and the bottom line that's,1528.88,4.659 where the power is so rather than fight,1531.38,4.38 that part of what this framework does is,1533.539,4.321 say let's how let's figure out how we,1535.76,4.56 can align those heuristic imperatives,1537.86,4.26 reduce suffering increase prosperity and,1540.32,4.38 increase understanding how can we align,1542.12,5.22 those fundamental human needs the,1544.7,4.26 fundamental needs of all living things,1547.34,3.66 with corporate interest,1548.96,4.68 and so one story that I like to share is,1551.0,4.5 that I've had a few patreon uh,1553.64,3.659 supporters reach out to me and they're,1555.5,3.36 like hey I've got this autonomous system,1557.299,3.0 that I'm working on but it's like it's,1558.86,3.96 getting stuck or I need help or whatever,1560.299,5.041 um or even without asking for my help uh,1562.82,4.2 they said like hey I implemented the,1565.34,3.24 heroes to comparatives in my autonomous,1567.02,3.0 Business Systems and they work better,1568.58,5.699 and I'm like thanks share so like if you,1570.02,5.82 have any of those examples please post,1574.279,2.941 them on Reddit on the heroes to,1575.84,3.48 comparative subreddit because we need we,1577.22,3.839 need more of those stories about how,1579.32,3.78 aligned AI systems are actually good for,1581.059,4.381 business it's that simple the bottom,1583.1,4.86 line like I I will always say that,1585.44,4.56 corporations are intrinsically amoral,1587.96,5.219 however what I will say is that is that,1590.0,4.919 their profit motive their primary,1593.179,3.12 incentive structure which is to make,1594.919,4.62 more money will benefit from adopting,1596.299,5.101 heuristic comparative aligned systems,1599.539,4.081 services and products which I also we,1601.4,3.54 also have some members of the community,1603.62,3.12 who are working on spinning this out,1604.94,3.18 into either for-profit or not,1606.74,4.679 not-for-profit services and of course,1608.12,4.799 we're going to be publishing open source,1611.419,3.601 data sets reference architectures that,1612.919,3.661 sort of stuff to make it as easy as,1615.02,3.899 possible for corporations all over the,1616.58,5.459 world to adopt aligned AI,1618.919,5.041 uh and we're going to work on convincing,1622.039,3.24 them that this is the way to go too,1623.96,3.86 number five National regulations,1625.279,5.041 obviously as I just mentioned you know,1627.82,5.08 corporations can or sorry Nations can do,1630.32,4.28 some stuff like people pointed out like,1632.9,5.159 gdpr uh European unions like you know,1634.6,6.459 big package about like a uh data privacy,1638.059,5.161 and stuff and certainly as an I.T,1641.059,3.261 professional,1643.22,3.9 people on the technology side are,1644.32,5.14 terrified of gdpr right that's got some,1647.12,4.02 teeth right you know right to be,1649.46,3.78 forgotten where the data is owned and,1651.14,4.44 housed and data governance okay great,1653.24,6.179 that's all fine but see the thing is is,1655.58,5.82 Nations have their own incentive,1659.419,4.14 structure where it comes to Ai and what,1661.4,4.56 I mean by that is uh the the national,1663.559,4.441 interests of companies has to do with,1665.96,4.92 their own GDP as a whole so this is a,1668.0,6.539 big difference gdpr was about uh like,1670.88,5.64 data privacy for Citizens and social,1674.539,5.341 media it wasn't as directly tied to like,1676.52,6.899 the national growth of their G ADP it,1679.88,5.52 wasn't necessarily directly tied to,1683.419,3.661 their geopolitical influence or their,1685.4,4.68 military or their National Security,1687.08,4.92 AI today though,1690.08,4.199 is all of those things and more,1692.0,5.46 because GDP growth geopolitical,1694.279,5.101 influence National Security border,1697.46,4.38 security whatever all of that has to do,1699.38,4.44 those are the national interests that we,1701.84,4.62 are going to be working on aligning AI,1703.82,6.0 with and basically the long story short,1706.46,6.36 is at a national level We're not gonna,1709.82,4.92 we're not going to say hey Nations maybe,1712.82,4.02 you shouldn't adopt AI maybe you should,1714.74,3.179 slow it down maybe you should just,1716.84,3.12 regulate it we're going to be actually,1717.919,3.481 more I'm not going to say that like,1719.96,3.42 we're accelerationists because like you,1721.4,3.779 don't need to push the to go any faster,1723.38,3.48 right I'm not advocating for,1725.179,3.721 accelerationism I'm just observing that,1726.86,4.26 acceleration is happening so how do we,1728.9,4.639 steer it right and the idea is,1731.12,5.34 encouraging Nations to adopt Heroes,1733.539,5.921 comparative aligned uh models services,1736.46,5.699 and systems because at every level of,1739.46,4.92 government that will help steer the,1742.159,4.081 nation in a better Direction and their,1744.38,4.5 implementations will be safer more,1746.24,4.799 reliable more trustworthy so on and so,1748.88,4.26 forth and of course stability is good,1751.039,3.24 for business it's good for the account,1753.14,3.32 economy it's good for National Security,1754.279,5.52 and all that other fun stuff next up is,1756.46,5.5 number six uh layer six International,1759.799,4.681 treaties so I actually did wasn't the,1761.96,4.02 first one to come up with this idea but,1764.48,3.48 basically we're going to be advocating,1765.98,4.679 for an international Consortium like,1767.96,4.92 CERN but for AI because here's the other,1770.659,4.02 thing and a lot of people pointed this,1772.88,4.14 out is that a lot of Nations,1774.679,5.22 cannot even afford to participate in AI,1777.02,5.7 research right AI research is carried,1779.899,4.561 out largely by the wealthiest companies,1782.72,3.0 on the planet and the wealthiest,1784.46,2.76 countries on the planet,1785.72,3.42 that's going to intrinsically leave a,1787.22,4.439 lot of other nations uh behind in the,1789.14,4.98 dust right and that's just not fair that,1791.659,4.201 is a malarchy outcome where there's a,1794.12,3.419 few wealthy bastions and the rest are,1795.86,4.98 poor and they end up basically like,1797.539,6.26 tossed on the on the rough Seas of an AI,1800.84,5.64 saturated world so what we're going to,1803.799,5.081 do is we're going to advocate for a,1806.48,5.64 global international Consortium where uh,1808.88,6.899 people people Nations pool resources,1812.12,5.88 share their scientists share their,1815.779,5.101 research share their data so that we can,1818.0,5.46 all benefit equally across the whole,1820.88,6.539 globe which that also uh has uh knock-on,1823.46,6.079 benefits with in terms of alliances,1827.419,4.681 economic benefits because you look at,1829.539,4.721 like everyone's going to benefits from,1832.1,4.319 from like CERN and the collaborations,1834.26,4.44 between like NASA and Esa and and that,1836.419,3.721 sort of stuff so International,1838.7,3.959 scientific treaties generally one,1840.14,3.72 they've got a pretty good track record,1842.659,3.24 and two we've got a good model for them,1843.86,3.6 so we're just basically saying let's,1845.899,4.861 copy the success of NASA Issa of CERN,1847.46,6.24 and let's do it for AI again that's not,1850.76,5.46 like you know we're not this is nothing,1853.7,3.839 Earth shattering right it's been done,1856.22,2.88 before we're just saying maybe it is,1857.539,3.781 time to do this with AI and finally,1859.1,5.1 layer 7 of the gato framework is global,1861.32,5.76 consensus so Global consensus has to do,1864.2,5.219 with messaging,1867.08,5.459 um uh working with universities academic,1869.419,5.401 institutions uh industrial sectors,1872.539,5.701 National sectors uh social media right,1874.82,6.3 or all media really because if we can,1878.24,5.819 build consensus in every sector in every,1881.12,5.059 domain and at every level of society,1884.059,6.961 then consensus around how to uh align AI,1886.179,7.0 so that we all end up in a more utopian,1891.02,3.06 state,1893.179,3.48 the utopian attractor State rather than,1894.08,5.219 dystopia or Extinction then we're going,1896.659,4.561 to have a lot more energy right that,1899.299,3.961 Overton window is going to be aligned in,1901.22,4.14 the correct direction rather than you,1903.26,3.06 know because right now the Overton,1905.36,3.24 window is highly highly centered over,1906.32,4.979 we're all going to die or nothing is,1908.6,6.179 happening but really the truth is well,1911.299,5.461 those are possibilities but the Overton,1914.779,4.081 window needs to be broadened and that is,1916.76,4.139 one of the key components of global,1918.86,3.78 consensus,1920.899,4.38 so I just threw a lot at you and this,1922.64,5.82 all sounds really good Pie in the Sky uh,1925.279,5.461 you know blah blah right there's,1928.46,3.66 probably some skepticism so let's,1930.74,4.08 address that this all started as a very,1932.12,5.039 small Discord Community where I just,1934.82,3.9 wanted to bring some people together to,1937.159,3.061 help me do Heroes to comparatives,1938.72,4.92 research and it quickly very quickly,1940.22,5.22 scaled up,1943.64,4.74 um we to as of this recording we have I,1945.44,5.04 think right around just shy of 70 people,1948.38,5.22 involved and more people coming all the,1950.48,4.62 time we're actually having to work on,1953.6,4.14 figuring out ways of automating the,1955.1,4.74 recruiting the applications and the,1957.74,3.419 onboarding which we haven't figured out,1959.84,4.079 yet but we need to,1961.159,4.681 um we're organizing teams and projects,1963.919,4.38 around each layer of gato that I just uh,1965.84,4.02 outlined and so you can see those here,1968.299,4.5 on the right hand side so if you're a,1969.86,4.62 reinforcement learning researcher or an,1972.799,4.26 ml researcher or a data scientist we,1974.48,5.1 need your help with layer one if you're,1977.059,4.081 a software architect or a cloud,1979.58,4.079 architect or someone or devops someone,1981.14,5.34 who understands Automation and complex,1983.659,4.681 systems we need your help in Layer Two,1986.48,4.319 autonomous systems we've got a whole,1988.34,5.4 bunch of blockchain endow people working,1990.799,5.1 with us on layer three which is such a,1993.74,3.72 cool topic because where this is like,1995.899,3.541 super Cutting Edge also we're going to,1997.46,3.42 eat our own dog food we're already,1999.44,5.88 working on using Dows to help voting and,2000.88,6.0 decision making and allocation of,2005.32,4.32 resources within this project obviously,2006.88,4.98 as I've said in many of my videos a lot,2009.64,4.139 of blockchain and DOW technology is not,2011.86,4.319 ready but we are going to eat our own,2013.779,4.38 dog food and make sure that we are,2016.179,4.021 testing these things so that they'll do,2018.159,3.36 the things that we say that they need to,2020.2,2.819 do right we're going to figure it out as,2021.519,4.441 as we go number four corporate adoption,2023.019,5.821 we have a few entrepreneurs and Business,2025.96,5.819 Leaders we've got uh several ctOS in the,2028.84,4.8 group we need more connections to,2031.779,3.9 business and industry this means,2033.64,4.2 conferences this means,2035.679,4.38 um meetups this means,2037.84,4.8 um people on boards right A lot of my,2040.059,4.261 patreon supporters are business people,2042.64,3.659 and so like I work with them directly,2044.32,4.039 but we need more of that we need people,2046.299,5.401 uh working to evangelize,2048.359,5.081 um not just not just like saying hey,2051.7,3.659 Corporation you should adopt your,2053.44,3.239 heuristic imperatives and then leaving,2055.359,3.54 it at that we have startups that we're,2056.679,4.801 working with because the the companies,2058.899,4.801 offering aligned Services don't exist,2061.48,4.439 yet so we're helping incubate those,2063.7,3.36 things and I don't mean from a financial,2065.919,3.18 perspective but from a consultation,2067.06,5.039 perspective and so because if the if hi,2069.099,5.401 aligned Services goods and services,2072.099,5.401 exist companies can adopt them but until,2074.5,5.399 they exist they can't be adopted really,2077.5,5.159 number five National regulation we're,2079.899,5.7 just starting to have this conversation,2082.659,4.561 um actually just a conversation I had,2085.599,3.901 just a little while ago had to do with,2087.22,5.34 uh talking with some of the uh policy,2089.5,6.3 makers and lawyers and legislators that,2092.56,5.039 are concerned about this kind of stuff,2095.8,3.24 so for instance,2097.599,3.541 um the vice president uh I don't know if,2099.04,4.2 it's today but soon we'll be talking,2101.14,4.86 with all of the big Tech Giants right so,2103.24,3.66 we need to have more of those,2106.0,3.3 conversations and we need to add some of,2106.9,4.92 uh some of our perspective from the gato,2109.3,3.9 framework,2111.82,4.32 um into those National conversations but,2113.2,4.7 not just from it not just from a,2116.14,3.719 regulatory standpoint of the nation,2117.9,4.36 looking down into the nation the,2119.859,4.081 nation's looking up and out to the rest,2122.26,3.24 of the world because as I mentioned,2123.94,4.32 National Security that is a huge thing,2125.5,4.8 GDP growth that is a big thing in,2128.26,4.5 geopolitical influence AI is going to,2130.3,5.039 affect all of these domains number six,2132.76,5.7 uh the international treaty again we,2135.339,5.821 need we need people that are connected,2138.46,5.159 to the UN,2141.16,3.24 um,2143.619,4.321 uh maybe NATO I don't know oecd all,2144.4,6.48 kinds of stuff uh UNESCO there's all,2147.94,4.5 kinds of international organizations,2150.88,4.02 that we would like to be connected with,2152.44,5.34 and work with and talk to in order to,2154.9,5.4 have these conversations and,2157.78,4.02 by and large just make the right,2160.3,3.299 connections so that these conversations,2161.8,4.14 are happening and we can articulate the,2163.599,5.161 gato framework and get it published and,2165.94,5.58 then finally layer 7 Global consensus we,2168.76,4.14 have writers we have graphic,2171.52,3.0 communicators we've got editors we've,2172.9,3.66 got audio Engineers,2174.52,4.76 um we're working with uh people all over,2176.56,6.12 even more influencers have excuse me,2179.28,6.22 reached out to me so I'm uh I'm going to,2182.68,4.74 be having conversations with them so,2185.5,6.18 that we can all align on this consensus,2187.42,6.6 and then here's our uh here's our our,2191.68,4.38 mascot it's our own version of taco cat,2194.02,5.16 so again you know gato cat and then you,2196.06,7.26 know seven layered Taco you get the idea,2199.18,6.48 um okay so you're probably glazing over,2203.32,4.14 at this point but you've got the meat of,2205.66,3.3 it so if you're really really super,2207.46,4.02 interested in the layers let's take a,2208.96,4.56 look at the layers of Gato in a little,2211.48,5.22 bit bigger depth so number one of uh,2213.52,5.22 layer one model alignment fine tuning,2216.7,3.72 the very first experiment that I,2218.74,4.02 published was on fine tuning large,2220.42,4.14 language models so that they are aligned,2222.76,4.38 number two reinforcement learning again,2224.56,4.44 that is the goal is how do you create,2227.14,4.08 the data sets in the systems and the,2229.0,6.0 signals in order to have uh models that,2231.22,6.06 not only are initially aligned to,2235.0,4.32 heuristic imperatives and human needs,2237.28,4.2 and the needs of all life but how do you,2239.32,3.72 make sure that they get better at that,2241.48,3.3 over time right that is the entire,2243.04,4.2 purpose of heuristics heuristics uh,2244.78,3.839 heuristic imperatives and reinforcement,2247.24,4.56 learning basically the same thing,2248.619,5.281 um at least here's the comparatives are,2251.8,3.84 reinforcement learning on a specific,2253.9,2.939 trajectory,2255.64,4.74 model bias so there's uh there's a lot,2256.839,5.821 of intrinsic bias in models there's been,2260.38,4.56 uh some really interesting studies even,2262.66,4.02 chat GPT with reinforcement learning,2264.94,3.72 with human feedback is still pretty,2266.68,4.439 sexist it's also pretty racist depending,2268.66,3.9 on the kinds of prompts that you use,2271.119,3.661 there's a lot of implicit bias then,2272.56,4.68 there's also um Mesa optimization which,2274.78,4.74 I'm not sure I'm not entirely sure that,2277.24,4.74 Mesa optimization is a problem for,2279.52,5.16 language models but it could be,2281.98,4.68 um so we'll see but we need to be aware,2284.68,3.659 of that and we need to study it and if,2286.66,3.54 it is there we need to address it but,2288.339,4.321 Mesa optimization is like a tiny,2290.2,4.44 component of this whole framework,2292.66,5.699 open source data sets so one of the,2294.64,6.3 things that I mentioned is open source,2298.359,4.98 open source Open Source by by creating,2300.94,4.26 and Publishing open source data sets,2303.339,5.28 that can uh one they're transparent but,2305.2,6.06 two that can foster collaboration and,2308.619,4.261 ultimately one of the things that I hope,2311.26,4.26 to achieve is what I call axiomatic,2312.88,5.219 alignment so axiomatic alignment is what,2315.52,4.8 happens when through conversation,2318.099,4.98 through experimentation through repeated,2320.32,5.64 augmentation of Open Source data sets,2323.079,5.341 practically every data set out there,2325.96,5.76 that AI is trained on intrinsically has,2328.42,5.88 some alignment baked into it and if,2331.72,5.1 every data set or all or if enough data,2334.3,5.94 sets are aligned then you can end up,2336.82,6.18 with a a virtuous cycle or a positive,2340.24,5.46 feedback loop where every subsequent,2343.0,5.339 data set is also more and more aligned,2345.7,5.46 so from a model perspective the,2348.339,5.041 overarching goal is to arrive at a place,2351.16,4.56 of axiomatic alignment,2353.38,4.14 so this will require us to solve,2355.72,4.2 problems around training model,2357.52,5.04 architecture and then finally the data,2359.92,5.1 ecosystem that we build and when I say,2362.56,4.559 we I don't mean just those of us in gato,2365.02,4.74 the gato framework project but everyone,2367.119,4.74 everyone participating this whether,2369.76,4.2 they're academic researchers corporate,2371.859,4.321 government military so on and so forth,2373.96,5.28 now Layer Two autonomous systems I,2376.18,3.96 already talked a little bit about,2379.24,3.66 cognitive architecture we don't need to,2380.14,4.92 um you know beat the dead horse there,2382.9,4.439 but one of the things that we want to,2385.06,4.62 talk about and and publish is an open,2387.339,4.26 source reference architecture that's,2389.68,3.72 really the primary one of the primary,2391.599,3.961 goals here is what are the components,2393.4,3.84 what are the system components that you,2395.56,4.559 need in order to have a fully aligned,2397.24,5.52 and fully autonomous system so this,2400.119,3.901 includes some of these things like,2402.76,3.78 self-evaluation and stability we are,2404.02,5.339 working on how do you how do you design,2406.54,4.92 tasks how do you evaluate past,2409.359,3.661 performance how do you automatically,2411.46,3.78 label data and how do you create modular,2413.02,4.98 design patterns that allow for anyone,2415.24,6.18 and everyone to create their own fully,2418.0,5.82 autonomous systems that are also aligned,2421.42,5.22 to the heuristic imperatives and,2423.82,6.0 therefore should be benevolent so by,2426.64,5.1 getting by having the ultimate goal of,2429.82,3.84 publishing these open source reference,2431.74,3.72 architectures that'll make it really,2433.66,4.02 easy for all corporations out there and,2435.46,3.78 all private individuals and all,2437.68,4.86 governments to adopt these uh these,2439.24,4.859 patterns these software architecture,2442.54,4.62 patterns which again just by providing,2444.099,4.621 that answer and making it as easy as,2447.16,5.3 possible will be a one component in,2448.72,5.82 solving alignment and the control,2452.46,4.24 problem globally,2454.54,5.52 so decentralized networks this is not,2456.7,5.52 just blockchain not just Dows but also,2460.06,5.34 federations so keep that in mind,2462.22,4.28 um,2465.4,4.08 there's two primary components here one,2466.5,5.02 first we just have to figure out how to,2469.48,3.72 do these Technologies because by and,2471.52,3.72 large these are still highly,2473.2,4.32 experimental Technologies,2475.24,3.839 um and I will be the first to admit that,2477.52,3.18 maybe blockchain and DOW is not the,2479.079,4.381 correct way but in principle some kind,2480.7,4.8 of Federated system or decentralized,2483.46,4.68 network is probably the way to go in,2485.5,4.74 order to have some of these things such,2488.14,4.68 as algorithmic consensus when we're in a,2490.24,3.96 world where we have billions upon,2492.82,3.6 billions of autonomous agents all,2494.2,4.62 working on their own we need a way for,2496.42,4.8 them to work with each other and with us,2498.82,6.06 to come up with a consensus mechanisms,2501.22,7.02 that will slow slow the roll basically,2504.88,5.04 so there's a couple components that can,2508.24,3.96 go into that one is trust and reputation,2509.92,6.179 mechanisms so if you have you know some,2512.2,6.659 arbitrary AI agent operating out on the,2516.099,4.26 net on its own,2518.859,4.561 if it is an untrusted agent then maybe,2520.359,4.74 you don't want to give it resources or,2523.42,3.199 you don't want to give it any Credence,2525.099,3.541 that's what I mean by trust and,2526.619,4.0 reputation mechanisms resource control,2528.64,4.8 and allocation is another aspect of,2530.619,4.861 using blockchain or Dao or Federated,2533.44,4.86 Technologies which basically means if an,2535.48,5.16 agent is behaving in a way that is not,2538.3,5.4 aligned if the consensus of all agents,2540.64,4.74 says hey that's a little bit destructive,2543.7,3.84 maybe you shouldn't do it you revoke its,2545.38,4.979 access to computational resources data,2547.54,5.22 that sort of thing which can be a way to,2550.359,6.301 allow and Empower uh autonomous agents,2552.76,5.7 to police each other,2556.66,5.16 and then finally incentivizing alignment,2558.46,5.1 um so one of the things uh that people,2561.82,3.48 are concerned about is instrumental,2563.56,3.84 convergence so instrumental convergence,2565.3,6.72 is the idea that um AI uh no matter what,2567.4,6.78 goals you give it will be incentivized,2572.02,4.92 to pursue basic similar things like,2574.18,5.22 control of power more data that sort of,2576.94,5.639 stuff but so that's that's based on its,2579.4,5.699 in intrinsic motivations right an AI,2582.579,5.401 needs electricity to run so therefore it,2585.099,4.081 will always have some intrinsic,2587.98,3.9 motivation to do that now through these,2589.18,4.32 Network systems whether it's Federated,2591.88,3.9 decentralized however the network,2593.5,5.579 architecture is ultimately designed if,2595.78,6.18 you incentivize their behavior to get,2599.079,4.801 the behavior that you want so that they,2601.96,4.139 can get what they want then that is the,2603.88,5.52 way to go so for instance if you use,2606.099,5.581 resource tokens or cryptocurrency or,2609.4,3.78 whatever to say hey,2611.68,3.72 everything that you do that is aligned,2613.18,4.8 the the the the the the rest of the,2615.4,4.38 network says we agree with that behavior,2617.98,3.66 we agree with that decision we'll give,2619.78,3.6 you a little bit more data or a little,2621.64,4.679 bit more computational horsepower that,2623.38,5.1 sort of stuff so you incentivize the,2626.319,4.5 behavior that you want to see,2628.48,5.22 number four corporate adoption so again,2630.819,4.981 like I said for everyone that's talked,2633.7,3.6 to me about it implementing Heroes to,2635.8,3.779 comparatives ultimately just creates,2637.3,4.799 better Solutions so if the best AI,2639.579,4.981 services and products are aligned,2642.099,4.861 the solution sells itself that's that,2644.56,4.08 can that could literally be the end of,2646.96,4.2 the conversation is that working with,2648.64,3.9 corporations whether it's the tech,2651.16,3.84 Giants providing these services or,2652.54,4.22 everyone else consuming those services,2655.0,5.579 to realize and develop those services so,2656.76,6.4 that all AI services are intrinsically,2660.579,4.74 aligned and of course open AI has done,2663.16,4.26 their best you know they have,2665.319,3.661 um they have their own internal research,2667.42,3.24 one problem though is that they're not,2668.98,3.42 sharing that research,2670.66,3.959 um so their their work on alignment is a,2672.4,3.959 total black box which means nobody else,2674.619,3.061 can,2676.359,3.781 um Can can duplicate it so we need an,2677.68,4.8 open source way so that everyone can,2680.14,4.679 duplicate alignment research and make,2682.48,4.44 sure that all their apis all their AIS,2684.819,4.921 are aligned and then corporations don't,2686.92,4.08 even need to think about it right,2689.74,3.06 because again corporations are like okay,2691.0,3.0 whatever whatever is going to make us,2692.8,3.36 the most money will do that and that's,2694.0,5.579 yeah so if we if we create a corporate,2696.16,6.54 ecosystem if an economic ecosystem in,2699.579,5.221 which the best option Finance actually,2702.7,4.68 is also the most aligned option problem,2704.8,4.74 solved now that's a big if,2707.38,4.5 there's a few other reasons though that,2709.54,4.86 adopting aligned AI services and systems,2711.88,4.5 would be good for corporations one,2714.4,3.48 public relations,2716.38,3.479 you know whatever whatever is popular in,2717.88,4.86 Vogue so for instance like LGBT rights,2719.859,4.98 super popular right now all the rage so,2722.74,4.079 guess what a lot of corporations are,2724.839,4.081 jumping on that bandwagon bandwagon,2726.819,4.081 mentality is good as long as it aligns,2728.92,4.74 on also something that is good employee,2730.9,5.1 satisfaction now obviously I think that,2733.66,4.14 employment conventional employment is,2736.0,2.76 going to be going the way of the,2737.8,2.819 dinosaurs by and large but for the,2738.76,3.78 employees that are there it really feels,2740.619,4.681 good to know that your company as part,2742.54,4.86 of a higher mission to make the world,2745.3,4.62 better for everyone so just gonna throw,2747.4,3.9 that out there and then finally,2749.92,3.179 stakeholder capitalism,2751.3,4.559 stakeholder capitalism is an is a a,2753.099,6.061 paradigm whereby it's not just you you,2755.859,5.281 the corporation and your customers it's,2759.16,3.78 everyone as a stakeholder so that's,2761.14,3.959 employees customer suppliers environment,2762.94,5.04 the rest of society so by adopting,2765.099,4.921 aligned AI that can also bring,2767.98,3.9 corporations in a line with stakeholder,2770.02,4.68 capitalism as that idea continues to,2771.88,4.14 develop,2774.7,3.899 oh this is a long video uh number five,2776.02,4.5 National regulations I already mentioned,2778.599,6.0 GDP GDP growth obviously AI is a gonna,2780.52,7.26 be a huge powerful economic engine for,2784.599,5.401 the foreseeable future so we need to,2787.78,4.44 make sure that as Nations you know try,2790.0,4.5 to maximize their GDP which they are all,2792.22,3.96 incentivized to do so that's fine I'm,2794.5,2.64 not going to tell them that they're,2796.18,1.919 wrong,2797.14,2.28 um I don't think that it's necessarily,2798.099,3.24 the best thing to optimize for but,2799.42,4.14 that's how the world works right now you,2801.339,4.201 can wish in one hand and you know you,2803.56,3.18 know what you can do on the other hand,2805.54,3.48 guess which one fills up,2806.74,4.02 um National Security so this is the,2809.02,4.02 biggest thing right the US's uh chips,2810.76,4.319 act where we you know did the the AI,2813.04,4.319 chips embargo against China right that's,2815.079,5.341 an example of the geopolitical game of,2817.359,4.381 chess that is going to be playing out,2820.42,4.86 for the foreseeable future around Ai and,2821.74,6.96 adversarial uses of AI so by working,2825.28,6.539 with nations in in line in alignment,2828.7,5.639 with their their national interests we,2831.819,5.401 can also work with them to adopt more,2834.339,5.421 aligned AI solicit Solutions and systems,2837.22,5.76 Democratic institutions so uh voter,2839.76,7.059 rights electric transparency Judicial,2842.98,7.08 Systems AI is going to impact every,2846.819,7.321 element every aspect of uh liberal,2850.06,6.6 Democratic societies including the,2854.14,4.38 agencies that the that those governments,2856.66,4.14 run on so by working with them to say,2858.52,5.04 here's how you can Implement AI to both,2860.8,5.46 save money and be a better Society to,2863.56,4.259 strengthen your Democratic institutions,2866.26,3.839 that will benefit everyone,2867.819,4.8 geopolitical influence ditto there's,2870.099,5.881 going to be things about trade for,2872.619,6.24 instance alliances all of those things,2875.98,4.98 are going to be impacted by AI which we,2878.859,4.621 need to study and we need to become the,2880.96,4.26 world experts on so that we can advise,2883.48,4.02 and consult properly and then finally,2885.22,4.139 sustainability which comes down to,2887.5,3.78 environmental challenges in the grand,2889.359,4.021 scheme of things I think that if we,2891.28,4.38 solve these other problems then by,2893.38,4.68 virtue of solving those problems around,2895.66,5.179 consensus we'll probably also figure out,2898.06,5.1 environmental control,2900.839,5.141 layer 6 International treaty I already,2903.16,5.159 mentioned um basically CERN but for AI,2905.98,4.56 so just a really quick recap of the,2908.319,4.621 benefits one membership and governance,2910.54,5.16 where all uh all nations are,2912.94,4.98 stakeholders and so they can join and,2915.7,3.48 make decisions collectively,2917.92,4.38 collaborative research again same exact,2919.18,4.2 thing that we already see with CERN,2922.3,3.5 shared resources and infrastructure,2923.38,4.5 Education and Training so this is,2925.8,3.94 another thing is there's probably going,2927.88,4.199 to be a shortfall of qualified Ai,2929.74,4.619 blockchain and and cognitive Architects,2932.079,4.921 for a while so by working together to,2934.359,4.201 make sure that we train up the people,2937.0,3.96 that we need to solve this problem that,2938.56,3.779 is something that International,2940.96,4.08 cooperation could do a lot for open,2942.339,4.5 science and knowledge sharing again that,2945.04,4.02 has been well established with,2946.839,3.541 um with some of these existing things,2949.06,4.98 International cooperation ditto huh see,2950.38,6.479 above statements and then finally uh,2954.04,4.74 Global consensus I already mentioned uh,2956.859,3.48 pretty much all of these,2958.78,3.42 um academic institutions we've got we've,2960.339,3.78 already got a few professors and,2962.2,4.08 students in the group so we've got a few,2964.119,3.901 lines in you know we've got feelers and,2966.28,4.92 fingers into um into the academic,2968.02,4.5 establishment,2971.2,2.879 um I've actually personally had probably,2972.52,3.18 a hundred different students reach out,2974.079,2.76 to me,2975.7,4.02 um either on patreon Discord or LinkedIn,2976.839,5.101 or Twitter and every time they ask me,2979.72,3.599 like Dave what should I what should I do,2981.94,4.08 and I'm like AI man like it's going that,2983.319,5.04 way if you care about the future like,2986.02,4.2 take a look at like some of the stuff,2988.359,3.96 that I've written and advocate for yours,2990.22,3.66 to comparatives research and they're,2992.319,3.841 like cool that's what I'll do,2993.88,5.64 um so you know because education is the,2996.16,6.9 future as as as many criticisms as I,2999.52,5.22 have of particularly American,3003.06,4.14 institutions universities are here,3004.74,4.079 they're here to stay they're important,3007.2,4.58 stakeholders in this entire conversation,3008.819,5.941 media engagement so this is this has to,3011.78,5.44 do with mainstream media this has to do,3014.76,5.16 with social media uh all of the above,3017.22,4.2 one of the things that we're working on,3019.92,3.24 is we're working on producing materials,3021.42,4.08 to make all this as accessible and,3023.16,4.74 shareable as possible so we're creating,3025.5,3.96 graphical slide decks we're creating,3027.9,4.62 educational materials I've got my videos,3029.46,4.74 um that sort of stuff because the more,3032.52,3.54 information that we get out there the,3034.2,4.2 easier it is to consume the more widely,3036.06,4.08 it's shared the better off we're all,3038.4,3.02 going to be,3040.14,4.26 next up is industry Partnerships again,3041.42,5.199 as I mentioned just a minute ago one of,3044.4,3.419 the things that we're that we're working,3046.619,3.48 on is publishing those open source,3047.819,5.28 standards advocating for startups and,3050.099,4.861 other companies to build and adopt,3053.099,5.161 aligned AI services and Pro products and,3054.96,5.46 just by working with them to say hey we,3058.26,4.559 recognize that your bottom line is the,3060.42,4.139 most important thing to companies let's,3062.819,3.78 make sure that that that that you,3064.559,4.681 implement and deploy these things in a,3066.599,3.72 way that doesn't have unintended,3069.24,3.599 negative consequences and then finally,3070.319,5.401 policy advocacy so this has to do with,3072.839,5.041 back going back every layer which is,3075.72,5.78 working with legislators lawyers,3077.88,5.939 and other groups you know whether it's,3081.5,4.9 think tanks whoever in order to better,3083.819,4.981 understand this stuff so an example of,3086.4,4.26 this is I've got a few meetings coming,3088.8,4.62 up later in May where I'll be meeting,3090.66,5.52 with people to help bring them up to,3093.42,4.919 speed with some of these ideas and help,3096.18,4.26 guide them as to like okay this is,3098.339,4.02 what's happening this is how it works,3100.44,4.08 and here's a here's an approach that we,3102.359,4.681 can take to make sure that it doesn't uh,3104.52,5.22 go uh belly side up,3107.04,4.62 now,3109.74,4.02 um we all have a good story,3111.66,4.98 for understanding this so in Avengers,3113.76,5.099 which I talk about this probably more,3116.64,4.5 than I should near the very end when,3118.859,4.861 Thanos said I am inevitable,3121.14,6.54 that is a fictional representation of,3123.72,7.92 Malik so the the idea is that Thanos was,3127.68,6.24 an Unstoppable destructive force that,3131.64,4.14 nobody wanted he wanted an outcome that,3133.92,4.56 nobody wanted but it seemed inevitable,3135.78,5.22 and he even said I am inevitable,3138.48,3.839 the snap,3141.0,3.78 the idea that there could be a moment in,3142.319,5.641 time that everything goes sideways,3144.78,5.94 everything goes wrong that is what,3147.96,5.399 Singularity or hard takeoff or whatever,3150.72,5.58 could represent the Infinity Stones,3153.359,5.641 think of those as the power of AI as as,3156.3,4.68 we get more and more AI capabilities,3159.0,4.8 it's like we're loading up our Gauntlet,3160.98,4.68 um the sacrifice that various people,3163.8,3.84 make like Tony Stark we have a lot of,3165.66,4.02 hard choices to make including just the,3167.64,3.78 investment that people like me and,3169.68,3.84 everyone in the community are making in,3171.42,5.1 terms of time and energy and the risks,3173.52,5.4 that we're taking in order to say hey we,3176.52,4.26 see this problem coming and we're going,3178.92,3.72 to try and do something about it,3180.78,4.74 in the story of undoing the snap the,3182.64,5.1 idea is that there is always hope that,3185.52,4.68 with the right people the right team and,3187.74,5.22 the right effort you can either avert,3190.2,6.119 disaster or undo disaster now obviously,3192.96,4.74 a lot of doomers say we don't get a,3196.319,3.721 do-over we don't get we we get one shot,3197.7,3.84 at this I don't know whether or not,3200.04,4.68 that's true but the idea is that we are,3201.54,6.6 barreling towards our end game right we,3204.72,5.82 have we must have the right people the,3208.14,3.9 right team,3210.54,4.98 um in a concerted Global effort in order,3212.04,6.18 to solve this problem safely and not,3215.52,4.2 just not just solve it like,3218.22,3.06 satisfactorily,3219.72,3.48 because again there's many possible,3221.28,3.6 outcomes I don't want a dystopian,3223.2,4.02 outcome any more than I want Extinction,3224.88,6.179 or collapse there's one possible outcome,3227.22,6.839 that is win-win that is Utopia and we,3231.059,4.621 got to thread that needle and we'll be,3234.059,3.361 working as hard as we can to make sure,3235.68,4.679 that that happens so this is The,3237.42,6.24 Avengers Assemble moment if you want to,3240.359,6.361 join this effort the link to apply is in,3243.66,5.34 the description of this video if you,3246.72,4.08 don't want to participate directly you,3249.0,3.96 can also support me on patreon I'm also,3250.8,3.66 happy to support you if you support me,3252.96,4.32 on patreon I have a private patreon,3254.46,5.159 Discord where I answer questions we,3257.28,4.799 actually just started having office,3259.619,5.881 hours Town Hall Days where all my,3262.079,4.74 patreon supporters can interact with,3265.5,4.14 each other and with me in real time if,3266.819,4.141 you've been laid off and you've got,3269.64,3.24 technical skills or political skills or,3270.96,3.84 communication skills or whatever,3272.88,3.9 maybe now's the time for you to join the,3274.8,4.74 effort if you're scared one of the one,3276.78,4.38 of the most powerful things that people,3279.54,4.14 have told me in in in the heuristics,3281.16,4.56 imperatives Discord is that for the,3283.68,4.56 first time in since forever they feel,3285.72,5.099 empowered to make a difference in the,3288.24,4.26 outcome that we're heading towards and,3290.819,3.421 if you're optimistic like me we also,3292.5,6.0 need that so Avengers assembled thank,3294.24,6.44 you,3298.5,2.18