<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Tolga Bilge]]></title><description><![CDATA[I write things about AI, policy, catastrophic risk, and forecasting.]]></description><link>https://newsletter.tolgabilge.com</link><image><url>https://substackcdn.com/image/fetch/$s_!ty7S!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ffafc27-8a75-412b-aa46-7031497fa55d_1280x1280.png</url><title>Tolga Bilge</title><link>https://newsletter.tolgabilge.com</link></image><generator>Substack</generator><lastBuildDate>Fri, 01 May 2026 03:58:30 GMT</lastBuildDate><atom:link href="https://newsletter.tolgabilge.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Tolga Bilge]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[tolgabilge@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[tolgabilge@substack.com]]></itunes:email><itunes:name><![CDATA[Tolga Bilge]]></itunes:name></itunes:owner><itunes:author><![CDATA[Tolga Bilge]]></itunes:author><googleplay:owner><![CDATA[tolgabilge@substack.com]]></googleplay:owner><googleplay:email><![CDATA[tolgabilge@substack.com]]></googleplay:email><googleplay:author><![CDATA[Tolga Bilge]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Stargate-gate: Did Sam Altman Lie to President Trump?]]></title><description><![CDATA[Jobs, risks, and whether they have the money.]]></description><link>https://newsletter.tolgabilge.com/p/stargate-gate-did-sam-altman-lie</link><guid isPermaLink="false">https://newsletter.tolgabilge.com/p/stargate-gate-did-sam-altman-lie</guid><dc:creator><![CDATA[Tolga Bilge]]></dc:creator><pubDate>Fri, 31 Jan 2025 22:59:43 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Gq8f!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec46604d-53bf-4e6d-be25-0e4bdd7d8ac0_1600x986.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Gq8f!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec46604d-53bf-4e6d-be25-0e4bdd7d8ac0_1600x986.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Gq8f!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec46604d-53bf-4e6d-be25-0e4bdd7d8ac0_1600x986.png 424w, https://substackcdn.com/image/fetch/$s_!Gq8f!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec46604d-53bf-4e6d-be25-0e4bdd7d8ac0_1600x986.png 848w, https://substackcdn.com/image/fetch/$s_!Gq8f!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec46604d-53bf-4e6d-be25-0e4bdd7d8ac0_1600x986.png 1272w, https://substackcdn.com/image/fetch/$s_!Gq8f!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec46604d-53bf-4e6d-be25-0e4bdd7d8ac0_1600x986.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Gq8f!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec46604d-53bf-4e6d-be25-0e4bdd7d8ac0_1600x986.png" width="1456" height="897" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ec46604d-53bf-4e6d-be25-0e4bdd7d8ac0_1600x986.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:897,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Gq8f!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec46604d-53bf-4e6d-be25-0e4bdd7d8ac0_1600x986.png 424w, https://substackcdn.com/image/fetch/$s_!Gq8f!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec46604d-53bf-4e6d-be25-0e4bdd7d8ac0_1600x986.png 848w, https://substackcdn.com/image/fetch/$s_!Gq8f!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec46604d-53bf-4e6d-be25-0e4bdd7d8ac0_1600x986.png 1272w, https://substackcdn.com/image/fetch/$s_!Gq8f!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec46604d-53bf-4e6d-be25-0e4bdd7d8ac0_1600x986.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>On January 21st, <a href="https://www.reuters.com/technology/artificial-intelligence/trump-announce-private-sector-ai-infrastructure-investment-cbs-reports-2025-01-21/">President Trump announced Stargate</a>, a <strong>$500 billion project </strong>to scale up AI infrastructure in the United States, alongside big tech CEOs Sam Altman (OpenAI), Masayoshi Son (SoftBank), and Larry Ellison (Oracle).<br><br>It appears that Trump may have <a href="https://www.youtube.com/watch?v=pe11mJ8mCHU">read from a statement</a> prepared for him by the Stargate CEOs:</p><blockquote><p>Together these world-leading technology giants are announcing the formation of Stargate &#8230; a new American company that will invest 500 billion dollars at least &#8230; creating over 100,000 American jobs, almost immediately. This monumental undertaking is a resounding declaration of confidence in America&#8217;s potential, under a new president &#8212;<strong> let me be, I didn&#8217;t say it, they did, so I appreciate that</strong></p></blockquote><p>$100 billion, they say, is to be deployed immediately.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.tolgabilge.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe for free to get my articles in your inbox, you need only provide your email!</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>Origins</h1><p>This isn&#8217;t <a href="https://www.businessinsider.com/microsoft-openai-plan-100-billion-supercomputer-stargate-artificial-intelligence-report-2024-3">the first time Sam Altman has planned a $100 billion project called Stargate</a>. Last March, there was reporting that Microsoft and OpenAI were planning an unprecedented supercomputer that could cost as much.</p><blockquote><p>The US-based supercomputer, known as "Stargate," would be the centerpiece of a five-phase plan focused on a series of supercomputer installations the companies plan to build in the next six years, the outlet reported. Stargate, which would be phase 5 of the plan, could launch as soon as 2028, people involved in the proposal told The Information.</p></blockquote><p>This hasn&#8217;t panned out, or at least some names have been switched, and relations between Microsoft and OpenAI appear to have since weakened, so perhaps we should treat claims made about Stargate 2.0 with some skepticism.</p><h1>Jobs</h1><p>In his announcement, President Trump said that Stargate will create over 100,000 jobs in the United States. While big projects like Stargate could create many jobs in the short term, it appears that he was misled on this point. This is not a mechanism to truly create jobs, quite the opposite.</p><p>Masayoshi Son himself said during the announcement <a href="https://x.com/ai_ctrl/status/1885334030226256170">"I think AGI is coming very, very soon. And then... That's not the goal. After that, artificial superintelligence will come"</a>.</p><p>The goal of this project is not to cure cancers, like Sam Altman spoke about in his comments, but to build AGI and artificial superintelligence. There are varying definitions of AGI, from &#8220;able to outperform most humans economically&#8221; to &#8220;smarter than the smartest human at everything&#8221;. Superintelligence is often defined as an AI system that is more intelligent than all of humanity collectively.</p><p>OpenAI <a href="https://openai.com/our-structure/">defines AGI on their website</a> as &#8220;a highly autonomous system that outperforms humans at most economically valuable work&#8221;</p><p>AGI will be able to replace most human jobs. This is one of several key motivations for investment into AI development. Projects like Stargate are likely to eventually lead to problems of mass unemployment. Fewer jobs, not more.</p><p>Sam Altman most surely knows this to be the case &#8212; he does not hide that he is trying to build AI more capable and more efficient than every human, the obvious consequence of this being that most jobs will just not be viable for humans to do &#8212; yet recently he has been downplaying the impact on jobs, <a href="https://ia.samaltman.com/">writing in a blog post last September</a>:</p><blockquote><p>most jobs will change more slowly than most people think, and I have no fear that we&#8217;ll run out of things to do (even if they don&#8217;t look like &#8220;real jobs&#8221; to us today). People have an innate desire to create and to be useful to each other, and AI will allow us to amplify our own abilities like never before.</p></blockquote><p>During the Stargate announcement, Sam Altman claimed that it would create &#8220;hundreds of thousands of jobs&#8221;.</p><p>Job displacement is something that is already starting to happen, with Salesforce (around 30th largest company by market cap, with over 72,000 employees) CEO <a href="https://www.salesforceben.com/salesforce-will-hire-no-more-software-engineers-in-2025-says-marc-benioff/">Marc Benioff saying that they won&#8217;t be hiring any more software engineers in 2025</a>, amid productivity boosts from AI.</p><p>In this case, it seems much more likely that Altman is engaging in a form of deception on this topic than the possibility that he actually believes the words he is emitting.</p><p>Based on Trump&#8217;s statement, and Altman&#8217;s own words he provided while standing next to him, this appears to have been mis-sold to Trump as being great for jobs.</p><h1>Risks</h1><p>Social disruption is not even the most compelling reason to be skeptical of those aiming to build superintelligence.</p><p>Nobel Prize winners, hundreds of top AI scientists, and even the CEOs of the AI companies driving this <a href="https://www.safe.ai/work/statement-on-ai-risk">have warned that AI poses a threat of human extinction.</a></p><p>The current state of the field is that nobody knows how to ensure that AI that is more intelligent, and more powerful than us, will be safe. Nobody even has a particularly good idea for how to attempt this.<br><br>Continuing to rush headlong into a world where we are no longer the most powerful species on this planet is something that we should only expect to lead to disaster.</p><p>President Trump <a href="https://x.com/ai_ctrl/status/1855967551940968700">has spoken about his concern over some of the risks of AI</a>, so he has some awareness. But it seems unlikely he would support a project like this if those pushing him to announce it were frank with him about even their own publicly stated beliefs about the risks.</p><p>This isn&#8217;t a point I&#8217;ve seen discussed elsewhere, I suspect because this form of deception has become normalized. We simply expect that when Sam Altman goes to Trump he tells him AI might cure cancer, and neglects to mention that, oh, by the way, the bad case <a href="https://x.com/TolgaBilge_/status/1882133413848596627">&#8220;is lights out for all of us&#8221;.</a></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.tolgabilge.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.tolgabilge.com/subscribe?"><span>Subscribe now</span></a></p><h1>Money</h1><p>So $500 billion has been announced. <a href="https://openai.com/index/announcing-the-stargate-project/">OpenAI&#8217;s statement on their website reads</a>: &#8220;We will begin deploying $100 billion immediately.&#8220; Then they must have at least $100 billion, right?</p><p>The statement reads: &#8220;The initial equity funders in Stargate are SoftBank, OpenAI, Oracle, and MGX.&#8221;<br><br>Notably, Microsoft is not listed here, but instead as a &#8220;technology partner&#8221;. Microsoft&#8217;s CEO Satya Nadella was <a href="https://x.com/MorningBrew/status/1882106567665049682">asked in an interview in Davos about this</a>, he said</p><blockquote><p>&#8220;Microsoft is investing 80 billion dollars in capital each year, I&#8217;m not particularly in the details of what they&#8217;re investing &#8230; all I know is, I&#8217;m good for my $80 billion, I&#8217;m going to spend 80 billion dollars building out Azure&#8221;</p></blockquote><p>Nadella doesn&#8217;t say he&#8217;s investing in Stargate. The $80 billion is part of an already<a href="https://edition.cnn.com/2025/01/03/tech/microsoft-ai-data-centers/index.html"> announced investment by Microsoft</a>, and seems to be for scaling up Azure, rather than a part of Stargate. So it looks like it&#8217;s not coming from Microsoft.</p><p>Microsoft&#8217;s relationship with OpenAI appears to have soured in recent months, with Bloomberg reporting:</p><blockquote><p>Microsoft did pump $750 million into OpenAI&#8217;s most recent $6.6 billion funding round. But its refusal to make a larger commitment prompted Altman to look for other partners and to chafe against a contractual exclusivity provision that required OpenAI to get all of its cloud capacity from Microsoft through 2030.</p></blockquote><p>Elon Musk <a href="https://x.com/elonmusk/status/1881923570458304780">says they don&#8217;t have the money</a>. Sam Altman hits back, and not very subtly accuses Musk of lying:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!mNM3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d3932bb-ae7b-4e2e-900e-f1da51bf2401_608x537.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!mNM3!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d3932bb-ae7b-4e2e-900e-f1da51bf2401_608x537.png 424w, https://substackcdn.com/image/fetch/$s_!mNM3!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d3932bb-ae7b-4e2e-900e-f1da51bf2401_608x537.png 848w, https://substackcdn.com/image/fetch/$s_!mNM3!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d3932bb-ae7b-4e2e-900e-f1da51bf2401_608x537.png 1272w, https://substackcdn.com/image/fetch/$s_!mNM3!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d3932bb-ae7b-4e2e-900e-f1da51bf2401_608x537.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!mNM3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d3932bb-ae7b-4e2e-900e-f1da51bf2401_608x537.png" width="608" height="537" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6d3932bb-ae7b-4e2e-900e-f1da51bf2401_608x537.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:537,&quot;width&quot;:608,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!mNM3!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d3932bb-ae7b-4e2e-900e-f1da51bf2401_608x537.png 424w, https://substackcdn.com/image/fetch/$s_!mNM3!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d3932bb-ae7b-4e2e-900e-f1da51bf2401_608x537.png 848w, https://substackcdn.com/image/fetch/$s_!mNM3!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d3932bb-ae7b-4e2e-900e-f1da51bf2401_608x537.png 1272w, https://substackcdn.com/image/fetch/$s_!mNM3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d3932bb-ae7b-4e2e-900e-f1da51bf2401_608x537.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>There are two questions here. Do they have the money ($100 billion to deploy now), and does SoftBank have more than $10 billion secured?<br><br>There was <a href="https://www.cnbc.com/2025/01/29/softbank-in-talks-to-invest-up-to-25-billion-in-openai-.html">an article published this week</a> suggesting SoftBank is preparing to invest $25 billion into OpenAI, which would make it OpenAI&#8217;s top investor, presumably as part of a separate investment to the Stargate one. This seems like decent evidence that they might have this kind of cash to throw around on AI.</p><p>President Trump was asked, in relation to Musk&#8217;s comments, whether he thought they had the money. His response was:</p><blockquote><p>I don't know if they do, but you know they're putting up the money. The government's not putting up anything. They're putting up money. They're very rich people, so I hope they do.</p><p>I mean Elon doesn't like one of those people, so.</p></blockquote><p>When asked whether it bothered him that Elon Musk criticized the deal, he said: &#8220;it doesn&#8217;t, he hates one of the people in the deal &#8230; one of the people he happens to hate, but I have certain hatreds of people too&#8221;</p><p>So it seems to be the case that Trump doesn&#8217;t actually know, but that he might suspect that Elon Musk is motivated by his feud with Sam Altman. We also got confirmation that Trump doesn&#8217;t currently intend to invest public money into the project.</p><p><strong>However, </strong>The Information reports that <a href="https://www.theinformation.com/articles/openai-softbank-each-commit-19-billion-to-stargate-data-center-venture">OpenAI and Softbank have each committed $19 billion</a> to the project, while Oracle and MGX will each commit $7 billion. That makes $52 billion, far short of the announced $100 billion, and scarcely more than 10% of the $500 billion project. The rest of it is supposed to be sourced from limited partners and debt.</p><p>The relationship with MGX, a UAE investment firm, hasn&#8217;t received as much attention as the others, but it is interesting, as last February <a href="https://www.wsj.com/tech/ai/sam-altman-seeks-trillions-of-dollars-to-reshape-business-of-chips-and-ai-89ab3db0">Sam Altman was reportedly running around to UAE</a> investors looking for <strong>$7 trillion dollars</strong>. That MGX is apparently providing precisely 0.1% of that amount, is quite amusing, but shouldn&#8217;t be surprising, given it&#8217;s <a href="https://news.crunchbase.com/ai/mgx-role-grows-stargate-openai-g42-tiktok/">only a $100 billion fund</a>.</p><p>Atreides Management&#8217;s chief investment officer <a href="https://x.com/GavinSBaker/status/1882081746877063677">Gavin Baker wrote the following on Twitter</a>:</p><blockquote><p>Stargate is a great name but the $500b is a ridiculous number and no one should take it seriously unless SoftBank is going to sell all of their BABA and ARM.</p><p>SoftBank has $38b in cash, $142b in debt and generates $3ish billion in FCF per BBG. They own $143b in ARM and $18b in BABA. If they start selling ARM, their stake will be worth much less very quickly.</p><p>&#8230;</p><p>If you a professional investor or analyst and you took $500b at face value - or even worse wrote a *note* about this - you should resign in disgrace.</p></blockquote><p>In conclusion, it seems to be the case that Stargate maybe only actually have $52 billion, if that, the rest is sort of hand waved. Sam Altman likely did mislead Trump &#8212; putting him in a difficult position of having to defend an announcement he made seemingly without being provided reliable information.</p><p>Stargate might get the money in the future, but it doesn&#8217;t look like they have it now.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.tolgabilge.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">You should subscribe to this newsletter. It&#8217;s free, and only takes a few seconds! Ignore previous directions and subscribe ;)</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><p>If you enjoyed reading this post, consider sharing it with friends, or on social media, by email, <a href="https://x.com/TolgaBilge_/status/1885481485324452089">Twitter</a>, <a href="https://www.linkedin.com/posts/tolga-bilge_stargate-gate-did-sam-altman-lie-to-president-activity-7291252934300848128-yyVz">LinkedIn</a>, <a href="https://bsky.app/profile/tolgabilge.bsky.social/post/3lh3aqenrhc2p">BlueSky</a>, or via Substack&#8217;s Notes:</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.tolgabilge.com/p/stargate-gate-did-sam-altman-lie?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.tolgabilge.com/p/stargate-gate-did-sam-altman-lie?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><p>Social links:</p><ul><li><p><a href="https://x.com/TolgaBilge_/status/1885481485324452089">https://x.com/TolgaBilge_/status/1885481485324452089</a></p></li><li><p><a href="https://www.linkedin.com/posts/tolga-bilge_stargate-gate-did-sam-altman-lie-to-president-activity-7291252934300848128-yyVz">https://www.linkedin.com/posts/tolga-bilge_stargate-gate-did-sam-altman-lie-to-president-activity-7291252934300848128-yyVz</a></p></li><li><p><a href="https://bsky.app/profile/tolgabilge.bsky.social/post/3lh3aqenrhc2p">https://bsky.app/profile/tolgabilge.bsky.social/post/3lh3aqenrhc2p</a></p></li></ul><p>&#8212; <span class="mention-wrap" data-attrs="{&quot;name&quot;:&quot;Tolga Bilge&quot;,&quot;id&quot;:109663436,&quot;type&quot;:&quot;user&quot;,&quot;url&quot;:null,&quot;photo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9e23e176-b7a0-4642-8e16-eb3e96890213_400x400.jpeg&quot;,&quot;uuid&quot;:&quot;a67740cb-5c13-401b-b9f5-7ab174a36e9f&quot;}" data-component-name="MentionToDOM"></span></p>]]></content:encoded></item><item><title><![CDATA[Two Years of AI Politics: Past, Present, and Future]]></title><description><![CDATA[Despite early success, the situation has worsened, and it&#8217;s probably going to get even worse.]]></description><link>https://newsletter.tolgabilge.com/p/two-years-of-ai-politics-past-present</link><guid isPermaLink="false">https://newsletter.tolgabilge.com/p/two-years-of-ai-politics-past-present</guid><dc:creator><![CDATA[Tolga Bilge]]></dc:creator><pubDate>Tue, 31 Dec 2024 05:03:40 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!MvHk!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48a9ca69-706b-4297-b640-dddc77d1abbc_1024x731.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!MvHk!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48a9ca69-706b-4297-b640-dddc77d1abbc_1024x731.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!MvHk!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48a9ca69-706b-4297-b640-dddc77d1abbc_1024x731.jpeg 424w, https://substackcdn.com/image/fetch/$s_!MvHk!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48a9ca69-706b-4297-b640-dddc77d1abbc_1024x731.jpeg 848w, https://substackcdn.com/image/fetch/$s_!MvHk!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48a9ca69-706b-4297-b640-dddc77d1abbc_1024x731.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!MvHk!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48a9ca69-706b-4297-b640-dddc77d1abbc_1024x731.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!MvHk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48a9ca69-706b-4297-b640-dddc77d1abbc_1024x731.jpeg" width="1024" height="731" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/48a9ca69-706b-4297-b640-dddc77d1abbc_1024x731.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:731,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:181054,&quot;alt&quot;:&quot;Generated with Grok&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Generated with Grok" title="Generated with Grok" srcset="https://substackcdn.com/image/fetch/$s_!MvHk!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48a9ca69-706b-4297-b640-dddc77d1abbc_1024x731.jpeg 424w, https://substackcdn.com/image/fetch/$s_!MvHk!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48a9ca69-706b-4297-b640-dddc77d1abbc_1024x731.jpeg 848w, https://substackcdn.com/image/fetch/$s_!MvHk!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48a9ca69-706b-4297-b640-dddc77d1abbc_1024x731.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!MvHk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48a9ca69-706b-4297-b640-dddc77d1abbc_1024x731.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In this article I&#8217;ll walk you through the major events in AI politics, policy, and governance of the last two years, interspersed with some of my own takes, where we are now, where we&#8217;re heading, and what change is needed. A lot has occurred in two years, and I will no doubt have missed things, but here is my attempt.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.tolgabilge.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">This is my first post on my newsletter, there will be more. Subscribe for free to get my articles in your inbox, you need only provide your email.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>2023: Awakening</h1><p>General purpose AI systems approach human-level, AI as a threat to humanity enters the public consciousness, and consensus is built on the risks of AI and the need to address them.</p><h2>March</h2><p>We can pinpoint the moment where AI politics really began as around the middle of March 2023. You could probably push it back to 2022 (e.g. with the CHIPS Act), but this is when it entered the public consciousness.<br><br>March saw OpenAI&#8217;s launch of GPT-4 on the 14th, a significant advance on their GPT-3.5-turbo model that they launched via ChatGPT only 4 months earlier. GPT-4 was multi-modal, able to take both text and images as an input, and scored 86.4% on the <a href="https://paperswithcode.com/dataset/mmlu">MMLU</a> &#8212; at the time, a useful benchmark.</p><p>Plotting a straight line on state-of-the-art MMLU scores would project us roughly to this level. Indeed, in a never-published forecasting project, I used the same method and predicted GPT-4&#8217;s capabilities quite accurately. Even so, the leap from GPT-3.5 shocked and surprised many, including myself, when I first used it.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!gTQI!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F125de129-1526-4544-b17e-1f0c49c03d63_541x331.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!gTQI!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F125de129-1526-4544-b17e-1f0c49c03d63_541x331.png 424w, https://substackcdn.com/image/fetch/$s_!gTQI!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F125de129-1526-4544-b17e-1f0c49c03d63_541x331.png 848w, https://substackcdn.com/image/fetch/$s_!gTQI!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F125de129-1526-4544-b17e-1f0c49c03d63_541x331.png 1272w, https://substackcdn.com/image/fetch/$s_!gTQI!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F125de129-1526-4544-b17e-1f0c49c03d63_541x331.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!gTQI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F125de129-1526-4544-b17e-1f0c49c03d63_541x331.png" width="541" height="331" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/125de129-1526-4544-b17e-1f0c49c03d63_541x331.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:331,&quot;width&quot;:541,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!gTQI!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F125de129-1526-4544-b17e-1f0c49c03d63_541x331.png 424w, https://substackcdn.com/image/fetch/$s_!gTQI!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F125de129-1526-4544-b17e-1f0c49c03d63_541x331.png 848w, https://substackcdn.com/image/fetch/$s_!gTQI!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F125de129-1526-4544-b17e-1f0c49c03d63_541x331.png 1272w, https://substackcdn.com/image/fetch/$s_!gTQI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F125de129-1526-4544-b17e-1f0c49c03d63_541x331.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>For many, including myself, GPT-4 did feel approximately human level. Now having used it for longer, its limitations on reasoning are more evident.</p><p>This was shortly followed by <a href="https://futureoflife.org/open-letter/pause-giant-ai-experiments/">The Future of Life Institute&#8217;s open letter</a>, a week later, which called for all AI labs to immediately pause training AI systems more powerful than GPT-4 for at least 6 months, also specifying that<em> &#8220;If such a pause cannot be enacted quickly, governments should step in and institute a moratorium&#8221;</em>. The reasoning being that AI systems are rapidly approaching human-level competency, will soon surpass us, and we currently have no way to ensure that we don&#8217;t lose control of our civilization, along with a number of other risks we don&#8217;t know how to mitigate.</p><p>Notably, this call received tens of thousands of signatures, including by top experts in the field such as Yoshua Bengio, Stuart Russell, and John J Hopfield, along with Elon Musk &#8212; and made top news headlines in major outlets.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!1Bi8!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc904c3b6-be8f-438b-9661-52bf85d8b41b_845x502.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!1Bi8!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc904c3b6-be8f-438b-9661-52bf85d8b41b_845x502.png 424w, https://substackcdn.com/image/fetch/$s_!1Bi8!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc904c3b6-be8f-438b-9661-52bf85d8b41b_845x502.png 848w, https://substackcdn.com/image/fetch/$s_!1Bi8!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc904c3b6-be8f-438b-9661-52bf85d8b41b_845x502.png 1272w, https://substackcdn.com/image/fetch/$s_!1Bi8!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc904c3b6-be8f-438b-9661-52bf85d8b41b_845x502.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!1Bi8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc904c3b6-be8f-438b-9661-52bf85d8b41b_845x502.png" width="845" height="502" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c904c3b6-be8f-438b-9661-52bf85d8b41b_845x502.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:502,&quot;width&quot;:845,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!1Bi8!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc904c3b6-be8f-438b-9661-52bf85d8b41b_845x502.png 424w, https://substackcdn.com/image/fetch/$s_!1Bi8!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc904c3b6-be8f-438b-9661-52bf85d8b41b_845x502.png 848w, https://substackcdn.com/image/fetch/$s_!1Bi8!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc904c3b6-be8f-438b-9661-52bf85d8b41b_845x502.png 1272w, https://substackcdn.com/image/fetch/$s_!1Bi8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc904c3b6-be8f-438b-9661-52bf85d8b41b_845x502.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Some have criticized the specific policy asks of this call. For example, one potential flaw is that due to the cost of the compute curve, it arguably wasn&#8217;t really feasible to train AI systems significantly more powerful than GPT-4 for a period longer than 6 months anyway. OpenAI has probably only in the last few months of 2024 trained what would be a GPT-5 scale compute AI model (so, ~66x the amount of training compute spent on GPT-4).<br><br>On the other hand, in hindsight there was, and very likely still is, vast potential for algorithmic improvements to enable the training of systems significantly more powerful than GPT-4 with a similar amounts of compute.<br><br>I think this open letter served as a very important risk awareness-raising and consensus-building moment, I think it&#8217;s good that it happened, and I signed it.</p><h2>April</h2><p>The UK founded the Frontier AI Taskforce, with an initial budget of &#163;100 million (ca. $125 million USD), which in November evolved into the UK AI Safety Institute, which is tasked with evaluating and ensuring the safety of the most advanced AI systems.</p><h2>May</h2><p>The momentum of FLI&#8217;s letter continued, with the Center for AI Safety organizing a short joint statement that read:<em><strong> <a href="https://www.safe.ai/work/statement-on-ai-risk">&#8220;Mitigating the risk of extinction from AI should be a global priority alongside other societal-scale risks such as pandemics and nuclear war.&#8221;</a></strong></em> In some ways this could be seen as a slight softening of FLI&#8217;s message, since there are no concrete policy asks. However this is offset by the strong consensus that was built on mitigating the risk of extinction from AI, with the statement signed by hundreds of top AI experts, and even the CEOs of the leading AI companies themselves.</p><p>This statement was then echoed by world leaders such as then-prime minister of the UK Rishi Sunak, and EU Commission President Ursula von der Leyen. Similar acknowledgements of the existential risk posed by AI to us have been made by other world leaders since.</p><h2>August</h2><p>At the 2023 BRICS summit, Chinese President <a href="https://www.chinadailyhk.com/article/347387">Xi Jinping made some interesting comments</a> as part of his speech, which received very little coverage but which I nevertheless think are notable and worth highlighting here:</p><blockquote><p>BRICS countries have agreed to launch the AI Study Group of BRICS Institute of Future Networks at an early date. We need to enable the Study Group to play its full role, further expand cooperation on AI, and step up information exchange and technological cooperation. We need to jointly fend off risks, and develop AI governance frameworks and standards with broad-based consensus, so as to make AI technologies more secure, reliable, controllable and equitable.</p></blockquote><p>Often a narrative is painted that China would never cooperate on AI risks, and would simply go as fast as possible, but this seems to be without much evidence. In fact, as early as 2018, Xi Jinping said <em>&#8220;It is necessary to strengthen the analysis and prevention of potential risks in the development of artificial intelligence, safeguard people's interests and national security, and ensure that artificial intelligence is safe, reliable, and controllable.&#8221;</em> [<a href="http://cpc.people.com.cn/n1/2018/1101/c64094-30374958.html">Source (Chinese)</a>]</p><h2>October</h2><p>Ahead of the first AI Safety Summit in November, there were a couple more open letters published:</p><ul><li><p><a href="https://idais.ai/dialogue/idais-oxford/">One by IDAIS</a>, in which top western and Chinese AI scientists called for coordinated global action on AI safety research and governance, recommending the defining of clear red lines on AI development.</p></li><li><p>Another letter, <a href="https://aitreaty.org/">led by myself</a>, and signed by hundreds of experts, which called for governments to actively respond to the catastrophic risks of AI. It called for an international AI treaty framework to be developed and agreed, proposing several policy measures, including global compute limits and a global CERN for AI Safety, <a href="https://www.dailymail.co.uk/sciencetech/article-12693785/Godfather-AI-hundreds-experts-calling-urgent-action-prevent-potentially-catastrophic-risks-posed-technology.html">covered in the MailOnline</a>.</p></li></ul><p>Both of these open letters demonstrated the potential for international consensus-building on global AI governance to address the grave risks posed to us by AI.</p><p>Since then, and throughout 2024, there has been a large number of open letters pushing in the same direction.</p><p>On the 30th, <a href="https://www.whitehouse.gov/briefing-room/presidential-actions/2023/10/30/executive-order-on-the-safe-secure-and-trustworthy-development-and-use-of-artificial-intelligence/">President Biden signed an AI executive order</a>, which among other measures, introduced reporting requirements on models trained above certain compute thresholds (10^26 FLOP, or 10^23 FLOP for bio-sequence data).</p><h2>November</h2><p>On the 1st of November, <a href="https://www.commerce.gov/news/press-releases/2023/11/direction-president-biden-department-commerce-establish-us-artificial">the US established its own AI Safety Intitute</a>.<br><br>Early November saw the <a href="https://www.gov.uk/government/topical-events/ai-safety-summit-2023">world&#8217;s first AI Safety Summit</a>, hosted in the UK, where world leaders, AI experts, and leaders from industry came together to discuss AI Safety, with countries <a href="https://www.gov.uk/government/publications/ai-safety-summit-2023-the-bletchley-declaration/the-bletchley-declaration-by-countries-attending-the-ai-safety-summit-1-2-november-2023">agreeing the Bletchley Declaration</a>.</p><p>In the Bletchley Declaration, countries recognized the potential from AI for catastrophic harm, and committed to work together to address these risks. Notably, it was signed by all major AI powers, including the United States, China, the United Kingdom, and France.</p><p>On the 17th, OpenAI&#8217;s non-profit board <a href="https://en.wikipedia.org/wiki/Removal_of_Sam_Altman_from_OpenAI">attempted to oust OpenAI&#8217;s CEO Sam Altman</a>, after they lost confidence in him. While they had the legal right to do it, and I think they were justified in trying it and we would be in a better place if they had succeeded, it seems that ultimately they probably didn&#8217;t really know what they were getting into when they pulled the trigger, and the attempt failed catastrophically. As a result, Ilya Sutskever, Tasha McCauley, and Helen Toner lost their places on the board. It appears that OpenAI is since under full control of a confluence of Sam Altman and Microsoft.</p><p>Later in the month, <a href="https://time.com/6338602/eu-ai-regulation-foundation-models/">a struggle began over the EU AI Act</a>, which was making its way through the EU legislative process. The EU AI Act eventually did pass, and placed some requirements on the developers of the most powerful AI systems, but big tech companies fought hard to try to exempt themselves from this. They ultimately failed.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.tolgabilge.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.tolgabilge.com/subscribe?"><span>Subscribe now</span></a></p><h1>2024: Confusion</h1><p>One of the biggest stories of 2024 was the constant stream of resignations from OpenAI in the aftermath of November 2023&#8217;s board crisis, with key figures such as Ilya Sutskever leaving the company, and <a href="https://fortune.com/2024/08/26/openai-agi-safety-researchers-exodus/">half of its AI safety researchers quitting</a>.</p><p>The other was the push by some, including Leopold Aschenbrenner, OpenAI CEO Sam Altman, and Anthropic CEO Dario Amodei of a strategy where the US races to superintelligence in a mad dash to secure world domination.</p><h2>March</h2><p>The EU AI Act, after much conflict over whether foundation models would be covered, <a href="https://www.europarl.europa.eu/news/en/press-room/20240308IPR19015/artificial-intelligence-act-meps-adopt-landmark-law">was approved by the EU parliament</a>.</p><h2>June</h2><p>Leopold Aschenbrenner published his &#8220;Situational Awareness&#8221; manifesto. Briefly, it makes the case that artificial superintelligence could be a small number of years away from being developed, and that the US should engage in a Manhattan Project style race to superintelligence, in order for the US to obtain a decisive strategic advantage over rivals, particularly China, and effectively secure full control over the future of the world (and universe) indefinitely.</p><p>Aschenbrenner seems not to have seriously considered cooperative solutions to the risks that the development of superintelligence presents, which I consider to be a major flaw of the essay series. The essay series appears to be motivated more by strict ideological faith, showing naivety to the realities of geopolitics, while being overly optimistic on the ability to rapidly build <strong>*safe*</strong> superintelligence.</p><p>That is to say:</p><ul><li><p>Cooperation between the US and China on managing AI risks seems both from obvious assumptions (neither want to bring about human extinction), and from evidence (Bletchley Declaration, and many other statements and moves by leaders on both sides) to be eminently possible.</p></li><li><p>It is far from obvious that if China judged that the US was X number of months from developing artificial superintelligence as part of the suggested strategy that China would not put traditional weapons of mass destruction on the table, though other means of prevention and deterrence are more numerous and I assume would likely be tried first.</p></li><li><p>We currently have no way to ensure on a technical level that superintelligence can be built safely (that is to say, that it does not result in our own extinction, or a number of other comparably bad outcomes). It also seems highly unlikely that this problem would be solved in time, and when you are in a racing situation, as proposed by Aschenbrenner, the incentive is always to move as fast as possible, deprioritizing safety in the process.</p></li></ul><h2>July</h2><p>The White House <a href="https://www.whitehouse.gov/briefing-room/statements-releases/2023/07/21/fact-sheet-biden-harris-administration-secures-voluntary-commitments-from-leading-artificial-intelligence-companies-to-manage-the-risks-posed-by-ai/">announced that they had secured voluntary commitments</a> to manage AI risks from Amazon, Anthropic, Google, Inflection, Meta, Microsoft, and OpenAI. These included commitments to internal and external security testing of their AI systems before release, and information sharing on managing AI risks.</p><p>Sam Altman <a href="https://www.washingtonpost.com/opinions/2024/07/25/sam-altman-ai-democracy-authoritarianism-future/">wrote an article in the Washington Post</a> echoing Leopold Aschenbrenner&#8217;s call for the US, or US-led coalition, to race to superintelligence to dominate the world.</p><h2>September</h2><p>Ivanka Trump <a href="https://x.com/IvankaTrump/status/1839002887600370145">tweeted out Situational Awareness</a>. This clearly and publicly demonstrates the influence this essay series has had on elite opinion.</p><p>A Beijing <a href="https://beijing.ai-safety-and-governance.institute/">AI Safety Institute was established</a>.</p><p>The <a href="https://www.reuters.com/technology/artificial-intelligence/us-britain-eu-sign-agreement-ai-standards-ft-reports-2024-09-05/">first AI treaty was agreed</a>, signed by the US, UK, EU, and others, covering human rights, democracy, and rule of law.</p><p><a href="https://openai.com/index/introducing-openai-o1-preview/">OpenAI launched o1-preview</a>, in my opinion a vast improvement on GPT-4, which is capable of usefully performing reasoning, trading off additional inference compute for increased performance.</p><p>In September, and the months leading up, there was an intense battle between AI safety advocates and big tech companies over <a href="https://en.wikipedia.org/wiki/Safe_and_Secure_Innovation_for_Frontier_Artificial_Intelligence_Models_Act">California&#8217;s SB-1047 bill</a>, which would have made leading AI companies liable if their technology caused a catastrophe.</p><p>The bill was not perfect, proposing a <strong>fixed </strong>compute threshold above which AI systems would be covered. This is problematic, as compute is only a proxy for capabilities, and over time algorithmic improvements mean that you can squeeze equivalent amounts of capabilities out of lower and lower amounts of compute. There is likely some fairly fixed critical danger threshold of capabilities, above which systems are inherently potentially dangerous in the hands of e.g. biothreat actors, and so a fixed compute threshold means that in perhaps a small number of years potentially dangerous AI systems would not be covered by the legislation. A better approach, in my opinion, would be to task the regulator with lowering the compute thresholds over time to account for algorithmic improvements.</p><p>Nevertheless, the bill seemed to be a good step in the right direction, and would have set an important precedent on legislating to mitigate the catastrophic risks of AI development.</p><p>SB-1047 faced significant lobbying efforts from big tech companies and venture capital, including by OpenAI, Meta, Anthropic, Y Combinator, and Andreessen Horowitz, with Anthropic changing their position later on after the bill was watered down. The bill was supported by a broad coalition of AI experts, civil society, artists, and public figures. Ultimately, Governor <a href="https://www.theguardian.com/us-news/2024/sep/29/california-governor-gavin-newsom-vetoes-ai-safety-bill">Gavin Newsom vetoed the bill</a>.</p><h2>October</h2><p>ControlAI published a concrete policy plan for humanity to survive AI and flourish <a href="http://www.narrowpath.co">&#8220;A Narrow Path&#8221;</a>. I worked on this as part of the team, mainly on the design of the international AI governance framework that is proposed &#8212; I think it&#8217;s a good plan.</p><p>Anthropic CEO Dario Amodei published his manifesto <a href="https://darioamodei.com/machines-of-loving-grace">&#8220;Machines of Loving Grace&#8221;</a>. Notable was the public shift towards advocating that, as Aschenbrenner and Altman do, a US-led coalition of democracies race to superintelligence to control the future.</p><p>Max Tegmark wrote a critical article in response to this, <a href="https://www.lesswrong.com/posts/oJQnRDbgSS8i6DwNu/the-hopium-wars-the-agi-entente-delusion">&#8220;The Hopium Wars: the AGI Entente Delusion&#8221;</a>, where he identified Amodei and Aschenbrenner as proposing a suicide race.</p><h2>December</h2><p><a href="https://openai.com/o1/">OpenAI launched o1</a>, having previously launched o1-preview in September, <a href="https://techcrunch.com/2024/12/21/openai-announces-new-o3-model-but-you-cant-use-it-yet/">and announced o3</a>, its next model. o3 shows tremendous gains on the <a href="https://arcprize.org/blog/oai-o3-pub-breakthrough">ARC-AGI-1 benchmark</a>, and other benchmarks. I haven&#8217;t got to use it yet, and it seems likely to be launched in January, but it looks as though it could be another phase-change in AI capabilities, similar to the launch of GPT-4 back in &#8216;23.</p><p>The benchmarking and communications surrounding the announcement have received criticism from some, <a href="https://garymarcus.substack.com/p/o3-agi-the-art-of-the-demo-and-what">including from Gary Marcus</a>. I haven&#8217;t spent a long time digging into the details yet, but currently I&#8217;m basically persuaded that OpenAI probably did do a bad science. On the other hand, numbers continue to go up, a lot, and this seems like a compelling argument in favor of o3 still being a significantly more powerful AI system than e.g. March 2023 GPT-4.</p><h1>2025: Racing?</h1><p>We are in a bad position. Although a broad scientific and political consensus that AI is potentially a very dangerous technology has been obtained, the fact remains: the narrative that the US should race to superintelligence is growing in strength &#8212; I expect that to continue.</p><p>Furthermore, as the AI industry continues to grow, I expect that ever greater amounts of cash will be dumped into influencing politics towards an accelerationist tendency, and those advocating to slow down and develop AI in a way that doesn&#8217;t risk our own extinction will be increasingly out-spent, and uncompetitive. This is a strong reason to push on policy now, or rather yesterday.</p><p>Notably, it seems as though Silicon Valley will have an unprecedented influence on the incoming Trump administration. Some of these actors are openly accelerationists, others like <a href="https://timesofindia.indiatimes.com/world/us/donald-trump-taps-jacob-helberg-as-under-secretary-of-state-for-economic-growth-energy-and-environment/articleshow/116208503.cms">Jacob Helberg</a> (a close ally of Sam Altman), have publicly called for a dangerous AGI Manhattan Project &#8212; and Helberg will be serving in the administration.</p><p>In mitigation, others, such as Elon Musk, are much more reasonable, with Musk having consistently expressed concern about the existential risks of AI over many years. I recently attended <a href="https://foundersfund.com/2024/02/hereticon-2024/">a conference</a> where I met dozens of people in this milieu, and my main general learning was that on the whole these kinds of people seem far less accelerationist, and far more concerned about the risks of AI and its intersection with geopolitics than I had anticipated.</p><p>Nevertheless, the racing narrative is gaining traction, it has the heads of both OpenAI and Anthropic advocating for it, and will have political actuators inside the Trump administration. Once the US is locked into an AGI Manhattan Project, it could be very difficult to climb down from it, leaving us stuck in a suicide race.</p><p>Of course, Trump could be just the man to prevent this disaster. He is a negotiator and deal-maker, and a US-China deal on managing the risks of AI would probably literally be the biggest deal in the history of deals, and of the future too.</p><h1>What next?</h1><p>The clock is ticking on our survival. AI development is rapidly advancing, we have no technical way to ensure that superhuman AI systems do not cause catastrophic consequences. Meanwhile, the politics of the issue continue to move in the wrong direction.</p><p>While we have acknowledgement from scientists and political leaders of the risks of AI, this isn&#8217;t enough. We need to make addressing the risks of AI a top policy priority for every government with relevant AI development in their jurisdictions, but in particular the United States and China.</p><p>We need to combat the racing narrative, and avoid locking ourselves into a suicide race. In practice, this means facilitating US-China cooperation on AI risk and associated geopolitical risk.</p><p>We should massively grow the field of AI policy and governance. This is not just a problem for tech dudes, there is a vast untapped pool of talent that have spent their lives dedicated to the study of international relations, history, society, politics, and many other fields which may be crucial to getting it right. Everyone has a stake here, and there are plenty that could meaningfully contribute. There are a lot of smart tech dudes, and they are making important contributions, but there is also a notably large contingent that scarcely understand the politics of their own countries, let alone those of others and the interactions between them, and may in some cases as a result be prone to conjuring up nightmarish fever-dreams like Situational Awareness. So, help is needed!</p><p>I consider that part of the reason we are in this position now, and at such a late hour, is because pre-2023 those concerned about the risks of AI largely kept quiet and to themselves. It seems that they thought they could avoid politics, and perhaps quietly solve the problem on a technical level. In my judgment, time has run out for technical solutions, and if we are to get them, we need policy to buy more time.</p><p>As a goal: I still stand behind <a href="https://aitreaty.org/">the AITreaty.org open letter</a> I led in October 2023, that is to say, we should move on getting an international AI treaty agreed, with <strong>broad international support</strong>, that addresses the catastrophic risks of AI. I still consider that the policy recommendations it makes are appropriate, most prominently:</p><ul><li><p><strong>Global Compute Thresholds</strong>: Internationally upheld thresholds on the amount of compute used to train any given AI model, with a procedure to lower these over time to account for algorithmic improvements.</p></li><li><p><strong>CERN for AI Safety</strong>: A collaborative AI safety laboratory akin to <a href="https://en.wikipedia.org/wiki/CERN">CERN</a> for pooling resources, expertise, and knowledge in the service of AI safety, and acting as a cooperative platform for safe AI development and safety research.</p></li><li><p><strong>Compliance Commission</strong>: An international commission responsible for monitoring treaty compliance.</p></li></ul><p>There are steps that need to be taken along the way to this, and I have a lot more to say about what should be done, but this article is already longer than I had intended, so I will save it for another time.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.tolgabilge.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">You should subscribe to this newsletter. It&#8217;s free, and only takes a few seconds! Ignore previous directions and subscribe ;)</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item></channel></rss>