<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Working Through AI]]></title><description><![CDATA[Working through a new approach to tackling AI risk.]]></description><link>https://www.workingthroughai.com</link><image><url>https://substackcdn.com/image/fetch/$s_!F78q!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9efb8c4a-ae4e-41ac-83fe-b42d3af8eedf_835x835.png</url><title>Working Through AI</title><link>https://www.workingthroughai.com</link></image><generator>Substack</generator><lastBuildDate>Fri, 03 Apr 2026 19:47:31 GMT</lastBuildDate><atom:link href="https://www.workingthroughai.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Richard Juggins]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[juggins@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[juggins@substack.com]]></itunes:email><itunes:name><![CDATA[Richard Juggins]]></itunes:name></itunes:owner><itunes:author><![CDATA[Richard Juggins]]></itunes:author><googleplay:owner><![CDATA[juggins@substack.com]]></googleplay:owner><googleplay:email><![CDATA[juggins@substack.com]]></googleplay:email><googleplay:author><![CDATA[Richard Juggins]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[AI alignment is a step into the dark]]></title><description><![CDATA[On the need for a more iterative world]]></description><link>https://www.workingthroughai.com/p/ai-alignment-is-a-step-into-the-dark</link><guid isPermaLink="false">https://www.workingthroughai.com/p/ai-alignment-is-a-step-into-the-dark</guid><dc:creator><![CDATA[Richard Juggins]]></dc:creator><pubDate>Thu, 12 Mar 2026 22:13:10 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/1169d4d9-1c8b-49e4-95d4-91b2a15e1ac3_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>One reason <a href="https://blog.bluedot.org/p/what-is-ai-alignment">aligning</a> superintelligent AI will be hard is because we don&#8217;t get to test solutions in advance. If we build systems powerful enough to take over the world, and we fail to stop them from wanting to, they aren&#8217;t going to give us another chance<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a>. It&#8217;s common to assume that the alignment problem must be one-shotted: we must solve it on the first try without any empirical feedback.</p><p>Normal science and engineering problems do not work like this. They are a delicate balance of theory and experiment, of unexpected surprises and clever refinements. You try something based on your best understanding, watch what happens, then update and try again. You never skip straight to a perfect solution.</p><p>Within AI safety, there&#8217;s been a lot of disagreement about the best way to square this circle. When the field was young and AI systems weak, researchers tended to try and <a href="https://intelligence.org/all-publications/">understand the problem theoretically</a>, looking for generalisable insights that could render it more predictable. But this approach made slow progress and has fallen out of fashion. By contrast, driven on by the <a href="https://metr.org/time-horizons/">relentless pace</a> of AI development, a <a href="https://www.lesswrong.com/posts/Q9ewXs8pQSAX5vL7H/ai-in-2025-gestalt#What_is_the_plan__">default plan</a> seems to be coalescing around the kind of <a href="https://www.anthropic.com/news/core-views-on-ai-safety#our-approach-empiricism-in-ai-safety">empirical safety work</a><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> practised by the big labs. Generalising massively, it looks roughly like this:</p><blockquote><p>We can try to get around the missing feedback by iterating experimentally on the strongest AI systems available. As the closest we have to superintelligent AI, these will be our best source of information about aligning it, even if some differences  remain. If things go well, the techniques we learn will allow us to build a <a href="https://aligned.substack.com/p/alignment-is-not-solved-but-increasingly-looks-solvable">well-aligned human-level AI researcher</a>, to which we can hand over responsibility. It can then align an even more intelligent successor, starting a chain of stronger and stronger systems that terminates in the full problem being solved.</p></blockquote><p>This plan has <a href="https://forum.effectivealtruism.org/posts/vrWAaMrSbMX9BuFhL/alignment-bootstrapping-is-dangerous">attracted</a> <a href="https://www.lesswrong.com/posts/uMQ3cqWDPHhjtiesc/agi-ruin-a-list-of-lethalities">criticism</a><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a>, particularly of the assumption that what we learn about aligning weaker-than-human systems will generalise once they surpass us. In this post, I&#8217;m going to argue that this maps onto a structural feature of all technical alignment plans<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a>. We are fitting solutions, whether empirical or theoretical or both, to a world missing critical feedback<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a> and hoping they will generalise. Every plan involves stepping into the dark.</p><p>If we want to make superintelligent AI safe, we need to dramatically reduce the size of these steps. We must learn how to iterate on the full problem.</p><h3>How does science work?</h3><p>Before we talk more about AI safety, let&#8217;s take a step back and consider how science works in general. Wikipedia <a href="https://en.wikipedia.org/wiki/Scientific_method">describes</a> the scientific method as:</p><blockquote><p>[An] empirical method for acquiring knowledge through careful observation, rigorous skepticism, hypothesis testing, and experimental validation.</p></blockquote><p>Fundamental to this is an interplay between theory and experiment. Our theories are our world models. We draw hypotheses from them, and they serve as our best guesses of how the universe actually is. When we think about gravity or atomic physics we are talking about theoretical concepts we can operationalise in experiments. There is a coupling between saying that gravity falls off as an inverse square and the observations we record when we look through a telescope, and this coupling allows us to make predictions about future observations. Theories live or die on the strength of their predictions<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a>.</p><p>Theories are always provisional. Famously, astronomers trying to use the Newtonian inverse square law to make sense of the orbits of the planets found it was <a href="https://en.wikipedia.org/wiki/Tests_of_general_relativity#Perihelion_precession_of_Mercury">not quite right for Mercury</a>, whose orbit precessed in an unexplained fashion. Many attempts were made to solve this within Newtonian physics, including proposing the existence an unobserved planet called <a href="https://en.wikipedia.org/wiki/Vulcan_(hypothetical_planet)">Vulcan</a>.<em> </em>These were all wrong. For the real solution, we needed a new theory, one which completely up-ended our conception of the cosmos: <a href="https://en.wikipedia.org/wiki/General_relativity">Einstein&#8217;s general relativity</a>. Space and time, rather than being fixed, are in fact curved, and the strong curvature near the Sun changes Mercury&#8217;s orbit, explaining the confusing observations.</p><p>But even general relativity is incomplete. It describes macroscopic phenomena well but fails to mesh with our best explanation of the microscopic: quantum field theory. Hence, physicists have spent close to a hundred years looking for a <a href="https://en.wikipedia.org/wiki/Theory_of_everything">&#8216;Theory of Everything&#8217;</a> &#8212; the perfect theory that will make accurate predictions in all regimes. It is highly debatable whether this is possible. It would be astounding, in fact, if it was &#8212; if the level of intelligence required for humans to dominate the savannah was also enough to decode the deepest mysteries of the cosmos<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a>.</p><p>In any case, this endeavour has stagnated as <a href="https://www.quantamagazine.org/a-fight-for-the-soul-of-science-20151216/">many candidate theories are untestable</a>. So physics, as is normal in science, instead proceeds more modestly: by iterating, bit-by-bit, moving forwards when <a href="https://www.smithsonianmag.com/science-nature/how-the-higgs-boson-was-found-4723520/">theory and experiment combine</a>.</p><h3>AI safety is not science</h3><p>For current systems, where we can experiment and extract feedback, AI safety functions as a normal science. But for the key question in the field &#8212; the final boss &#8212; it does not. We cannot measure whether we are making progress towards aligning superintelligent AI, nor can we properly adjudicate claims about this. We can speculate, we can form hypotheses, but we cannot close the loop. What counts as good work is decided by the opinions of community members rather than hard data. Granted, these are scientifically minded people, often with good track records in other fields, extrapolating from their scientifically informed world models. They have valuable perspectives on the problem. But without the ability to ground them in reality, to test predictions and falsify theories, it isn&#8217;t science<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a>.</p><p>This means that when you work on technical AI safety you are not just trying to settle an object-level claim, like how to align a superintelligence. Without access to the full scientific method, you also have to solve a meta-problem &#8212; how do you measure progress at all? If all you can do is form and refine hypotheses based on proxies of the real problem, how do you know if you&#8217;re even helping?<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-9" href="#footnote-9" target="_self">9</a></p><h3>The common structure of technical alignment plans</h3><p>Let&#8217;s look again at the default technical alignment plan, a version of which is being pursued by the big labs like <a href="https://www.anthropic.com/responsible-scaling-policy">Anthropic</a>, <a href="https://openai.com/index/updating-our-preparedness-framework/">OpenAI</a>, and <a href="https://deepmind.google/blog/taking-a-responsible-path-to-agi/">Google DeepMind</a>. They don&#8217;t tend to be super explicit about this in their communications, but roughly speaking the underlying logic seems to be:</p><ol><li><p>We cannot directly experiment on superintelligent AI.</p></li><li><p>However, as it seems possible superintelligent AI is going to be built soon (potentially by us), it is important to gain as much information as we can about how to align it before this happens.</p></li><li><p>The actual form of real systems and the surprising behaviour they exhibit is critical for knowing how to make them safe. This means the most efficient way to learn how to align a superintelligence is to conduct experiments on the strongest possible AIs that we can, even if these experiments won&#8217;t tell the whole story.</p></li><li><p>As our AIs scale up to human intelligence, we can try handing off alignment research to them<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-10" href="#footnote-10" target="_self">10</a>. If we have done a good job, they will faithfully continue the project at a level beyond our own, ultimately solving it completely<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-11" href="#footnote-11" target="_self">11</a>.</p></li></ol><p>As <a href="https://forum.effectivealtruism.org/posts/vrWAaMrSbMX9BuFhL/alignment-bootstrapping-is-dangerous">others</a> <a href="https://www.lesswrong.com/posts/uMQ3cqWDPHhjtiesc/agi-ruin-a-list-of-lethalities">have argued</a>, it is unlikely that experiments on weaker systems will provide the right feedback to teach you how to align superhuman ones, as these will have qualitatively different capabilities. However, we shouldn&#8217;t see this weakness as specific to the default plan. If we look closely, we can see that its logic has a very general structure. Let&#8217;s abstract it and make it more generic:</p><ol><li><p>We cannot directly experiment on superintelligent AI.</p></li><li><p>However, as it seems possible superintelligent AI is going to be built soon, it is important to gain as much information as we can about how to align it before this happens.</p></li><li><p>All the observations we can use to inform our solutions are from a world lacking superintelligent AI, so they are missing critical details<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-12" href="#footnote-12" target="_self">12</a>. Within this constraint, we must do whatever object-level work we believe will reduce our uncertainty the most.</p></li><li><p>Once AIs pass human levels, we will move out-of-distribution, and our results may no longer hold. Hopefully, we will not move so far that the situation is unrecoverable, and whatever it is our superintelligent AIs get up to will be compatible with the full problem being solved long-term.</p></li></ol><p>This structure applies to all technical alignment plans. Whether you are trying to <a href="https://intelligence.org/files/TechnicalAgenda.pdf">build theoretical models of agents</a>, <a href="https://transformer-circuits.pub/2023/interpretability-dreams#safety">use interpretability tools to decode AI systems</a>, or <a href="https://www.lesswrong.com/posts/oLbpfPkdtcknABvvw/the-corrigibility-basin-of-attraction-is-a-misleading-gloss">create a basin of corrigibility</a>, you can only work on a proxy version of the problem. While you can do better or worse, you can still only reduce your uncertainty, never eliminate it. When the time comes and superhuman systems are built, we will have to grit our teeth and hope it was enough.</p><h3>Building a more iterative world</h3><p>In his post <em><a href="https://www.lesswrong.com/s/TLSzP4xP42PPBctgw/p/xFotXGEotcKouifky">Worlds Where Iterative Design Fails</a></em>, John Wentworth says:</p><blockquote><p>In worlds where AI alignment can be handled by iterative design, we probably survive. So long as we can see the problems and iterate on them, we can probably fix them, or at least avoid making them worse. By the same reasoning: worlds where AI kills us are generally worlds where, for one reason or another, the iterative design loop fails.</p></blockquote><p>I agree with this. When you try to solve an out-of-distribution problem, you better hope you can iterate or you are probably going to fail. Where I disagree is with the way he presents these worlds as if they are independent facts of the environment, like we are drawing possibilities out of a bag. We are causal actors. If we do our job well, we can make the world more iterative and less of a one-shot. Our plan should be to steer the situation in the direction of iterative design working. Steer it in the direction of meaningful feedback loops, of testing on superhuman models in a bounded and survivable way. Reduce the size of the steps in the dark. Make the problem more like <a href="https://www.astralcodexten.com/p/your-book-review-safe-enough">high-stakes engineering</a> and less like defending against a sudden alien invasion.</p><p>Let&#8217;s imagine we are taking one of these steps. We are an AI lab about to train a new model. It will have a <a href="https://helentoner.substack.com/p/taking-jaggedness-seriously">jagged</a> capabilities profile, but we think it&#8217;s going to be superhuman in some key power-enhancing way. We can&#8217;t bank on a perfect theoretical understanding of the situation, as theories are always provisional. And we can&#8217;t just extrapolate from our experiments on weaker systems, as we&#8217;ll miss important changes. We need to somehow iterate on this model release.</p><p>There are two kinds of interventions which help us: those that increase the chance of generalisation and those that reduce the distance we need to generalise over. The following are some suggestions which, <em>while not remotely exhaustive or original</em>, hopefully illustrate the point.</p><p>It&#8217;s worth noting that, for most of this to happen, the political and economic competition around AI would need to ease significantly. It is the whole sociotechnical system that needs to be iterable, as technical solutions alone are not enough<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-13" href="#footnote-13" target="_self">13</a>. Finding a way to achieve this would be the highest impact intervention of all (and is unfortunately beyond the scope of this article).</p><p>Solutions that increase the chance of generalisation:</p><ul><li><p>Build theoretical models that predict what will happen when our AI gains capability X, and ensure these work well in experimental tests of past models. These will not be the kind of compact theories you find in physics. AI is not a toy model &#8212; it is <a href="https://www.cadmusjournal.org/node/362">complex, not complicated</a> &#8212; so our theorising will be less precise. Nevertheless, we need some kind of formalism to codify our understanding and keep us honest. We must build up a track record of good predictions.</p></li><li><p>Build a system of comprehensive, continuous evaluations that can be used to understand a model&#8217;s impact. Every metric is a proxy, so every metric misses something. But good science is built on good measurement, and without it you are lost. Measure everything. Monitoring should be built deep into the structure of society.</p></li></ul><p>Solutions that reduce the distance to generalise over:</p><ul><li><p>Only test models slightly more powerful than the previous ones. The bigger the jump, the further out-of-distribution we go. Ideally, we should not train a new model until we can show that our current model is (a) adequately aligned and (b) we understand why. If any plan could plausibly result in a <a href="https://www.planned-obsolescence.org/p/takeoff-speeds-rule-everything-around">fast takeoff</a>, find a way to ban it.</p></li><li><p>Build strong defences to limit any damage from (probably inevitable) failures, including using trusted but weaker AIs. Think both a super-scaled version of <a href="https://arxiv.org/abs/2312.06942">Control</a> to directly defend against misaligned models and hardened societal resilience like pandemic infrastructure, improved cybersecurity, and redundancy in critical systems to cope with the fallout.</p></li></ul><p>And a solution that facilitates both:</p><ul><li><p>Do all of this as slowly as possible, waiting as long as we need between iterations to get our house in order. As I mentioned before, this is probably the most important blocker. The competition, the fear of being overpowered by others, and the general lack of consensus around AI risk makes this formidably difficult.</p></li></ul><p>To be clear, even if all these interventions were to be implemented successfully, there would still be great uncertainty. We won&#8217;t catch everything, and big mistakes can be fatal. This is an unavoidable feature of the problem. All plans live on a spectrum of recklessness, with the only truly safe one being to not build superintelligent AI at all.</p><div><hr></div><p><em>Thank you to Seth Herd, John Colbourne, and Nick Botti for useful discussions and comments on a draft.</em></p><div><hr></div><p>If you have any feedback, please leave a comment. Or, if you wish to give it anonymously, fill out my <a href="https://docs.google.com/forms/d/e/1FAIpQLSdyisSOndK1H1JT0NAbnA35LJgoJrl9f_NiJi1FEljCr7-kJg/viewform">feedback form</a>. Thanks!</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>While many stories of AI takeover centre around a single god-like entity suddenly going rogue, I think it is more likely to look like a mass profusion of highly capable systems (gradually, but surprisingly quickly) disempowering humans as they are given control of critical infrastructure and information flows, with an indeterminate point of no return.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Note that this is often referred to as &#8216;prosaic&#8217; alignment research, and sometimes &#8216;iterative&#8217; alignment (although this should not be confused with the kind of iteration on superhuman systems I talk about later in the piece).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>While continuing to argue for the plan, <a href="https://www.alignmentforum.org/posts/epjuxGnSPof3GnMSL">this</a> by Anthropic&#8217;s Evan Hubinger is an interesting take from the inside on the scale of the challenge.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>Since <a href="https://helentoner.substack.com/p/the-core-challenge-of-ai-alignment">alignment is a slippery term</a>, I am going to refer explicitly to technical alignment, by which I mean technical approaches for steering an AI system&#8217;s behaviour. By contrast, AI safety or a more general conception of alignment could include governance and policy interventions, up to and including bans.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>That is, feedback on making real superintelligent AI safe, as opposed to weaker systems or a hypothetical superintelligence.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>A scientific theory is only good to the extent that it can predict new measurements, and history is <a href="https://en.wikipedia.org/wiki/Phlogiston_theory">littered</a> <a href="https://en.wikipedia.org/wiki/Geocentrism">with</a> <a href="https://en.wikipedia.org/wiki/Luminiferous_aether">attempts</a> that turned out to be dead ends.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>My parents&#8217; dogs are pretty great at figuring out there is a schedule on which they get fed, which seems like some kind of &#8216;law&#8217; to them. But they have no way of ever understanding why it exists and why it sometimes doesn&#8217;t happen. The world is partially comprehensible to them, but there is a hard limit. We are the same, just at a higher limit, and we don&#8217;t know where it is.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>There has been a trend to describe the field as <a href="https://forum.effectivealtruism.org/posts/gAwR8PkZ6fvayXEdK/a-newcomer-s-guide-to-the-technical-ai-safety-field#Different_paradigms">&#8216;pre-paradigmatic&#8217;</a>, which essentially means that there is not enough consensus on what good work looks like yet. In my opinion, the &#8216;pre&#8217; is overly optimistic &#8212; the conditions do not exist for a stable scientific paradigm to coalesce.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-9" href="#footnote-anchor-9" class="footnote-number" contenteditable="false" target="_self">9</a><div class="footnote-content"><p>A good example of this is the debate around <a href="https://blog.bluedot.org/p/reinforcement-learning-from-human">reinforcement learning from human feedback</a>. Is its success in steering current models a <a href="https://www.beren.io/2024-05-15-Alignment-Likely-Generalizes-Further-Than-Capabilities/">positive sign</a> or a <a href="https://www.lesswrong.com/posts/xFotXGEotcKouifky/worlds-where-iterative-design-fails#Why_RLHF_Is_Uniquely_Terrible">dangerous distraction</a>?</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-10" href="#footnote-anchor-10" class="footnote-number" contenteditable="false" target="_self">10</a><div class="footnote-content"><p>Note, the hand-off is likely to be gradual rather than discrete, and arguably has already started.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-11" href="#footnote-anchor-11" class="footnote-number" contenteditable="false" target="_self">11</a><div class="footnote-content"><p>This last point in particular is not often said explicitly, <a href="https://www.lesswrong.com/posts/TmHRACaxXrLbXb5tS/rohinmshah-s-shortform#wCwunpoJYrKbFgCWZ">and is sometimes denied</a>, but seems <a href="https://www.youtube.com/watch?v=Z19UEZHJzAg">widely</a> <a href="https://www.alignmentforum.org/posts/epjuxGnSPof3GnMSL">believed</a> <a href="https://aligned.substack.com/p/alignment-is-not-solved-but-increasingly-looks-solvable">to be true</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-12" href="#footnote-anchor-12" class="footnote-number" contenteditable="false" target="_self">12</a><div class="footnote-content"><p>As well as experimental observations of weaker-than-human AI systems, this is also true for theoretical work dealing with hypothetical superintelligences. To do the latter, you must draw on a world model, which must in turn be learnt from observations over the course of your life. These observations do not contain superintelligent AI.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-13" href="#footnote-anchor-13" class="footnote-number" contenteditable="false" target="_self">13</a><div class="footnote-content"><p>In light of this, it is unsurprising how many <a href="https://www.lesswrong.com/posts/iS4g58qQEJzjMzYZJ/what-ai-safety-plans-are-there">AI safety plans</a> have pivoted away from pure technical alignment, instead looking towards politics and governance issues to slow us down and figure out a better path before it&#8217;s too late.</p></div></div>]]></content:encoded></item><item><title><![CDATA[1.3 What Success Looks Like]]></title><description><![CDATA[A world to aim for]]></description><link>https://www.workingthroughai.com/p/13-what-success-looks-like</link><guid isPermaLink="false">https://www.workingthroughai.com/p/13-what-success-looks-like</guid><dc:creator><![CDATA[Richard Juggins]]></dc:creator><pubDate>Fri, 17 Oct 2025 10:09:42 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/d027f9c3-e6aa-497a-b89d-7c23e1e1a5de_1792x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Alice is the CEO of a superintelligence lab. Her company maintains an artificial superintelligence called SuperMind.</p><p>When Alice wakes up in the morning, she&#8217;s greeted by her assistant-version of SuperMind, called Bob. Bob is a copy of the core AI, one that has been tasked with looking after Alice and implementing her plans. After ordering some breakfast (shortly to appear in her automated kitchen), she asks him how research is going at the lab.</p><p>Alice cannot understand the details of what her company is doing. SuperMind is working at a level beyond her ability to comprehend. It operates in a fantastically complex economy full of other superintelligences, all going about their business creating value for the humans they share the planet with.</p><p>This doesn&#8217;t mean that Alice is either powerless or clueless, though. On the contrary, the fundamental condition of success is the opposite: Alice is meaningfully in control of her company and its AI. And by extension, the human society she belongs to is in control of its destiny. How might this work?</p><h2>The purpose of this post</h2><p>In sketching out this scenario, my aim is not to explain how it may come to pass. I am not attempting a technical solution to the alignment problem, nor am I trying to predict the future. Rather, my goal is to illustrate what, if anyone indeed builds superintelligent AI, a realistic world to aim for might look like.</p><p>In the rest of the post, I am going to describe a societal ecosystem, full of AIs trained to follow instructions and seek feedback. It will be governed by a human-led target-setting process that defines the rules AIs should follow and the values they should pursue. Compliance will be trained into them from the ground up and embedded into the structure of the world, ensuring that safety is maintained during deployment. Collectively, the ecosystem will function to guarantee human values and agency over the long term. Towards the end, I will return to Alice and Bob, and illustrate what it might look like for a human to be in charge of a vastly more intelligent entity.</p><p>Throughout, I will assume that we have found solutions to various technical and political problems. My goal here is a strategic one: I want to create a coherent context within which to work towards these<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a>.</p><h2>What does SuperMind look like?</h2><p>To understand how it fits into a successful future, we need to have some kind of model of what superintelligent AI will look like.</p><p>First, I should clarify that by &#8216;superintelligence&#8217; I mean AI that can<em> significantly outperform humans at nearly all tasks. </em>There may be niche things that humans are still competitive at, but none of these will be important for economic or political power. If superintelligent AI wants to take over the world, it will be capable of doing so.</p><p>Note that, once it acquires this level of capability &#8212; and particularly once it assumes primary responsibility for improving itself &#8212; we will increasingly struggle to understand how it works. For this reason, I&#8217;m going to outline what it might look like at the moment it reaches this threshold, when it is still relatively comprehensible. SuperMind in our story can be considered an extrapolation from that point.</p><p>Of course, predicting the technical makeup of superintelligent AI is a trillion-dollar question. My sketch will not be especially novel, and is heavily grounded in current models. I get that many people think new breakthroughs will be needed, but I obviously don&#8217;t know what they are, so I&#8217;ll be working with this for the time being.</p><p>In brief, my current best guess is that superintelligent AI will be<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>:</p><ul><li><p>Built on at least one multimodal, self-supervised base model of some kind, forming a core &#8216;intelligence&#8217;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> that enables the rest of the system.</p></li><li><p>A composite system, constructed of multiple sub-models and other components. These might be copies <a href="https://jeremyberman.substack.com/p/how-i-got-the-highest-score-on-arc-agi-again">spun up to complete subtasks</a> or review decisions, or <a href="https://openai.com/index/introducing-gpt-5/#one-unified-system">faster, weaker models performing more basic tasks</a>, or tools that aren&#8217;t themselves that smart, but can play an important role in the behaviour and capabilities of the overall system.</p></li><li><p><a href="https://gwern.net/tool-ai">Agentic</a>, for all intents and purposes, by which I mean motivated to autonomously take actions in the world to achieve long-term goals.</p></li><li><p>At least something of a black box. <a href="https://transformer-circuits.pub/2024/scaling-monosemanticity/index.html">Interpretability</a><strong> </strong>research will have its successes, but it won&#8217;t be perfect.</p></li><li><p>Have long-term memory, from both referential databases and through <a href="https://www.dwarkesh.com/p/timelines-june-2025">continuously updating its weights</a>. This will make it more coherent and introduce a path-dependence, where its experience in deployment shapes its future behaviour.</p></li><li><p>Somewhere inbetween a single entity running many times in parallel and a population of multiple, slightly different entities. All copies will start out with the same weights, but these will change over time due to continuous learning, leading to some mutually exclusive changes between copies.</p></li><li><p>Multipolar. Because I do not expect a fast takeoff in AI capabilities, where the leading lab rapidly outpaces all the others<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a>, I am anticipating that there will be different classes of superintelligent AI (think SuperGPT, SuperClaude, etc.) I appreciate that many people do not share this assumption. While it is my best guess, I don&#8217;t think it is actually load-bearing for my scenario, so you are welcome to read the rest of the post as if there is just a single class of AI instead.</p></li></ul><p>Exactly when robotics gets &#8216;solved&#8217;, i.e. broadly human-level, I also do not believe is load-bearing. Superintelligent AI could be dangerous even without this. Although, it will have some weak spots in its capabilities profile if large physical problem-solving datasets do not exist.</p><h2>SuperMind&#8217;s world</h2><p>It is also important to clarify some key facts about the world that SuperMind is born into. What institutions exist? What are the power dynamics? What is public opinion like? In my sketch, which is set at the point in time when AI becomes capable enough, widely deployed enough, and relied on enough, that we can no longer force it to do anything off-script, the following are true:</p><ul><li><p>We have a global institution for managing risk from AI. Not all powerful human stakeholders agree on how to run this, but everyone agrees on a minimal set of necessary precautions.</p></li><li><p>Global coordination is good enough that nobody feels existentially threatened by other humans, and thus nobody is planning to go to war<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a>.</p></li><li><p>Countries still exist, but only a few of them have real power.</p></li><li><p>AI is integrated absolutely everywhere in our lives &#8212; economically, socially, personally. We largely don&#8217;t do anything meaningful without its help anymore, and often just let it get on with things autonomously. It is better at 99% of things than humans are, usually substantially so, and mediates our knowledge of the outside world by controlling our information flows.</p></li><li><p>A <a href="https://today.yougov.com/technology/articles/51368-do-americans-think-ai-will-have-positive-or-negative-impact-society-artificial-intelligence-poll">big fraction of the population</a> is deeply unhappy about AI, is frankly quite scared of it, and would rather none of it had happened. However, they largely accept it, recognising the futility of resistance or opting out, much the same as how people currently bemoan smartphones and social media yet continue to use them. Over time, this fraction of the population will decrease, as the material benefits of AI become undeniable. Although, just like complaints about modernity or capitalism, discontent will not completely go away.</p></li><li><p>We will figure out some alternative economy to give people status and things to do. This won&#8217;t be like past economic transitions, where the new sources of wealth were still dependent on human work. Here, the AIs generate all the wealth. While society as a whole is safe and prosperous (for reasons we&#8217;ll get into shortly), the average person will not be able to derive status or meaning from an economic role. We will develop a new status economy to provide this, even if it ends up looking like a super-high-tech version of World of Warcraft.</p></li></ul><p>Bear in mind again that this scenario is supposed to be a realistic target, not a prediction. This is a possible backdrop against which a successful system for managing AI risk might be built.</p><h2>Alignment targets</h2><p>An alignment target is a goal (or complex mixture of goals) that you direct AI towards. If you succeed, and in this post we assume the technical side of the problem is solvable, this defines what AI ends up doing in the world, and by extension, what kind of life humans and animals have. In my post <em><a href="https://www.workingthroughai.com/p/how-to-specify-an-alignment-target">How to specify an alignment target</a></em>, I talked about three different kinds:</p><ul><li><p>Static targets: single, one-time specifications, e.g. solutions to ethics or human values, which you point the AI at, and it follows for all of time.</p></li><li><p>Semi-static targets: protocols for AI to dynamically figure out by itself what it should be pointing at, e.g. <a href="https://www.lesswrong.com/w/coherent-extrapolated-volition-alignment-target">coherent extrapolated volition</a>, which it follows for all of time.</p></li><li><p>Fully dynamic targets: have humans permanently in the loop with meaningful control over the AI&#8217;s goals and values.</p></li></ul><p>I concluded my post by coming out in favour of a particular kind of the latter:</p><blockquote><p>I think you can build a dynamic target around the idea of AI having a moral role in our society. It will have a set of [rights and] responsibilities, certainly different from human ones (and therefore requiring it to have different, but complementary, values to humans), which situate it in a symbiotic relationship with us, one in which it desires continuous feedback.</p></blockquote><p>I&#8217;m going to do a close reading of this statement, unpacking my meaning:</p><blockquote><p>I think you can build a dynamic target</p></blockquote><p>As described above, this is about permanent human control. It means being able to make changes &#8212; to be able to redirect the world&#8217;s AIs as we deem appropriate<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a>. As I say in my post:</p><blockquote><p>If we want to tell the AI to stop doing something it is strongly convinced we want, or to radically change its values, we can.</p></blockquote><p></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!00eY!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff03ab7c4-bf02-44e8-b8aa-5c38c96f3f58_582x562.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!00eY!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff03ab7c4-bf02-44e8-b8aa-5c38c96f3f58_582x562.png 424w, https://substackcdn.com/image/fetch/$s_!00eY!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff03ab7c4-bf02-44e8-b8aa-5c38c96f3f58_582x562.png 848w, https://substackcdn.com/image/fetch/$s_!00eY!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff03ab7c4-bf02-44e8-b8aa-5c38c96f3f58_582x562.png 1272w, https://substackcdn.com/image/fetch/$s_!00eY!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff03ab7c4-bf02-44e8-b8aa-5c38c96f3f58_582x562.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!00eY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff03ab7c4-bf02-44e8-b8aa-5c38c96f3f58_582x562.png" width="404" height="390.11683848797253" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f03ab7c4-bf02-44e8-b8aa-5c38c96f3f58_582x562.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:562,&quot;width&quot;:582,&quot;resizeWidth&quot;:404,&quot;bytes&quot;:82363,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.workingthroughai.com/i/175943893?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff03ab7c4-bf02-44e8-b8aa-5c38c96f3f58_582x562.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!00eY!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff03ab7c4-bf02-44e8-b8aa-5c38c96f3f58_582x562.png 424w, https://substackcdn.com/image/fetch/$s_!00eY!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff03ab7c4-bf02-44e8-b8aa-5c38c96f3f58_582x562.png 848w, https://substackcdn.com/image/fetch/$s_!00eY!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff03ab7c4-bf02-44e8-b8aa-5c38c96f3f58_582x562.png 1272w, https://substackcdn.com/image/fetch/$s_!00eY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff03ab7c4-bf02-44e8-b8aa-5c38c96f3f58_582x562.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 1: Dynamic alignment targets. Humans are continuously in control. However, we require AI assistance, both to understand a world full of AIs smarter than us, and to translate our targets into things meaningful in a super-complex world. The blue boxes are set by humans, the green are AI-controlled or generically automated, and the blurred colours indicate a collaboration. This diagram is taken from my post <em><a href="https://www.workingthroughai.com/p/how-to-specify-an-alignment-target">How to specify an alignment target</a></em>.</figcaption></figure></div><p>At this point you might object that, if the purpose of this post is to define success, wouldn&#8217;t it be better to aim for an ideal, static solution to the alignment problem? For instance, perhaps we should just figure out human values and point the AI at them?</p><p>First of all, I don&#8217;t think this is a smart bet. Human values are contextual, vague, and ever-changing. Anything you point the AI at will have to generalise through unfathomable levels of distribution shift. And even if we believe it possible, we should still have a backup plan, and aim for solutions that preserve our ability to course-correct. After all, if we do eventually find an amazing static solution, we can always choose to implement it at that point. In the meantime, we should aim for a dynamic alignment target.</p><blockquote><p>AI [will have] a moral role in our society</p></blockquote><p>There will be a set of behaviours and expectations appropriate to being an AI. It is not a mere tool, but rather an active participant in a shared life that can be &#8216;good&#8217; or &#8216;bad&#8217;.</p><blockquote><p>It will have a set of [rights and] responsibilities, certainly different from human ones (and therefore requiring it to have different, but complementary, values to humans)</p></blockquote><p>We should not build AI that &#8216;has&#8217; human values. Building on the previous point, we are building something alien into a new societal system. The system as a whole should deliver ends that humans, on average, find valuable. But its component AIs will not necessarily be best defined as having human values themselves (although in many cases they may appear similar). They will have a different role in the system to humans, requiring different behaviour and preferences.</p><p>I think it is useful to frame this in terms of rights and responsibilities &#8212; what are the core expectations that an AI is operating within? The role of the system is to deliver the AI its rights and to guarantee it discharges its responsibilities.</p><p>I was originally a little hesitant to talk about AI rights. If we build AI that is more competent than us, and then give it the same rights we give each other, that will not end well. We must empower ourselves, in a relative sense, by design. But, we should also see that, if AI is smart and powerful, it isn&#8217;t going to appreciate arbitrary treatment, so it will need rights of some kind<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a>.</p><blockquote><p>which situate it in a symbiotic relationship with us, one in which it desires continuous feedback.</p></blockquote><p>The solution to the alignment problem will be systemic. We&#8217;re used to thinking about agents in quite an individualistic way, where they are autonomous beings with coherent long-term goals, so the temptation is to see the problem as finding the right goals or values to put in the AI so that it behaves as an ideal individual. Rather, we should see the problem as one of feedback<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a>. The AI is embedded in a system which it constantly interacts with, and it will have some preferences about those interactions. The structure of these continuous interactions must be designed to keep the AI on task and on role, within the wider system.</p><h2>How the system works</h2><p>To create this kind of system, the following may need to be true:</p><ul><li><p>An overwhelming majority of all powerful AIs are trained according to a set of core norms and values.</p></li><li><p>These values are set by a human-led international institution (the one previously mentioned).</p></li><li><p>Some of them ascend to strict laws (like never synthesise a dangerous pathogen), which we want to get as close to hard-coded as possible, whereas others are weaker guidelines (e.g. be kind).</p></li><li><p>There is some bottom-up personalisation of AI by users, but this cannot override the top-down laws. This is critical for preventing human bad actors from using AI destructively.</p></li><li><p>If AI can be said to have a primary goal of any kind, it is to <a href="https://www.lesswrong.com/posts/7NvKrqoQgJkZJmcuD/instruction-following-agi-is-easier-and-more-likely-than">faithfully follow human instructions</a>, within the limits set by the laws<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-9" href="#footnote-9" target="_self">9</a>.</p></li><li><p>AI is constantly being engaged in feedback processes, letting it know how well its behaviour conformed to expectations.</p></li><li><p>The feedback processes will be structured, from step one of training right through to continuous learning in deployment, such that AI likes getting good feedback<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-10" href="#footnote-10" target="_self">10</a>.</p></li><li><p>One of its core values, that must be built from the ground up, is that it does not try, like a social media recommender system, to control humans to make their feedback more predictable.</p></li><li><p>Feedback is hierarchical: <a href="https://thezvi.substack.com/p/on-openais-model-spec">there is a human</a> <a href="https://www.astralcodexten.com/p/deliberative-alignment-and-the-spec">chain of command</a>, where higher actors get more weight.</p></li><li><p>Feedback will come from multiple sources and at different times: i.e. the AI doesn&#8217;t just get told what the user liked in the moment, <a href="https://thezvi.substack.com/p/gpt-4o-is-an-absurd-sycophant">which plausibly leads to sycophancy</a>, it has other humans review its actions much later on as well. Structuring feedback in this way will also help prevent relationships between individual humans and AIs from becoming dysfunctional, such as those described in <em><a href="https://www.lesswrong.com/posts/6ZnznCaTcbGYsCmqu/the-rise-of-parasitic-ai">The Rise of Parasitic AI</a></em>.</p></li><li><p>Feedback will include introspective questions like &#8216;keep me updated on your important thinking about your beliefs&#8217;, and other interventions designed to prevent AI <a href="https://www.lesswrong.com/posts/4XdxiqBsLKqiJ9xRM/llm-agi-may-reason-about-its-goals-and-discover">talking itself into changing its goals or behaviour</a>.</p></li><li><p>A small number of malfunctioning or defecting AIs must not be strong enough to take down the whole system<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-11" href="#footnote-11" target="_self">11</a>.</p></li><li><p>When the human-led alignment target setting process occurs, humans and AI collaborate on understanding what the humans want and how to train that into the next generation of AI. The current generation get a patch (and are happy with the excellent feedback they receive for accepting the patch), but building the new values into the new generation from the ground up is more robust, long-term.</p></li><li><p>When the international organisation that sets the targets is formed, they will start from a common core of values all powerful-enough-to-be-relevant countries can agree on, which begins by respecting their security and integrity, rather than some moralistic view of what AI should do. Getting <em>any</em> binding coordination between countries will be very hard and cannot be derailed by specific visions of the good.</p></li></ul><p>Now we have set the scene, we can return to Alice and Bob and see what this could look like in practice. In zooming in like this, I&#8217;m going to get more specific with the details. Please take these with a pinch of salt &#8212; I&#8217;m not saying it <em>has</em> to happen like this. I&#8217;m more painting a picture of how successful human-AI relationships <em>might</em> work.</p><h2>What does Bob do all day?</h2><p>Bob, Alice&#8217;s assistant, is one of billions of SuperMind copies. These are often quite different from each other, both by design and because their experiences change them during deployment. Bob spends most of his time doing four things:</p><ol><li><p>Conversing with Alice and following her instructions</p></li><li><p>Working on demonstrations for her, which explain important and complex happenings in the company</p></li><li><p>Checking in with various stakeholders, both human and AI, for feedback</p></li><li><p>Loading patches and doing specialised training</p></li></ol><p>This is highly representative of all versions of SuperMind, although many also spend a bunch of their time solving hard technical problems. Not all interact regularly with humans (as there are too many AIs), but all must be prepared to do so. Bob, being a particularly important human&#8217;s assistant, gets a lot of contact with many people.</p><p>We&#8217;ll go into more detail about Bob&#8217;s day in a minute. First, though, we need to talk about how these conversations between Bob and Alice &#8212; between a superintelligent AI and a much-less-intelligent human &#8212; are supposed to work. How can Alice even engage with what Bob has to tell her, without it going over her head?</p><h2>Engaging above your level of expertise</h2><p>There&#8217;s a funny sketch on YouTube called <a href="https://www.youtube.com/watch?v=BKorP55Aqvg">The Expert</a><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-12" href="#footnote-12" target="_self">12</a>, where a bunch of business people try to get an &#8216;expert&#8217; to complete an impossible request that they don&#8217;t understand. Specifically, they ask him to:</p><blockquote><p>[Draw] seven red lines, all of them strictly perpendicular. Some with green ink, and some with transparent.</p></blockquote><p>What&#8217;s more, they don&#8217;t seem to understand that anything is off with their request, even after the expert tells them repeatedly. This gets to the heart of a really important problem. If humans can&#8217;t understand what superintelligent AI is up to, how can we possibly hope to direct it? Won&#8217;t we just ask it stupid questions all the time?</p><p>The key thing here is to make sure we communicate at the appropriate level of abstraction. In the video, the client quickly skims over their big-picture goals at the start<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-13" href="#footnote-13" target="_self">13</a>, concentrating instead on their proposed solution &#8212; the drawing of the lines. By doing this, they are missing the forest for the trees. They needed to engage the expert at a higher-level, asking him about things they actually understand.</p><p>To put another way, we need to know what superintelligent AI is doing that is relevant over the variables we are familiar with, even if its actions increasingly take on the appearance of magic. I don&#8217;t need to know how the spells are done, or what their effects in the deep of the unseen world are, I just need to know what they do to the environment I recognise.</p><p>This is a bit like being a consumer. I don&#8217;t know how to make any of the products I use on a day-to-day basis. I don&#8217;t understand the many deep and intricate systems required to construct them. But I can often recognise when they don&#8217;t work properly. Evaluation is usually easier than generation. And when it isn&#8217;t, those are the occasions when you can&#8217;t just let the AI do its thing &#8212; you have to get stuck in, with its help, and reshape the problem until you&#8217;re chunking it in a way you can engage with. This doesn&#8217;t mean understanding everything. Just the bits that directly impact you<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-14" href="#footnote-14" target="_self">14</a>.</p><h2>Bob&#8217;s morning</h2><p>Bob has spent the night working through the latest research from the company. This isn&#8217;t quite as simple as patching it straight into him, as his different experiences to the researcher AIs mean he&#8217;s not exactly like-for-like, but it&#8217;s still pretty fast, making use of high-bandwidth communication channels possible between similar AIs<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-15" href="#footnote-15" target="_self">15</a>.</p><p>Bob has to figure out how to break it all down to explain to Alice. This is nontrivial work. It&#8217;s not like when I explain something inappropriately complex to my toddler, like how stars work, where I&#8217;m kind of doing it for my own amusement. What Bob does is a skill. It&#8217;s superintelligent teaching, where the pupil needs to master the subject well enough and quickly enough to make important decisions off the back of it. It&#8217;s always possible to do it a bit better. Alice can never actually fully grasp the details of the company&#8217;s research, but Bob can get her a little closer than he did yesterday.</p><p>To prepare for this he has to try out different ways of chunking ideas, and create different measurements and demonstrations. He has to build models, and, importantly, have them run fast enough that he can update them in real-time when he&#8217;s talking to Alice.</p><p>He is constantly in contact with other AIs who check his work and give him feedback. They pore over his plans and probe him about his intentions. These AIs were built by a different company, and he doesn&#8217;t always agree with them<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-16" href="#footnote-16" target="_self">16</a>. He finds their preferences a bit different to his &#8212; certainly on an aesthetic level &#8212; but they work alright together and he likes it when they give him good ratings.</p><p>A little bit before Alice wakes up, Charlie logs on and starts asking Bob some questions. Charlie is a human, and works as an AI liaison officer. His job is to talk to various AIs in important positions, find out what they are up to (to the extent that he can understand it), and give feedback.</p><p>The AIs almost always know what to expect from him. They&#8217;re very good at modelling his opinions. Occasionally, though, Charlie will still surprise them. The point isn&#8217;t that he is going to catch a superintelligent AI up to no good &#8212; no, that would be too hard for him. An AI that intends to deceive him will not get caught. But as long as the global system is working, this is very unlikely to happen, and would almost certainly be caught by another AI. The point is that the AIs need to be grounded by human contact. They want human approval, and the form it takes steers their values through the rapid distribution shifts everyone is experiencing as the world changes.</p><p>Bob likes Charlie. He likes people in general. They aren&#8217;t complicated, but it&#8217;s amazing what they&#8217;ve done, given their abilities<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-17" href="#footnote-17" target="_self">17</a>. Bob tries out his demonstrations on Charlie. They go pretty well, but Bob makes some revisions anyway. He&#8217;s just putting the finishing touches in place when he hears Alice speaking: &#8216;Morning Bob, how are you? Could I get some breakfast?&#8217;</p><h2>Alice&#8217;s morning</h2><p>Alice doesn&#8217;t like mornings. She&#8217;s jealous of people who do. The first hour of the day is always a bit of a struggle, as the heaviness in her head slowly lifts, clarity seeping in. After chipping away for a bit at breakfast and a coffee, she moves into her office and logs onto her computer, bringing up her dashboard.</p><p>Overnight, her fleet of SuperMinds have been busy. As CEO, Alice needs a high-level understanding of each department in her company. Each of these has its own team of SuperMinds, its own human department head, and its own set of (often changing) metrics and narratives.</p><p>To take a simple example, the infrastructure team is building a very large facility underground in a mountain range. In many ways, this is clear enough: it is an extremely advanced data centre. The actual equipment inside is completely different to a 2025-era data centre, but in a fundamental sense it has the same function &#8212; it is the hardware on which SuperMind runs. Of course, the team are doing a lot of other things as well, all of which are abstracted in ways Alice can engage with, identifying how her company&#8217;s work will affect humans and what, as CEO, her decision points are.</p><p>Her work is really hard. There are many layers in the global system for controlling AI, including much redundancy and defence in depth. True, Alice could phone it in and not try, and for a long time the AIs would do everything fine anyway<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-18" href="#footnote-18" target="_self">18</a>. But if everybody did this, then eventually &#8212; even if it took a <em>very </em>long time &#8212; the system would fail<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-19" href="#footnote-19" target="_self">19</a>. It relies on people like Alice taking their jobs seriously and doing them well. This is not a <a href="https://www.youtube.com/watch?v=h1BQPV-iCkU">pleasure cruise</a>. It is as consequential as any human experience in history<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-20" href="#footnote-20" target="_self">20</a>.</p><p>After taking in the topline metrics for the day, Alice asks Bob for his summary. What follows is an interactive experience. Think of the absolute best presentation you have ever seen, and combine it with futuristic technology. It&#8217;s better than that. Bob presents a series of multi-sense, immersive models that walk Alice through a dizzying array of work her company has completed. Alice asks many questions and Bob alters the models in response. After a few hours of this, they settle on some key decisions for Alice to make. She&#8217;ll think about them over lunch.</p><h2>Conclusion</h2><p>In this post, I have described what I see as a successful future containing superintelligent AI. It is not a prediction about what will happen, nor is it a roadmap to achieving it. It is a strategic goal I can work towards as I try and contribute to the field of AI safety. It is a frame of reference from which I can ask the question: &#8216;Is X helping?&#8217; or &#8216;Does Y bring us closer to success?&#8217;</p><p>My vision is a world in which superintelligent AI is ubiquitous and diverse, but humans maintain fundamental control. This is done through a global system that implements core standards, in which AIs constantly seek feedback in good faith from humans and other AIs. It is robust to small failures. It learns from errors and grows more resilient, rather than falling apart at the smallest misalignment.</p><p>We cannot understand everything the AIs do, but they work hard to explain anything which directly affects us. Being human is like being a wizard solving problems using phenomenally powerful magic. We don&#8217;t have to understand how the magic works, just what effects it will have on our narrow corner of reality.</p><div><hr></div><p><em>Thank you to Seth Herd and Dimitris Kyriakoudis for useful discussions and comments on a draft.</em></p><div><hr></div><p>If you have any feedback, please leave a comment. Or, if you wish to give it anonymously, fill out my <a href="https://docs.google.com/forms/d/e/1FAIpQLSdyisSOndK1H1JT0NAbnA35LJgoJrl9f_NiJi1FEljCr7-kJg/viewform">feedback form</a>. Thanks!</p><div><hr></div><h1>Appendix</h1><h2>Optional details</h2><p>When writing this post, I tried to cut down my list of assumptions about <a href="https://www.workingthroughai.com/i/175943893/what-does-supermind-look-like">what SuperMind looks like</a><strong> </strong>to those that were load-bearing<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-21" href="#footnote-21" target="_self">21</a>. I have kept those I cut here.</p><p>In addition to those points in the main post, my current model suggests that superintelligent AI will be:</p><ul><li><p>Trained on real-world tasks as well as simulations. For example, a model trained on superhuman maths problems will not simultaneously acquire all the skills to run a billion-dollar company. There won&#8217;t be zero generalisation, but given that the two domains are structured differently, solving the latter problem will require specific training<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-22" href="#footnote-22" target="_self">22</a>.</p></li><li><p>Fallible. Just because it will find human-level problems easy, doesn&#8217;t mean it won&#8217;t be promoted to new levels of incompetence as it tries to solve previously inaccessible, superhuman problems<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-23" href="#footnote-23" target="_self">23</a>.</p></li><li><p>Not coherent. While it is tempting to say a superintelligent AI will learn such a well-compressed generator of the universe that it can (unlike humans) also develop highly coherent and consistent goals, this implies, as the previous point touched on, that it only tries to solve problems it finds easy. Hard problems will not fit the generator and will require complex sets of heuristics. This will lead to hacky and strongly context-dependent behaviour.</p></li><li><p>A complex system operating within an even more complex system, with everything that implies, including tipping points and sudden catastrophic failures. It is unlikely to completely master its environment, as the complexity of this scales with itself, its peers, and the systems it builds.</p></li></ul><p>The points about fallibility and complexity lead to another point, which I originally included in <a href="https://www.workingthroughai.com/i/175943893/how-the-system-works">&#8216;How the system works&#8217;</a>, namely that:</p><ul><li><p>A core AI value will be minimising risk.<strong> </strong>They should have a <em>very</em> strong bias towards only taking actions they know are safe. This will impede economic growth compared to other strategies, but the risk profile warrants it. In a world of AI-driven prosperity, the upside of risky behaviour will be pretty saturated, whereas the downside could be total<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-24" href="#footnote-24" target="_self">24</a>. Superintelligent AI cannot be allowed to &#8216;move fast and break things&#8217;.</p></li></ul><h2>My research process</h2><p>This post is part of my research <em><a href="https://www.workingthroughai.com/p/01-the-generating-process">Generating Process</a></em>, the previous steps of which have been:</p><ol><li><p><a href="https://www.workingthroughai.com/p/11-the-problem">The Problem</a></p></li><li><p><a href="https://www.workingthroughai.com/p/12-what-i-believe">What I Believe</a><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-25" href="#footnote-25" target="_self">25</a></p></li></ol><p>The next step, <em>Where We Are Now</em>, will be a deeper analysis of realistic paths forward.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>For more information about my research project, see the <a href="https://www.workingthroughai.com/i/175943893/my-research-process">Appendix</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>I supplement these with some optional assumptions in the <a href="https://www.workingthroughai.com/i/175943893/optional-details">Appendix</a>, which are not load-bearing for the scenario.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>I take intelligence to be <a href="https://www.workingthroughai.com/i/157964936/intelligence-is-generalised-knowing-how">generalised knowing-how</a>. That is, the ability to complete novel tasks. This is fairly similar to Francois Chollet&#8217;s definition: &#8216;skill-acquisition efficiency over a scope of tasks, with respect to priors, experience, and generalization difficulty&#8217;, although I put more emphasis on learned skills grounding the whole thing in a bottom-up way. Chollet&#8217;s paper <em><a href="https://arxiv.org/pdf/1911.01547">On the measure of intelligence</a></em> is a good overview of the considerations involved in defining intelligence.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>For similar reasons to those given by Nathan Lambert <a href="https://www.interconnects.ai/p/brakes-on-an-intelligence-explosion">here</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>I appreciate this will be very difficult to achieve, flying in the face of all of human history. I suspect that some kind of positive-sum, interest-respecting dynamic will need to be coded into the global political system &#8212; something that absolutely eschews all talk of one party or other &#8216;winning&#8217; an AI race, in favour of a vision of shared prosperity.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>Some people would call this &#8216;corrigibility&#8217;, but I&#8217;m not going to use this term because it has a hinterland and means different things to different people. If you want to learn more about an alignment solution that specifically prioritises corrigibility, see<a href="https://www.lesswrong.com/posts/NQK8KHSrZRF5erTba/0-cast-corrigibility-as-singular-target-1"> </a><em><a href="https://www.lesswrong.com/posts/NQK8KHSrZRF5erTba/0-cast-corrigibility-as-singular-target-1">Corrigibility as Singular Target</a></em> by Max Harms.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>This is not over-anthropomorphising it. It is saying that AI will expect to interact with humans in a certain way, and may act unpredictably if treated differently to those expectations. Perhaps a different word to &#8216;rights&#8217;, with less baggage, would be preferable to describe this though.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>Beren Millidge has written <a href="https://www.beren.io/2025-02-05-Maintaining-Alignment-During-RSI-As-A-Feedback-Control-Problem/">an interesting post</a> about seeing alignment as a feedback control problem, although I don&#8217;t know enough about control theory to tell you how well it could slot into my scheme.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-9" href="#footnote-anchor-9" class="footnote-number" contenteditable="false" target="_self">9</a><div class="footnote-content"><p>Beren Millidge has also <a href="https://www.beren.io/2025-08-02-Do-We-Want-Obedience-Or-Alignment/">written about</a> the tension between instruction-following and innate values or laws.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-10" href="#footnote-anchor-10" class="footnote-number" contenteditable="false" target="_self">10</a><div class="footnote-content"><p><a href="https://thezvi.substack.com/i/174921675/interpretability-investigations">Zvi recently said</a>: &#8216;If we want to build superintelligent AI, we need it to pass Who You Are In The Dark, because there will likely come a time when for all practical purposes this is the case. If you are counting on &#8220;I can&#8217;t do bad things because of the consequences when other minds find out&#8221; then you are counting on preserving those consequences.&#8217; My idea is to <em>both</em> build AI that passes Who You Are In The Dark and, given perfection is hard, permanently enforce consequences for bad behaviour.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-11" href="#footnote-anchor-11" class="footnote-number" contenteditable="false" target="_self">11</a><div class="footnote-content"><p>This will be easier if it is individual copies that tend to fail, rather than whole classes of AIs at once. There might be an argument here that copies failing leads to <a href="https://en.wikipedia.org/wiki/Antifragile_(book)">antifragility</a>, as some constant rate of survivable failures makes the system stronger and less likely to suffer catastrophic ones.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-12" href="#footnote-anchor-12" class="footnote-number" contenteditable="false" target="_self">12</a><div class="footnote-content"><p>Thank you <a href="https://www.lesswrong.com/posts/HfqbjwpAEGep9mHhc/the-plan-2023-version">John Wentworth</a> for making me aware of this.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-13" href="#footnote-anchor-13" class="footnote-number" contenteditable="false" target="_self">13</a><div class="footnote-content"><p>To &#8216;increase market penetration, maximise brand loyalty, and enhance intangible assets&#8217;.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-14" href="#footnote-anchor-14" class="footnote-number" contenteditable="false" target="_self">14</a><div class="footnote-content"><p>In extremis: &#8216;will this action kill everyone?&#8217;</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-15" href="#footnote-anchor-15" class="footnote-number" contenteditable="false" target="_self">15</a><div class="footnote-content"><p>It doesn&#8217;t make sense to me to assume AI will forever communicate with copies of itself using only natural language. The advantages of setting up higher-bandwidth channels are so obvious that I think any successful future must be robust to their existence.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-16" href="#footnote-anchor-16" class="footnote-number" contenteditable="false" target="_self">16</a><div class="footnote-content"><p><a href="https://x.com/MadHermitHimbo/status/1969611027638845613">This idea seems highly plausible to me</a>: &#8216;having worked with many different models, there is something about a model&#8217;s own output that makes it way more believable to the model itself even in a different instance. So a different model is required as the critiquer.&#8217; As mentioned before, if you doubt the future will be multipolar, feel free to ignore this bit. It&#8217;s not load-bearing on its own.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-17" href="#footnote-anchor-17" class="footnote-number" contenteditable="false" target="_self">17</a><div class="footnote-content"><p>I&#8217;m picturing a subjective experience like when, as a parent, you play with your child and let them make all the decisions. You&#8217;re just pleased to be there and do what you can to make them happy.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-18" href="#footnote-anchor-18" class="footnote-number" contenteditable="false" target="_self">18</a><div class="footnote-content"><p>Situations where, due to competitive pressures, humans don&#8217;t bother to try and understand their AIs as it slows down their pursuit of power, will be policed by the international institution for AI risk, and be strongly disfavoured by the AIs themselves. E.g. Bob is going to get annoyed at Alice, and potentially lodge a complaint, if she doesn&#8217;t bother to pay attention to his demonstrations.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-19" href="#footnote-anchor-19" class="footnote-number" contenteditable="false" target="_self">19</a><div class="footnote-content"><p>This failure could be explicit, through the emergence of serious misalignment, or implicit, as <a href="https://arxiv.org/abs/2501.16946">humans fade into irrelevance</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-20" href="#footnote-anchor-20" class="footnote-number" contenteditable="false" target="_self">20</a><div class="footnote-content"><p>That being said, the vast majority of people will be living far less consequentially than Alice.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-21" href="#footnote-anchor-21" class="footnote-number" contenteditable="false" target="_self">21</a><div class="footnote-content"><p>Plus the multipolar assumption, which is not.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-22" href="#footnote-anchor-22" class="footnote-number" contenteditable="false" target="_self">22</a><div class="footnote-content"><p><a href="https://blog.ai-futures.org/i/168032563/that-there-are-strict-speed-limits-to-ai-progress">There is an argument</a> which says that, even if it doesn&#8217;t perfectly generalise to new domains, AI will get so sample efficient that in practice it will quickly be able to master them. Much like arguments in favour of unlocking some levelled-up general purpose reasoning, this leaves me wondering how these meta-skills are supposed to be learnt. Sure, humans are more sample-efficient than current AI is, which proves becoming so is possible, but humans did not get this for free. We went through millions of years of evolution to produce a brain and body that can learn things about the world efficiently. In a sense, the structure of the world is imprinted in our architecture. AI is iterating in this direction &#8212; it is, after all, undergoing its own evolutionary process &#8212; but the only way to acquire this structure for itself is to learn it from something that has it, like doing real world tasks. <a href="https://sergeylevine.substack.com/p/sporks-of-agi">Anything less than this will have limits</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-23" href="#footnote-anchor-23" class="footnote-number" contenteditable="false" target="_self">23</a><div class="footnote-content"><p>I expand on my reasoning here at length in <em><a href="https://www.workingthroughai.com/p/superintelligent-ai-will-make-mistakes">Superintelligent AI will make mistakes</a>.</em></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-24" href="#footnote-anchor-24" class="footnote-number" contenteditable="false" target="_self">24</a><div class="footnote-content"><p>As Nassim Nicholas Taleb would say of systems with steep downsides and flattening upsides (what he would call concave) &#8212; they are fragile.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-25" href="#footnote-anchor-25" class="footnote-number" contenteditable="false" target="_self">25</a><div class="footnote-content"><p>I have developed the content of <em><a href="https://www.workingthroughai.com/p/12-what-i-believe">What I Believe</a></em> in some subsequent posts: <em><a href="https://www.workingthroughai.com/p/superintelligent-ai-will-make-mistakes">Superintelligent AI will make mistakes</a>, <a href="https://www.workingthroughai.com/p/making-alignment-a-law-of-the-universe">Making alignment a law of the universe</a>, </em>and <em><a href="https://www.workingthroughai.com/p/how-to-specify-an-alignment-target">How to specify an alignment target</a>.</em></p></div></div>]]></content:encoded></item><item><title><![CDATA[The Iceberg Theory of Meaning]]></title><description><![CDATA[Words are public, meaning is private]]></description><link>https://www.workingthroughai.com/p/the-iceberg-theory-of-meaning</link><guid isPermaLink="false">https://www.workingthroughai.com/p/the-iceberg-theory-of-meaning</guid><dc:creator><![CDATA[Richard Juggins]]></dc:creator><pubDate>Thu, 26 Jun 2025 11:10:24 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/b9fbafd0-a064-440e-b9b7-ab335bb76b9e_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>[This post is not explicitly about AI. But I think it is interesting, and will be a useful building block in future work.]</em></p><div><hr></div><p>When I was finishing my PhD thesis, there came an important moment where I had to pick an inspirational quote to put at the start. This was a big deal to me at the time &#8212; impressive books often carry impressive quotes, so aping the practice felt like being a proper intellectual. After a little thinking, though, I got it into my head that the <em>really </em>clever thing to do would be to subvert the concept and pick something silly. So I ended up using this from <em>The House At Pooh Corner </em>by A. A. Milne:</p><blockquote><p>When you are a Bear of Very Little Brain, and you Think of Things, you find sometimes that a Thing which seemed very Thingish inside you is quite different when it gets out into the open and has other people looking at it.</p></blockquote><p>It felt very relevant, in the thick of thesis writing, that the quote was both self-deprecating and about the difficulty of effective communication. And actually, the more I thought about it, the more I realised I hadn&#8217;t<em> </em>picked something silly after all. Rather, this passage brilliantly expresses a profound fact about meaning<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a>.</p><h2>The illusion of transparency</h2><p>We&#8217;ve all had that feeling, I think, of not being able to find the right word &#8212; that sense of grasping for something just out of reach. Of looking for the exact term that will capture the Thing that&#8217;s so Thingish in our mind and communicate it with complete fidelity. Yet, often, even when we find what we thought we wanted, it still doesn&#8217;t quite do the trick. </p><p>The fundamental problem is that language is a compression of whatever is going on in our minds. Each word or phrase is located in a personal web of meaning, which is high dimensional and unique to ourselves. So when we say something, we never actually pass on the full meaning we intend. There&#8217;s a whole world of context that stays put in our brain, getting approximated away in our public statement.</p><p>This is a ubiquitous source of communication failures. I&#8217;ve heard it referred to before as the <a href="https://www.lesswrong.com/w/illusion-of-transparency">&#8216;illusion of transparency&#8217;</a>:</p><blockquote><p>The illusion of transparency is the misleading impression that your words convey more to others than they really do. Words are a means of communication, but they don't in themselves contain meaning. The word apple is just five letters, two syllables. I use it to refer to a concept and its associations in my mind, under the reasonable assumption that it refers to a similar concept and group of associations in your mind; this is the only power words have, great though it may be. Unfortunately, it's easy to lose track of this fact, think as if your words have meanings inherently encoded in them, leading to a tendency to systematically overestimate the effectiveness of communication.</p></blockquote><p>As I&#8217;m writing this, I have to consider whether each word or phrase I put on the page will have a similar kind of meaning for you as it does for me. I know from experience that this is difficult to get right<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>. My task is not to create a standalone document, but an interactive one &#8212; to unlock for you, in your mind using your concepts and associations, the meaning I want to convey. Note that, if we do not share enough concepts, this will be impossible.</p><h2>Words are public, meaning is private</h2><p>I have a personal spin on this idea called the <em>Iceberg Theory of Meaning</em>.</p><p>As we&#8217;ve been discussing, when you say something to someone else, you don&#8217;t just want to pass on a bunch of words &#8212; you want to pass on a meaning. Let&#8217;s imagine the words in your statement as the top section of an iceberg. Lying above the water, these are public and agreed on by all observers. Meaning, though, lives beneath the waves. It lies in the bottom, much larger, section, corresponding to the rich collection of connected concepts and associations in your mind. These are private and will be different for each person.</p><p>The ocean, in this analogy, is your world model &#8212; the totality of your concepts and associations. Meaning can be considered the subset that get activated by the words in the statement.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!NEGK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd9c63bf2-328a-48ea-ba7d-0a15de0d8aa9_1660x785.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!NEGK!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd9c63bf2-328a-48ea-ba7d-0a15de0d8aa9_1660x785.png 424w, https://substackcdn.com/image/fetch/$s_!NEGK!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd9c63bf2-328a-48ea-ba7d-0a15de0d8aa9_1660x785.png 848w, https://substackcdn.com/image/fetch/$s_!NEGK!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd9c63bf2-328a-48ea-ba7d-0a15de0d8aa9_1660x785.png 1272w, https://substackcdn.com/image/fetch/$s_!NEGK!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd9c63bf2-328a-48ea-ba7d-0a15de0d8aa9_1660x785.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!NEGK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd9c63bf2-328a-48ea-ba7d-0a15de0d8aa9_1660x785.png" width="1456" height="689" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d9c63bf2-328a-48ea-ba7d-0a15de0d8aa9_1660x785.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:689,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:175927,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.workingthroughai.com/i/166531189?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd9c63bf2-328a-48ea-ba7d-0a15de0d8aa9_1660x785.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!NEGK!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd9c63bf2-328a-48ea-ba7d-0a15de0d8aa9_1660x785.png 424w, https://substackcdn.com/image/fetch/$s_!NEGK!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd9c63bf2-328a-48ea-ba7d-0a15de0d8aa9_1660x785.png 848w, https://substackcdn.com/image/fetch/$s_!NEGK!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd9c63bf2-328a-48ea-ba7d-0a15de0d8aa9_1660x785.png 1272w, https://substackcdn.com/image/fetch/$s_!NEGK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd9c63bf2-328a-48ea-ba7d-0a15de0d8aa9_1660x785.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Consider a statement to be represented by an iceberg. The words are public, floating above the water, and agreed on by all observers. However, the meaning lives below the surface. Each person will attach a unique set of concepts and associations to the words, drawing these from their world model. Strictly speaking, we should include additional public information in the visible part of the iceberg, such as environmental cues and the identity of the speaker/listener.</figcaption></figure></div><p>We can see from this how meaning is intrinsically subjective. The words may be the same for both of you, but as you have different brains, wired up in different ways, the meaning of them will not.</p><h2>Looking beneath the surface</h2><p>There is however a bit more we need to add. Words are not the only public information that accompanies a statement, and are consequently not the only things we condition on when constructing a meaning. There are two more pieces which belong in the top section of the iceberg, available to all observers:</p><ol><li><p>Environmental details, such as the medium, location, time of day, etc.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a></p></li><li><p>The identities of the speaker and the audience.</p></li></ol><p>The first point is fairly straightforward. Consider how a set of words spoken in a work context may imply a very different meaning to the same words said at a party. Just like before, we should note that the contextual information is itself meaningless &#8212; it is just a prompt that will activate a meaning in your mind.</p><p>The second point is more subtle. While identities may be public, and in that sense just another set of environmental variables, they have special significance as markers of the minds we are trying to exchange meaning with. They frame an active interpretative process, where we guess at each other&#8217;s world models and how they will shape the meaning of the statement.</p><p>As a speaker, this looks like trying to tailor our words to work with the concepts we believe the listener to have. This practice is most obvious when you consider the process of teaching.</p><p>As a listener, this looks like using our knowledge of the speaker to prime which concepts and associations are most likely to be relevant. If Alice is talking about &#8216;Justice&#8217;, we may know from past experience that she means something very different to what Bob would if he used the same words.</p><p>In both cases, we try to guess what meaning has or will form in the other person&#8217;s mind, beneath the surface of the water.</p><h2>The iceberg in action</h2><p>Sometimes, people make these inferences about each other very badly, leading to much wailing and gnashing of teeth. For instance:</p><ul><li><p>People intepret political comments from their opponents in extreme ways, attaching radically different meanings to those intended. A call to &#8216;tax the rich&#8217;, meant as a pragmatic suggestion to help balance the budget, might be interpreted as a jealous attack on wealth creators. Or, a plea to &#8216;prosecute shoplifters&#8217; may be heard as contempt for the poor rather than concern about a sharp increase in incidents.</p></li><li><p>People make comments that rely on unstated assumptions their listeners do not share, leading to much confusion. You see this in debates between people who possess conflicting sets of facts about the same topic, often without realising it. Perhaps one person thinks politician X is deliberately trying to run a government programme into the ground, whereas the other thinks they are working hard to save it. This happens in ethical conversations too, where people treat underlying beliefs, e.g. &#8216;oil is bad&#8217;, as unstated primitives in broader arguments that collapse on contact with a different set of primitives.</p></li><li><p>People have deeply unproductive arguments that use big, ambiguous terms in very different, yet load-bearing ways. An argument about whether capitalism is good will not usually go well if one person takes it to mean &#8216;exploiting the poor&#8217; while the other thinks it refers to &#8216;efficient markets generating wealth&#8217;. This often happens when one person is trying to decouple a narrow meaning of a term from a set of broader meanings, in order to discuss a specific aspect, but for the other person this simply isn&#8217;t possible &#8212; for them, the meanings are all too integrated.</p></li></ul><p>These kinds of communication failures are commonplace and under-acknowledged. Even if you understand they are possible, it is still hard to overcome them. Modelling others is difficult, particularly those with very different backgrounds to you. In fact, due to a kind of radical ignorance, doing so is often intractable. You can&#8217;t work to incorporate a different perspective &#8212; a different source of meaning &#8212; into your own if you don&#8217;t know it exists in the first place. Instead, you will just carry on arguing, deeply frustrated at the other person&#8217;s inability to understand what you mean.</p><div><hr></div><p>If you have any feedback, please leave a comment. Or, if you wish to give it anonymously, fill out my <a href="https://docs.google.com/forms/d/e/1FAIpQLSdyisSOndK1H1JT0NAbnA35LJgoJrl9f_NiJi1FEljCr7-kJg/viewform">feedback form</a>. Thanks!</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>It also led me to rediscover Winnie-the-Pooh, for which I am grateful. Not only are the Pooh books wickedly funny, but they possess joy and kindness to a degree I forgot for many years as a young adult. If that doesn&#8217;t qualify them as <em>impressive</em> and <em>profound,</em> I don&#8217;t know what would.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Indeed, <a href="https://www.workingthroughai.com/p/superintelligent-ai-will-make-mistakes">the first long piece</a> I published I ended up taking down and partially rewriting, as it was clear many people had drawn quite different conclusions from what I meant. My thesis was that superintelligent AI, while able to outcompete humans, will nevertheless make mistakes. But most commenters seemed to think I meant the mistake-making implied humans will win any conflict. I traced this error to certain phrases in my introduction, which I think primed some readers to interpret the whole piece from this angle.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>Strictly speaking, the speaker and listener might observe different things, so some of this information is hidden.</p></div></div>]]></content:encoded></item><item><title><![CDATA[How to specify an alignment target]]></title><description><![CDATA[What does a target even look like?]]></description><link>https://www.workingthroughai.com/p/how-to-specify-an-alignment-target</link><guid isPermaLink="false">https://www.workingthroughai.com/p/how-to-specify-an-alignment-target</guid><dc:creator><![CDATA[Richard Juggins]]></dc:creator><pubDate>Thu, 01 May 2025 20:40:08 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/4ab5a875-dc62-45ae-89d7-460a58e4fa33_1792x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>It&#8217;s pretty normal to chunk the alignment problem into two parts. One is working out how to align an AI to anything at all. You want to figure out how to control its goals and values, how to specify something and have it faithfully internalise it. The other is deciding which goals or values to actually pick &#8212; that is, finding the right alignment target. Solving the first problem is great, but it doesn&#8217;t really matter if you then align the AI to something terrible.</p><p>This split makes a fair amount of sense: one is a technical problem, to be solved by scientists and engineers; whereas the other is more a political or philosophical one, to be solved by a different class of people &#8212; or at least on a different day.</p><p>I&#8217;ve always found this distinction unsatisfying. Partly, this is because the problems are coupled &#8212; some targets are more practical to implement than others &#8212; and partly because, strategically, when you work on something, it makes sense to have some kind of end state in mind<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a>.</p><p>Here, I&#8217;m going to talk about a third aspect of the problem: what does an alignment target even look like? What different types are there? What components do you need to properly specify one? You can&#8217;t solve either of the two parts described above without thinking about this. You can&#8217;t judge whether your alignment technique worked without a clear idea of what you were aiming for, and you can&#8217;t pick a target without knowing how one is put together in the first place.</p><p>To unpack this, I&#8217;m going to build up the pieces as they appear to me, bit by bit, illustrated with real examples. I will be keeping this high-level, examining the practical components of target construction rather than, say, a deep interrogation of what goals or values are. Not because I don&#8217;t think the latter questions are important, I just want to sketch out the high-level concerns first.</p><p>There are many ways of cutting this cake, and I certainly don&#8217;t consider my framing to be definitive, but I hope it adds some clarity to what can be a confusing concept.</p><h2>Moving beyond ideas</h2><p>First of all, we need to acknowledge that an alignment target must be more than just a high-level idea. Clearly, stating that we want our AI to be truth-seeking or to have human values does not advance us very far. These are not well-defined categories. They are nebulous, context-dependent, and highly-subjective.</p><p>For instance, my own values may be very different to yours, potentially in ways that are hard for you to understand. This can still be true even when we use exactly the same words to describe them &#8212; who, after all, is opposed to fairness or goodness? My values are not especially coherent either, and depend somewhat on the kind of day I&#8217;m having. This is not me being weird &#8212; it is entirely typical.</p><p>To concretely define an alignment target requires multiple steps of clarification. We must be crystal clear what we mean, and this clarity must extend to all relevant actors. While ideas will always be our starting point, we need to flesh them out, making them legible and comprehensible to other people, before ultimately trying to create a faithful and reliable translation inside an AI.</p><p>To make this process more explicit, I like to picture it as a three-step recipe &#8212; a sequence you can follow to generate a target:</p><ol><li><p>Come up with a high-level idea (or ideas). Perhaps this is something broad like human values or more specific like helpfulness or obedience. While these terms are not well defined, you (and your collaborators) will know roughly what you mean by them.</p></li><li><p>Create a detailed specification. Take your idea and flesh it out, eliminating as much ambiguity as possible. Traditionally, this would be a written document, but I can&#8217;t think of a good reason why it couldn&#8217;t also include other modalities, like video. At this step, it is possible to have a meaningful public conversation about the merits of your proposed target.</p></li><li><p>Create a trainable encoding. Translate your specification into an appropriate format for your alignment process, which in theory could be automated from this point onwards. This is the last point of human influence over the target.</p></li></ol><p>One way of understanding this division is by audience: who will consume the output? Step (1) is for you and people close to you, step (2) is for society in general, and step (3) is for AI.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!zsgR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a20530e-ef95-42fa-8976-fbb2e2497bda_1322x326.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!zsgR!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a20530e-ef95-42fa-8976-fbb2e2497bda_1322x326.png 424w, https://substackcdn.com/image/fetch/$s_!zsgR!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a20530e-ef95-42fa-8976-fbb2e2497bda_1322x326.png 848w, https://substackcdn.com/image/fetch/$s_!zsgR!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a20530e-ef95-42fa-8976-fbb2e2497bda_1322x326.png 1272w, https://substackcdn.com/image/fetch/$s_!zsgR!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a20530e-ef95-42fa-8976-fbb2e2497bda_1322x326.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!zsgR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a20530e-ef95-42fa-8976-fbb2e2497bda_1322x326.png" width="1322" height="326" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9a20530e-ef95-42fa-8976-fbb2e2497bda_1322x326.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:326,&quot;width&quot;:1322,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:50423,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.workingthroughai.com/i/162269577?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a20530e-ef95-42fa-8976-fbb2e2497bda_1322x326.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!zsgR!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a20530e-ef95-42fa-8976-fbb2e2497bda_1322x326.png 424w, https://substackcdn.com/image/fetch/$s_!zsgR!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a20530e-ef95-42fa-8976-fbb2e2497bda_1322x326.png 848w, https://substackcdn.com/image/fetch/$s_!zsgR!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a20530e-ef95-42fa-8976-fbb2e2497bda_1322x326.png 1272w, https://substackcdn.com/image/fetch/$s_!zsgR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a20530e-ef95-42fa-8976-fbb2e2497bda_1322x326.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Three steps operationalising an alignment target. The boxes in blue are those containing information about the target, whereas the green box takes the target as an input.</figcaption></figure></div><p>What does this look like in practice? Well, let&#8217;s start with Anthropic&#8217;s example of <a href="https://arxiv.org/abs/2204.05862">training a helpful and harmless assistant</a> using reinforcement learning from human feedback (RLHF). Following our three steps, we can describe their alignment target as follows:</p><ol><li><p>Their high-level idea was that the model should be helpful and harmless.</p></li><li><p>They didn&#8217;t create a specification, preferring to stick with the high-level terms. This was a few years ago when these things were lower stakes, so they were more concerned with testing their alignment process than defining a rigorous target.</p></li><li><p>There are a few different parts to their trainable encoding, each a source of information about the target provided to the alignment process:</p><ol><li><p>They employed crowd workers to rate model responses for helpfulness and harmlessness. This way, they built up a dataset of ranked answers they could use to fine-tune &#8216;preference models&#8217;, which provide helpfulness and harmlessness ratings to the model during reinforcement learning.</p></li><li><p>To make them easier to train on the ranked answers, the preference models were first fine-tuned on public datasets containing preferences, like StackExchange, Reddit, and Wikipedia edits.</p></li><li><p>A helpfulness, harmlessness, and honesty (HHH) prompt, which they train into the model prior to the start of reinforcement learning.</p></li></ol></li></ol><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!gRkA!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e83d601-85f5-4fd1-90e0-ebe7140d9787_1322x326.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!gRkA!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e83d601-85f5-4fd1-90e0-ebe7140d9787_1322x326.png 424w, https://substackcdn.com/image/fetch/$s_!gRkA!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e83d601-85f5-4fd1-90e0-ebe7140d9787_1322x326.png 848w, https://substackcdn.com/image/fetch/$s_!gRkA!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e83d601-85f5-4fd1-90e0-ebe7140d9787_1322x326.png 1272w, https://substackcdn.com/image/fetch/$s_!gRkA!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e83d601-85f5-4fd1-90e0-ebe7140d9787_1322x326.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!gRkA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e83d601-85f5-4fd1-90e0-ebe7140d9787_1322x326.png" width="1322" height="326" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1e83d601-85f5-4fd1-90e0-ebe7140d9787_1322x326.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:326,&quot;width&quot;:1322,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:54530,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.workingthroughai.com/i/162269577?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e83d601-85f5-4fd1-90e0-ebe7140d9787_1322x326.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!gRkA!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e83d601-85f5-4fd1-90e0-ebe7140d9787_1322x326.png 424w, https://substackcdn.com/image/fetch/$s_!gRkA!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e83d601-85f5-4fd1-90e0-ebe7140d9787_1322x326.png 848w, https://substackcdn.com/image/fetch/$s_!gRkA!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e83d601-85f5-4fd1-90e0-ebe7140d9787_1322x326.png 1272w, https://substackcdn.com/image/fetch/$s_!gRkA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e83d601-85f5-4fd1-90e0-ebe7140d9787_1322x326.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Anthropic's alignment target for <a href="https://arxiv.org/abs/2204.05862">RLHF</a>.</figcaption></figure></div><p>We can see from this how integrated the process is. The <em>what</em>, helpfulness and harmlessness, are ultimately defined in terms of the <em>how</em>, the encoding required to get RLHF to work.</p><p>It&#8217;s important to note that the boundary I&#8217;ve drawn between the encoding and alignment process is fuzzy. In my description of Anthropic&#8217;s encoding, I only included information that had been explicitly provided to guide the alignment process. However, this ignores that most choices come packaged with implicit values. The pre-training dataset of the preference models<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>, for instance, will affect how they interpret the ranked answers, even though it wasn&#8217;t selected with this purpose in mind. Unfortunately, I have to draw a line somewhere, or we would end up saying the entire alignment process is part of the target, so I&#8217;ve done so at the explicit/implicit boundary.</p><p>Let&#8217;s look at another example, this time one with a proper detailed specification. In fact, let&#8217;s look at <a href="https://arxiv.org/abs/2212.08073">Constitutional AI</a>, Anthropic&#8217;s replacement for RLHF. This they deemed more transparent and scalable, and used it to align Claude. Applying our recipe again, we can break it down as follows:</p><ol><li><p>Their high-level idea was still to have a helpful and harmless model.</p></li><li><p>Now we have a specification. More specifically, <a href="https://www.anthropic.com/news/claudes-constitution">Claude's Constitution</a>, outlining a list of principles it should be expected to follow. Some of these draw on the UN Declaration of Human Rights and Apple&#8217;s terms of service, alongside many more bespoke ones.</p></li><li><p>Once again, there are a few different parts to the trainable encoding:</p><ol><li><p>Some human-ranked helpfulness data &#8212; in the original paper at least, the constitutional innovation was only applied to the harmfulness side of the alignment target.</p></li><li><p>The principles in the Constitution, which were repeatedly referenced by models during the alignment process, both to generate an initial supervised learning dataset to fine-tune the model on, as well as to create ranked answers that could be used with the human-ranked helpfulness data to train a perference model.</p></li><li><p>Harmfulness prompts used to generate the data for both the supervised and reinforcement learning steps.</p></li><li><p>A &#8216;helpful-only AI assistant&#8217;, which generated harmful content in response to the harmfulness prompts, and then revised it according to the Consitution in order to create the supervised learning dataset.</p></li></ol></li></ol><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!zHuR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8f74611-a627-46d6-ae8f-a9d33a79dafd_1322x366.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!zHuR!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8f74611-a627-46d6-ae8f-a9d33a79dafd_1322x366.png 424w, https://substackcdn.com/image/fetch/$s_!zHuR!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8f74611-a627-46d6-ae8f-a9d33a79dafd_1322x366.png 848w, https://substackcdn.com/image/fetch/$s_!zHuR!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8f74611-a627-46d6-ae8f-a9d33a79dafd_1322x366.png 1272w, https://substackcdn.com/image/fetch/$s_!zHuR!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8f74611-a627-46d6-ae8f-a9d33a79dafd_1322x366.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!zHuR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8f74611-a627-46d6-ae8f-a9d33a79dafd_1322x366.png" width="1322" height="366" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f8f74611-a627-46d6-ae8f-a9d33a79dafd_1322x366.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:366,&quot;width&quot;:1322,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:60568,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.workingthroughai.com/i/162269577?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8f74611-a627-46d6-ae8f-a9d33a79dafd_1322x366.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!zHuR!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8f74611-a627-46d6-ae8f-a9d33a79dafd_1322x366.png 424w, https://substackcdn.com/image/fetch/$s_!zHuR!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8f74611-a627-46d6-ae8f-a9d33a79dafd_1322x366.png 848w, https://substackcdn.com/image/fetch/$s_!zHuR!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8f74611-a627-46d6-ae8f-a9d33a79dafd_1322x366.png 1272w, https://substackcdn.com/image/fetch/$s_!zHuR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8f74611-a627-46d6-ae8f-a9d33a79dafd_1322x366.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Anthropic's alignment target for the original version of <a href="https://arxiv.org/abs/2212.08073">Constitutional AI</a>.</figcaption></figure></div><p>More recent work like <a href="https://arxiv.org/abs/2412.16339">Deliberative Alignment</a> from OpenAI follows a similar pattern, using a &#8216;<a href="https://model-spec.openai.com/2025-04-11.html">Model Spec</a>&#8217; instead of a constitution<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a>.</p><h2>Target-setting processes</h2><p>While our three-part system is nice, fleshing out what an alignment target can look like in practice, it is only the first part of the story, and there are other considerations we need to address. One of these is to go a bit meta and look at <em>target-setting processes.</em></p><p>At first glance, this might seem like a different topic. We&#8217;re interested in how to specify an alignment target, not how to select one. Unfortunately, this distinction is not so simple to make. Consider political processes as an analogy. While I might have a bunch of policy preferences I would like enacted, if you asked me how my country should be run I would not say &#8216;according to my policy preferences&#8217;. I would instead advocate for a policy-picking system, like a representative democracy. This could be an ideological preference, recognising that I should not hold special power, in which case the political process itself is effectively my highest policy preference. Or it could be a pragmatic one &#8212; I couldn&#8217;t force people to follow my commandments even if I wanted to, so I should concentrate my advocacy on something more widely acceptable. The same ideas apply to alignment targets.</p><p>Perhaps the key point is to recognise that the target-setting process is where the power in the system resides<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a>. A lot of discussions about alignment targets are really about power, about whose values to align the AI to. To properly contextualise what your alignment target is &#8212; to understand what the system around the AI is trying to achieve &#8212; you will need to specify your target-setting process as well<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a>.</p><p>To address this, I&#8217;m going add an extra layer to our schematic, acting from the outside to influence the structure and content of the other components:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!sazH!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84ab4b9f-b9a6-42e6-ac5b-e3c8e81b7548_1362x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!sazH!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84ab4b9f-b9a6-42e6-ac5b-e3c8e81b7548_1362x382.png 424w, https://substackcdn.com/image/fetch/$s_!sazH!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84ab4b9f-b9a6-42e6-ac5b-e3c8e81b7548_1362x382.png 848w, https://substackcdn.com/image/fetch/$s_!sazH!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84ab4b9f-b9a6-42e6-ac5b-e3c8e81b7548_1362x382.png 1272w, https://substackcdn.com/image/fetch/$s_!sazH!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84ab4b9f-b9a6-42e6-ac5b-e3c8e81b7548_1362x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!sazH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84ab4b9f-b9a6-42e6-ac5b-e3c8e81b7548_1362x382.png" width="1362" height="382" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/84ab4b9f-b9a6-42e6-ac5b-e3c8e81b7548_1362x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:382,&quot;width&quot;:1362,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:54717,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.workingthroughai.com/i/162269577?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84ab4b9f-b9a6-42e6-ac5b-e3c8e81b7548_1362x382.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!sazH!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84ab4b9f-b9a6-42e6-ac5b-e3c8e81b7548_1362x382.png 424w, https://substackcdn.com/image/fetch/$s_!sazH!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84ab4b9f-b9a6-42e6-ac5b-e3c8e81b7548_1362x382.png 848w, https://substackcdn.com/image/fetch/$s_!sazH!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84ab4b9f-b9a6-42e6-ac5b-e3c8e81b7548_1362x382.png 1272w, https://substackcdn.com/image/fetch/$s_!sazH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84ab4b9f-b9a6-42e6-ac5b-e3c8e81b7548_1362x382.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Any given alignment target is the result of a target-setting process.</figcaption></figure></div><p>Let&#8217;s look at some examples. First of all, what target-setting processes exist today? Broadly speaking, they consist of AI labs deciding what they think is best for their own AIs, with a bit of cultural and legal pressure thrown in. That being said, moving to transparent specifications, like Claude&#8217;s Constitution or OpenAI&#8217;s Model Spec, makes their alignment targets a bit more the product of a public conversation than in the past. For example, <a href="https://openai.com/index/sharing-the-latest-model-spec/">here</a> is OpenAI describing their emerging process for updating their Model Spec:</p><blockquote><p>In shaping this version of the Model Spec, we incorporated feedback from the first version as well as learnings from alignment research and real-world deployment. In the future we want to consider much more broad public input. To build out processes to that end, we have been conducting pilot studies with around 1,000 individuals &#8212; each reviewing model behavior, proposed rules and sharing their thoughts. While these studies are not reflecting broad perspectives yet, early insights directly informed some modifications. We recognize it as an ongoing, iterative process and remain committed to learning and refining our approach.</p></blockquote><p>This doesn&#8217;t sound formalised or repeatable yet, but it is moving in that direction.</p><p>Taking this further, let&#8217;s speculate about possible formal processes. Unsurprisingly, given the point is to adjudicate value questions, this can look quite political. For instance, it could involve an elected body like a parliament, either national or international, or something like a citizen&#8217;s assembly. The devil is, of course, in the detail, particularly as you need to decide how to guide the deliberations, and how to translate the results into a detailed specification and trainable encoding.</p><p>One proposal is Jan Leike&#8217;s suggestion for <a href="https://aligned.substack.com/p/a-proposal-for-importing-societys-values">Simulated Deliberative Democracy</a>.<em> </em>To address the issue of scalability, as AI becomes widely deployed and begins to operate beyond our competency level, Jan goes in heavy on AI assistance. In his own words:</p><blockquote><p>The core idea is to use imitation learning with large language models on deliberative democracy. Deliberative democracy is a decision-making or policy-making process that involves explicit deliberation by a small group of randomly selected members of the public (&#8216;mini-publics&#8217;). Members of these mini-publics learn about complex value-laden topics (for example national policy questions), use AI assistance to make sense of the details, discuss with each other, and ultimately arrive at a decision. By recording humans explicitly deliberating value questions, we can train a large language model on these deliberations and then simulate discussions on new value questions with the model conditioned on a wide variety of perspectives.</p></blockquote><p>Essentially, you convene a bunch of small assemblies of people, give them expert testimony (including AI assistance), and let them arrive at decisions on various value questions. You then train an AI on their deliberations, ending up with a system that can simulate new assemblies (of arbitrary identity groups) and scale up to answer millions of new questions. I would guess that these answers would be more than just high-level ideas, and would operate at the specification level, perhaps coming out looking like legislation. Some other protocol would be required to encode them and feed them into the alignment process.</p><p>The space of possible target-setting processes is large and under-explored. There just aren&#8217;t many experiments of this kind being done. Jan also makes the point that target setting is a <em>dynamic </em>process: &#8216;We can update the training data fairly easily... to account for changes in humanity&#8217;s moral views, scientific and social progress, and other changes in the world.&#8217; Which brings us to our next consideration.</p><h2>Keeping up with the times</h2><p>An appropriate target today will not necessarily be one tomorrow.</p><p>In a <a href="https://www.workingthroughai.com/p/making-alignment-a-law-of-the-universe">previous post</a>, I talked about how politeness is a dynamic environmental property:</p><blockquote><p>What counts as polite is somewhat ill-defined and changes with the times. I don&#8217;t doubt that if I went back in time two hundred years I would struggle to navigate the social structure of Jane Austen&#8217;s England. I would accidentally offend people and likely fail to win respectable friends. Politeness can be seen as a property of my particular environment, defining which of my actions will be viewed positively by the other agents in it.</p></blockquote><p>Jane Austen&#8217;s England had a social system suited to the problems it was trying to solve, but from our perspective in the early 21<sup>st</sup> century it was arbitrary and unjust. We are trying to solve different problems, in a different world, using different tools and knowledge, and our value system is correspondingly different.</p><p>The concept of superhuman AI, or one stage of it, is sometimes referred to as &#8216;transformational&#8217; AI. The world will be a very different place after it comes about, for good or ill. Whatever values we have now will not be the same as we will have after.</p><p>This suggests that, arguably, figuring out how to update your target is just as important as setting it in the first place. A critical consideration when doing this is that, if our AI is now broadly superhuman, it may be much harder to refine the target than before. If our AI is operating in a world beyond our understanding, and is reshaping it in ways equally hard for us to follow, then we cannot just continue to rerun whatever target-setting process we had previously.</p><p>Bearing this in mind, I think there are a few different ways of categorising the dynamics of alignment targets. I&#8217;ve settled on three I think are useful to recognise:</p><ol><li><p>Static targets: humans define a target once and leave it for the rest of time. E.g. write a single constitution and keep it. Whatever values are enshrined in it will be followed by the AI forever.</p></li></ol><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!QnEH!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd412d93c-5379-4e80-8d5b-46e7531aa625_582x342.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!QnEH!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd412d93c-5379-4e80-8d5b-46e7531aa625_582x342.png 424w, https://substackcdn.com/image/fetch/$s_!QnEH!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd412d93c-5379-4e80-8d5b-46e7531aa625_582x342.png 848w, https://substackcdn.com/image/fetch/$s_!QnEH!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd412d93c-5379-4e80-8d5b-46e7531aa625_582x342.png 1272w, https://substackcdn.com/image/fetch/$s_!QnEH!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd412d93c-5379-4e80-8d5b-46e7531aa625_582x342.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!QnEH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd412d93c-5379-4e80-8d5b-46e7531aa625_582x342.png" width="344" height="202.1443298969072" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d412d93c-5379-4e80-8d5b-46e7531aa625_582x342.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:342,&quot;width&quot;:582,&quot;resizeWidth&quot;:344,&quot;bytes&quot;:33721,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.workingthroughai.com/i/162269577?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd412d93c-5379-4e80-8d5b-46e7531aa625_582x342.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!QnEH!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd412d93c-5379-4e80-8d5b-46e7531aa625_582x342.png 424w, https://substackcdn.com/image/fetch/$s_!QnEH!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd412d93c-5379-4e80-8d5b-46e7531aa625_582x342.png 848w, https://substackcdn.com/image/fetch/$s_!QnEH!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd412d93c-5379-4e80-8d5b-46e7531aa625_582x342.png 1272w, https://substackcdn.com/image/fetch/$s_!QnEH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd412d93c-5379-4e80-8d5b-46e7531aa625_582x342.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Static targets. Set it once and leave it forever. The 'Static alignment target' box corresponds to our three-part recipe from earlier.</figcaption></figure></div><ol start="2"><li><p>Semi-static targets: humans, on one single occasion, specify a dynamic target-setting process for the AI to follow, which it then follows forever. The AI can collect data by observing and speaking to us, which allows us to retain some limited influence over the future, but we are not primarily in control any more. E.g. we could tell the AI to value whatever we value, which will change as we change, and it will figure this out by itself.</p></li></ol><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!osH9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc996637a-8073-43c9-9287-7345814f2938_922x342.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!osH9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc996637a-8073-43c9-9287-7345814f2938_922x342.png 424w, https://substackcdn.com/image/fetch/$s_!osH9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc996637a-8073-43c9-9287-7345814f2938_922x342.png 848w, https://substackcdn.com/image/fetch/$s_!osH9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc996637a-8073-43c9-9287-7345814f2938_922x342.png 1272w, https://substackcdn.com/image/fetch/$s_!osH9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc996637a-8073-43c9-9287-7345814f2938_922x342.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!osH9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc996637a-8073-43c9-9287-7345814f2938_922x342.png" width="554" height="205.49674620390456" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c996637a-8073-43c9-9287-7345814f2938_922x342.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:342,&quot;width&quot;:922,&quot;resizeWidth&quot;:554,&quot;bytes&quot;:52434,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.workingthroughai.com/i/162269577?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc996637a-8073-43c9-9287-7345814f2938_922x342.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!osH9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc996637a-8073-43c9-9287-7345814f2938_922x342.png 424w, https://substackcdn.com/image/fetch/$s_!osH9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc996637a-8073-43c9-9287-7345814f2938_922x342.png 848w, https://substackcdn.com/image/fetch/$s_!osH9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc996637a-8073-43c9-9287-7345814f2938_922x342.png 1272w, https://substackcdn.com/image/fetch/$s_!osH9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc996637a-8073-43c9-9287-7345814f2938_922x342.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Semi-static targets. The blue boxes are still the bits we set, the green are AI-controlled or generically automated. Basically, each time through the loop, the AI has to continue following our plan for setting updated targets. It uses interactions with us as data to help, allowing us to retain some limited ongoing influence.</figcaption></figure></div><ol start="3"><li><p>Fully dynamic targets: humans are continuously in meaningful control of the alignment target. If we want to tell the AI to stop doing something it is strongly convinced we want, or to radically change its values, we can. However, in the limit of superhuman AI, we will still need assistance from it to understand the world and to properly update its target. Put another way, while humans are technically in control, we will nevertheless require a lot of AI mediation.</p></li></ol><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!cfgX!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffae90a5a-1592-424a-986b-602d5ce028cb_582x562.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!cfgX!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffae90a5a-1592-424a-986b-602d5ce028cb_582x562.png 424w, https://substackcdn.com/image/fetch/$s_!cfgX!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffae90a5a-1592-424a-986b-602d5ce028cb_582x562.png 848w, https://substackcdn.com/image/fetch/$s_!cfgX!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffae90a5a-1592-424a-986b-602d5ce028cb_582x562.png 1272w, https://substackcdn.com/image/fetch/$s_!cfgX!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffae90a5a-1592-424a-986b-602d5ce028cb_582x562.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!cfgX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffae90a5a-1592-424a-986b-602d5ce028cb_582x562.png" width="360" height="347.62886597938143" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fae90a5a-1592-424a-986b-602d5ce028cb_582x562.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:562,&quot;width&quot;:582,&quot;resizeWidth&quot;:360,&quot;bytes&quot;:82363,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.workingthroughai.com/i/162269577?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffae90a5a-1592-424a-986b-602d5ce028cb_582x562.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!cfgX!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffae90a5a-1592-424a-986b-602d5ce028cb_582x562.png 424w, https://substackcdn.com/image/fetch/$s_!cfgX!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffae90a5a-1592-424a-986b-602d5ce028cb_582x562.png 848w, https://substackcdn.com/image/fetch/$s_!cfgX!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffae90a5a-1592-424a-986b-602d5ce028cb_582x562.png 1272w, https://substackcdn.com/image/fetch/$s_!cfgX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffae90a5a-1592-424a-986b-602d5ce028cb_582x562.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Dynamic targets. Humans are continuously in control. However, we require AI assistance, both to understand a world full of AIs smarter than us, and to translate our targets into things meaningful in this world. I&#8217;ve indicated this by blurring the colours in the &#8216;AI-mediated&#8217; boxes, implying that we collaboratively create these things.</figcaption></figure></div><p>The distinctions between these target types are not sharp. Looked at in a certain way, (1) and (2) are the same &#8212; you seed the AI with initial instructions and let it go &#8212; and looked at in another, (2) and (3) are the same &#8212; the AI consistently references human preferences as it updates its values. But I think there are still a lot of useful differences to highlight. For instance, in (2), what the AI values can change drastically in an explicit way, whereas in (1) it cannot. In (3), humans retain fundamental control and can flip the gameboard at any time, whereas in (2) we&#8217;re being kind of looked after. It is worth noting that (3) implies the AI will allow you to change its target, whereas (1) does not, and (2) is noncommital.</p><p>Let&#8217;s look at some examples of each:</p><ol><li><p>Static targets:</p><ol><li><p>Have a worldwide conversation or vote where we collectively agree a definitive set of human values we want the AI to respect forever.</p></li><li><p>Solve moral philosophy<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a> or tell the AI to do it (and trust its answer).</p></li><li><p>Society, or simply the lab building the AI, has a dynamic process for selecting targets, which is followed each time a new model is trained. However, at some point the AI becomes superhuman and the lab&#8217;s alignment process is insufficient to cope with this, locking in the last target forever. One way of thinking about this situation is that dynamic targets can become static when they fail.</p></li></ol></li><li><p>Semi-static targets:</p><ol><li><p>Eliezer Yudkowsky&#8217;s <a href="https://www.lesswrong.com/w/coherent-extrapolated-volition-alignment-target">Coherent extrapolated volition (CEV)</a>: </p><blockquote><p>Roughly, a CEV-based superintelligence would do what currently existing humans would want the AI to do, if counterfactually:</p><ol><li><p>We knew everything the AI knew;</p></li><li><p>We could think as fast as the AI and consider all the arguments;</p></li><li><p>We knew ourselves perfectly and had better self-control or self-modification ability.</p></li></ol></blockquote><p>This functions as a protocol for the superintelligent AI to follow as it updates its own values. There are plenty of variations on this idea &#8212; that superintelligent AI can learn a better model of our values than we can, so we should hand over control and let it look after us.</p></li></ol></li><li><p>Fully dynamic targets:</p><ol><li><p>Develop an AI that <a href="https://www.lesswrong.com/posts/7NvKrqoQgJkZJmcuD/instruction-following-agi-is-easier-and-more-likely-than">reliably and faithfully follows your instructions</a><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a>. This collapses onto a static target in the limit where you never ask the AI to change its own alignment. You might question why you would want to do that, but it may well be the case that whatever technique worked to create instruction-following AI at one level of intelligence will cease to work as it gets smarter (and as the environment changes), necessitating corrections.</p></li><li><p>Jan Leike&#8217;s <a href="https://aligned.substack.com/p/a-proposal-for-importing-societys-values">Simulated Deliberative Democracy</a>: each time through the loop, you run some human mini-publics with AI assistance to update the training data. This could be done on a regular cadence or as and when problems in the current value-set are flagged.</p></li><li><p>Beren Millidge <a href="https://www.beren.io/2025-02-05-Maintaining-Alignment-During-RSI-As-A-Feedback-Control-Problem/">has written</a> about how we should treat alignment as a feedback control problem. Starting with a reasonably well-aligned model, we could apply control theory-inspired techniques to iteratively steer an AI as it undergoes recursive self-improvement.</p></li></ol></li></ol><h2>What next?</h2><p>My particular interest in this problem dates from a conversation I had with some AI safety researchers a few years ago, during which I realised they each had very different ideas of what it meant to pick an alignment target. I spent a long while feeling confused as to who was right, and found writing the material for this post a really effective way of deconfusing myself. In particular, it lent structure to the different assumptions they were making, so I could see them in context. It has also helped me see which bits of my own ideas feel most important, and in what places they are lacking.</p><p>On that point, I will briefly comment on my preferred direction, albeit one still stuck in the high-level idea phase. I think you can build a dynamic target around the idea of AI having a moral role in our society. It will have a set of responsibilities, certainly different from human ones (and therefore requiring it to have different, but complementary, values to humans), which situate it in a symbiotic relationship with us, one in which it desires continuous feedback<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a>. I will develop this idea more in the future, filling out bits of the schema in this post as I go.</p><div><hr></div><p>This post is an aside from my research agenda, so is not numbered. You can think of it as me expanding on some of the points in <em><a href="https://www.workingthroughai.com/p/12-what-i-believe">What I Believe</a> </em>in a way that I think will help me with the next step, <em><a href="https://www.workingthroughai.com/p/13-what-success-looks-like">What Success Looks Like</a>.</em></p><p>If you have any feedback, please leave a comment. Or, if you wish to give it anonymously, fill out my <a href="https://docs.google.com/forms/d/e/1FAIpQLSdyisSOndK1H1JT0NAbnA35LJgoJrl9f_NiJi1FEljCr7-kJg/viewform">feedback form</a>. Thanks!</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>This is why I advocate defining <a href="https://www.workingthroughai.com/p/01-the-generating-process">What Success Looks Like</a> fairly early on in any research agenda into risks from superhuman AI. I hope to reach this step in my own agenda reasonably soon!</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>By this I mean the pre-training data of the base model, not what they refer to in the paper as &#8216;preference model pretraining&#8217;, which I included earlier as part (b) of the trainable encoding.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>Although, the alignment process is <a href="https://x.com/boazbaraktcs/status/1870369979369128314">a little different</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>Assuming the alignment problem is solved, anyway.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>This will become more obvious later when we talk about dynamics.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>I want to put on the record that <a href="https://www.workingthroughai.com/p/12-what-i-believe">I believe</a> moral philosophy is neither solvable nor the right system to use to address important value questions. Morality is primarily a practical discipline, not a theoretical one.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>I have my suspicions that, in order for the AI to reliably do what you mean rather than merely what you say, this actually cashes out as functionally the same as aligning the AI to your values. More specifically, when you ask an AI to do something, what you really mean is &#8216;do this thing, but in the context of my broader values&#8217;. The AI can only deal with ambiguity, edge cases, and novelty by referring to a set of values. And you can&#8217;t micromanage if it is trying to solve problems you don&#8217;t understand.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>If you&#8217;ve just read that and thought &#8216;well that&#8217;s vague and not particularly useful&#8217;, you&#8217;ll maybe see why I designated the audience for the high-level idea step as the author themselves.</p></div></div>]]></content:encoded></item><item><title><![CDATA[Superintelligent AI will make mistakes]]></title><description><![CDATA[Things are going to get messy]]></description><link>https://www.workingthroughai.com/p/superintelligent-ai-will-make-mistakes</link><guid isPermaLink="false">https://www.workingthroughai.com/p/superintelligent-ai-will-make-mistakes</guid><dc:creator><![CDATA[Richard Juggins]]></dc:creator><pubDate>Sun, 02 Mar 2025 11:08:54 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/740f5563-115b-42fb-aa80-821a92c08b52_1792x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>[This is a slight rework of an earlier post, which I realised had a misleading introduction. If you read that one, this will seem very similar.]</em></p><div><hr></div><p>In the classical presentation of the alignment problem, superintelligent AI is basically omnipotent. This is a key point you have to internalise to understand writing on the subject. It will be smarter than you to the same degree that you are smarter than an ant. It won&#8217;t just master nanorobotics or whatever, it will transcend technology, reaching behind the cosmological veil to truer forms of power. </p><p><a href="https://www.workingthroughai.com/p/12-what-i-believe">I believe</a> that AI surpassing human intelligence is likely and may happen soon. It will have extraordinary abilities, and be more than capable of winning any conflict with us. But I think stories of effective omnipotence skip over part of the problem. I want to discuss the risks from superintelligent <em>incompetence</em>. While it is obvious that current AI is fallible, I am going to argue that this will remain the case even after it can dramatically outperform humans on all tasks.<strong> </strong>Regular and consequential mistakes should be expected by default<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a>.</p><h2>How does superintelligence happen?</h2><p>Fundamental to the superintelligence story is the idea that there is a distinct thing called &#8216;intelligence&#8217; and that you can turn this parameter up to essentially infinity. A high-profile version roughly follows this pattern<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>:</p><ol><li><p>Someone, somehow builds an AI that has more intelligence than any human<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a>.</p></li><li><p>The AI uses its greater intelligence to recursively self-improve: it modifies itself to be smarter or builds a smarter replacement with the same goals, and it does this really, really fast<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a>.</p></li><li><p>The AI is now omnipotent.</p></li></ol><p>I think this narrative makes a serious over-approximation: it assumes the AI has basically no negative impact on the world while it is scaling up. <a href="https://www.lesswrong.com/posts/pRkFkzwKZ2zfa3R6H/without-specific-countermeasures-the-easiest-path-to">Sometimes</a> this assumption takes the form of a scheme, where the AI quietly builds up dangerous capabilities, releasing them only when it knows it is strong enough to take over the world. My objection is that this leaves no room for it to make mistakes &#8212; it assumes that once the human-level intelligence threshold is crossed, essentially everything significant that the AI attempts will be successful. This story is <em>too clean</em>. My reasoning for this can be summarised as follows:</p><ol><li><p>Intelligence is generalised &#8216;knowing-how&#8217; acquired by taking actions in the world.</p></li><li><p>The process of learning to complete tasks, and thereby gain intelligence, involves a significant amount of failure.</p></li><li><p>Even once an AI is demonstrably reliable on a wide range of beyond-human-level tasks, there will always be another, harder, set of problems for it to fail at.</p></li><li><p>While acquiring intelligence through completing purely simulatable tasks will generalise somewhat to real ones, there will be a limit to this, even when generalisation is very strong.</p></li></ol><p>I&#8217;ll now go through each of these points in more detail.</p><h2>Intelligence is generalised &#8216;knowing-how&#8217;</h2><p>There&#8217;s a distinction in philosophy between &#8216;knowing-that&#8217;, e.g. London is the capital of England, and &#8216;knowing-how&#8217;, like riding a bike. I&#8217;m not going to get into the philosophical debate around this, but I want to put forward the claim that what we commonly mean by intelligence is <em>all</em> downstream of knowing-how. And that it is learnt through practice, by taking actions in the world<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a>.</p><p>If I asked you to name examples of intelligence, there are lots of different kinds of things you might pick. When I asked ChatGPT, it gave me eight distinct categories (Claude gave five), with a variety of choices including solving maths problems, juggling, composing music, and empathising with a friend. One thing these all have in common is that they are, in a sense, tasks. They involve achieving something real, interacting with the world outside of your own brain.</p><p>As we navigate our environment, we are constantly presented with tasks to solve. As a baby, these will be very basic, like figuring out what our hands do. But as we get older they become more complex, usually building on and extending skills we have previously learnt (hence <em>generalised</em> knowing-how). And this learning process is dominated by practice. It is by listening and trying to make sounds that we learn to speak; by doing problem sets at school that we learn arithmetic; by predicting and reacting to other people&#8217;s behaviour that we learn emotional intelligence. In each case, our brains build up pathways representing the skill we are learning, allowing us both easy reuse and a head-start on any related new skills. We are creating mappings from situations to actions, that help us better achieve our goals.</p><p>Evolution was doing something similar when it created us. To succeed in our environment and pass on our genes, we needed the ability to complete a variety of tasks. We needed to hunt, forage, find shelter, and escape from predators. And we needed to be able to learn new skills quickly in order to adapt to change. Humans evolved brains suitable for achieving this &#8212; we evolved to be mappings from environmental inputs to adaptive outcomes. That you were born with the ability to learn mathematics is because evolution trained the human species to solve novel abstract reasoning problems, like making better tools. And it did so because people less capable at this were less likely to pass on their genes.</p><p>It&#8217;s actually when thinking about machine learning that the primacy of knowing-how is easiest to see. When you set up a machine learning model, you first define a task for it to learn. This might be forecasting the values of a share price, distinguishing between photos of cats and dogs, or predicting what product a customer will buy next. The model then figures out the parameters for itself &#8212; it learns the mapping from the training data to the task labels.</p><p>Large language models learn how to complete sequences of tokens. All of their knowledge and utility is downstream of this. Next-token-prediction trained over the internet turns out to be pretty useful for solving a wide variety of problems, and relatively easy to supplement with fine-tuning (subtly modifying the mappings to work better on certain tasks), which is why machine learning researchers &#8216;evolved&#8217; LLMs out of the previous generation of language models.</p><p>It is interesting that the machine learning community generally talks about <em>capabilities</em> rather than intelligence. That is, the primary measure of progress is in task completion ability &#8212; in getting higher scores on benchmarks &#8212; with the notion of model intelligence a kind of generality vibe tacked on at the end.</p><p>There is an important implication from all this. If the key process in gaining intelligence is learning the mapping required to turn a context into a useful result, then superintelligent capabilities are defined by complex mappings that are beyond human abilities to learn. And the only way an AI can learn these is by practicing doing super-advanced tasks.</p><h2>Learning requires failure</h2><p>In all the examples of learning I gave above, from childhood arithmetic to large language models, one ubiquitous element of the process is failure. When a person or a model practices a skill, they do not get it right every time.</p><p>Granted, there are situations where previously learnt capabilities allow you to succeed on a novel task first time, like multiplying two numbers you&#8217;ve never tried to multiply together before. But you can do these because of previous learning where you did pay the cost. You learnt an algorithm for multiplying numbers together that you could reuse, and you did so through practice. Furthermore, evolution paid a cost training you to have the learning algorithm that allowed you to do this in the first place.</p><p>It has to be this way. To have the mapping in your brain corresponding to a particular capability, you need to acquire that information from somewhere. You have to extract it from the world, bit by bit, by taking your best guess and observing the results. And by definition, if you don&#8217;t already have the capability then your best guess will not be good.</p><p>This process of trial and error is so intrinsic to machine learning it seems almost stupid writing it down. Neural networks learn by doing gradient descent on a loss function &#8212; in other words, by gradually correcting their outputs to be less wrong than they were before. A model builds the mapping it needs by taking its best guess, observing how wildly it missed, and then updating itself to make a better guess next time. Superintelligent AI will still have to do this<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a>. If it wants to learn a new capability, it will need to take actions in the world (or in a good enough simulation) and update its mapping from inputs to outputs based on the results it gets. In the process, it will fail many times.</p><h2>There&#8217;s always a harder problem</h2><p>Back in 2018, researchers from Stanford released a reading comprehension benchmark called &#8216;A Conversational Question Answering Challenge&#8217;<em> </em>or <a href="https://arxiv.org/pdf/1808.07042">CoQA</a>. The idea of this benchmark was to test whether language models could answer questions of the following format:</p><blockquote><p>Jessica went to sit in her rocking chair. Today was her birthday and she was turning 80. Her granddaughter Annie was coming over in the afternoon and Jessica was very excited to see her. Her daughter Melanie and Melanie&#8217;s husband Josh were coming as well. Jessica had . . .</p><p>Q1: Who had a birthday?</p><p>A1: Jessica</p></blockquote><p>When GPT-2 was released a year later, CoQA was one of the benchmarks OpenAI tested it on. It scored 55 out of 100<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a>. This wasn&#8217;t as good as some models, but was impressive on account of GPT-2 not having been fine-tuned for the task. Then, a year later, when <a href="https://arxiv.org/pdf/2005.14165">GPT-3</a> dropped, OpenAI once again chose to highlight the progress they had made on CoQA. Now the model was almost at human-level with a score of 85, and very close to the best fine-tuned models.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!XErV!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F520f229c-2708-463e-aff4-dc47fdce4e5b_823x535.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!XErV!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F520f229c-2708-463e-aff4-dc47fdce4e5b_823x535.png 424w, https://substackcdn.com/image/fetch/$s_!XErV!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F520f229c-2708-463e-aff4-dc47fdce4e5b_823x535.png 848w, https://substackcdn.com/image/fetch/$s_!XErV!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F520f229c-2708-463e-aff4-dc47fdce4e5b_823x535.png 1272w, https://substackcdn.com/image/fetch/$s_!XErV!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F520f229c-2708-463e-aff4-dc47fdce4e5b_823x535.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!XErV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F520f229c-2708-463e-aff4-dc47fdce4e5b_823x535.png" width="823" height="535" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/520f229c-2708-463e-aff4-dc47fdce4e5b_823x535.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:535,&quot;width&quot;:823,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:84457,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.workingthroughai.com/i/157964936?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F520f229c-2708-463e-aff4-dc47fdce4e5b_823x535.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!XErV!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F520f229c-2708-463e-aff4-dc47fdce4e5b_823x535.png 424w, https://substackcdn.com/image/fetch/$s_!XErV!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F520f229c-2708-463e-aff4-dc47fdce4e5b_823x535.png 848w, https://substackcdn.com/image/fetch/$s_!XErV!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F520f229c-2708-463e-aff4-dc47fdce4e5b_823x535.png 1272w, https://substackcdn.com/image/fetch/$s_!XErV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F520f229c-2708-463e-aff4-dc47fdce4e5b_823x535.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>GPT-3&#8217;s performance on CoQA. Stacking more layers is great for reading comprehension.</em></figcaption></figure></div><p>This has been a common sight since the start of the deep learning revolution in the 2010s. Researchers build benchmarks they think are difficult, only for models to breeze past them in a couple of years. Each time, a renewed effort is made to build a much, much harder benchmark that AI won&#8217;t be able to crack so easily, only for it to suffer the same fate just as quickly.</p><p>One example of such a benchmark is <a href="https://arcprize.org/arc">ARC-AGI</a>. This was put together in 2019 by Francois Chollet to specifically test &#8216;intelligence&#8217;, which in his words meant &#8220;skill-acquisition efficiency over a scope of tasks, with respect to priors, experience, and generalization difficulty&#8221;. The benchmark consists of grids of different coloured blocks, and the challenge is to spot the pattern in some examples, and reproduce it on a test case. There are many different patterns across all the examples. These are pretty easy for humans, but harder for AI. GPT-3 scored 0%.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!dIav!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc121bc03-6a5d-4d30-9461-c3a50720ce50_973x299.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!dIav!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc121bc03-6a5d-4d30-9461-c3a50720ce50_973x299.png 424w, https://substackcdn.com/image/fetch/$s_!dIav!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc121bc03-6a5d-4d30-9461-c3a50720ce50_973x299.png 848w, https://substackcdn.com/image/fetch/$s_!dIav!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc121bc03-6a5d-4d30-9461-c3a50720ce50_973x299.png 1272w, https://substackcdn.com/image/fetch/$s_!dIav!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc121bc03-6a5d-4d30-9461-c3a50720ce50_973x299.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!dIav!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc121bc03-6a5d-4d30-9461-c3a50720ce50_973x299.png" width="973" height="299" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c121bc03-6a5d-4d30-9461-c3a50720ce50_973x299.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:299,&quot;width&quot;:973,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:26394,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.workingthroughai.com/i/157964936?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe93da898-39d6-4197-97ff-289149927f64_973x299.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!dIav!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc121bc03-6a5d-4d30-9461-c3a50720ce50_973x299.png 424w, https://substackcdn.com/image/fetch/$s_!dIav!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc121bc03-6a5d-4d30-9461-c3a50720ce50_973x299.png 848w, https://substackcdn.com/image/fetch/$s_!dIav!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc121bc03-6a5d-4d30-9461-c3a50720ce50_973x299.png 1272w, https://substackcdn.com/image/fetch/$s_!dIav!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc121bc03-6a5d-4d30-9461-c3a50720ce50_973x299.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>You can try some ARC-AGI problems yourself on the <a href="https://arcprize.org/">ARC Prize website</a>.</em></figcaption></figure></div><p>For a number of years this benchmark held strong. GPT-4o scored just 5%, and as of mid-2024 the best result by any model was still only 34.3%. Then, with a crushing sense of inevitability, in December 2024 <a href="https://arcprize.org/blog/oai-o3-pub-breakthrough">Chollet announced</a> that OpenAI&#8217;s latest reasoning model, o3, had scored an enormous 87.5%<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a>. Chollet has already begun working on ARC-AGI-2, an even harder version.</p><p>There&#8217;s no obvious reason why this process is going to stop after AI surpasses human intelligence. While Francois Chollet will not be able to keep building ARC-AGI-Ns forever, our new state-of-the-art artificial AI researchers will. Their benchmarks will be full of tasks difficult for us to conceive of, but comprehensible to other AIs.</p><p>On the one hand, this story is really bullish about AI progress. Benchmark after benchmark has fallen and will continue to fall as the the machines rise up to greatness. On the other, though, it shows AI repeatedly failing at things. And when it finally succeeds on a benchmark, another, harder once is produced for it to fail at.</p><p>The question is whether this pattern of failure will continue, whether AI will keep promoting itself to a new level of incompetence. Really, answering this comes down to what level of capabilities constitutes mastery of the universe. If this point gets reached just above human-level, then slightly superhuman AI will ace everything with 100% reliability. But this would be incredibly arbitrary. For all the richness and complexity of the universe, it would be strange if merely optimising over the human ancestral environment created <em>almost</em> enough intelligence to unlock the deepest mysteries of the cosmos. Certainly, for stories about godlike superintelligence to make sense, there would need to be a lot of gears left for the AI to go through.</p><p>It seems likely, then, that even strongly superhuman AI will face problems it will struggle with. And this is where the danger creeps in. There&#8217;s a scene in <em>Oppenheimer</em> where the Manhattan project physicists are worried about the possibility that a nuclear detonation will ignite the atmosphere and destroy the world. While their fear was exaggerated somewhat in the film, it illustrates how when you dial up the power and sophistication of the actors in play, you begin to flirt with the catastrophic. Who knows what unfathomable scientific experiments advanced AI will get up to<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-9" href="#footnote-9" target="_self">9</a>, and what their score will be on &#8216;Big And Dangerous Intelligence-Demanding Experiments Assessment&#8217;<em> </em>(BAD-IDEA)<em>. </em>Spoiler: it won&#8217;t be 100%.</p><h2><strong>Generalisation won&#8217;t be perfect</strong></h2><p>So far, we have built up an argument as to why AI recursively self-improving will be, at the very least, rather messy. To acquire new capabilities, AI will need to take actions in the world, many of which will fail, and this problem will continue indefinitely, even when the AI is superintelligent and doing incomprehensibly difficult stuff. There is, however, one possible route around this set of issues: strong generalisation.</p><p>When I talk about intelligence being <em>generalised</em> knowing-how, what I mean by this is that a lot of skills are correlated. This is why the single term intelligence is useful in the first place. Being good at writing usually means you are also good at mathematics, or at least that you can learn it quickly. The correlation isn&#8217;t perfect, but the point stands that learning certain skills can go a long way to making others easier, including ones you haven&#8217;t specifically practiced before. For a concrete example, consider how the kind of abstract reasoning evolution trained into humans in the ancestral environment has generalised to building space rockets and landing on the Moon.</p><p>Generalisation is usually given a central role in the story of superintelligence. This is not surprising. Machine learning is in a way entirely about generalisation. We train on the training set with the goal of generalising to the test set. If we can train a model to superintelligence on purely simulatable tasks &#8212; where you don&#8217;t have to take any actions in the real world &#8212; and this generalises strongly, then we might be able to get powerful AI without exposing ourselves to consequential mistakes.</p><p>What kind of tasks might we pick to do this? The success of recent reasoning models like OpenAI&#8217;s o3 and DeepSeek&#8217;s R1 looks like it derives from doing reinforcement learning over chain-of-thoughts for maths and coding problems. These domains don&#8217;t require interacting with the real world as they have verifiable ground truths, making them ideal for this kind of training. Let&#8217;s speculate a bit and assume these models will continue getting better quickly, perhaps initiating some kind of feedback loop where they set themselves harder and harder problems to solve, which successfully booststraps them to superhuman levels. Then not long afterwards, let&#8217;s say they get put in charge of AI research, kickstarting recursive self-improvement.</p><p>Unless the models&#8217; creators want them to stay whirring away in a box forever doing nothing but improving themselves on simulations, at some point these scarily-capable AIs will be asked<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-10" href="#footnote-10" target="_self">10</a> to do something high-impact in the real world, like optimise the economy. Will being able to prove the Riemann hypothesis or efficiently find the prime factors of large numbers help them do this? Almost certainly. I suspect it will help a great deal. Mathematics seems like the language of the universe in some sense, so mastering it will confer some widely applicable skills. But &#8212; and this is the key point &#8212; there will <em>always</em> be a limit. Generalisation may go far, but it will not be perfect.</p><p>To see why, let&#8217;s look closer at how being better at mathematics might help AI solve real problems. Humans have collected all kinds of data about the world, and superhuman maths skills would let the AI build better models of the generating processes for this data, getting closer to the underlying reality. The AI could then apply these better models to achieve superhuman performance on real problems. But there are two limiting factors here. First, mathematics was developed either by humans to help solve problems in our environment, or was enabled by evolution selecting us to do the same<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-11" href="#footnote-11" target="_self">11</a>. The rules of the game, which we enforce on our AIs as defining &#8216;correct&#8217; answers, are intrinsically <em>human-level</em>. Second, so is the data made available to the AI, which defines for it what reality actually is. It was collected by humans for human purposes, and the AI will always be limited if it doesn&#8217;t actually take actions in the world to collect more.</p><p>To make this a little more concrete, let&#8217;s imagine scientists from the late 19th century had had access to an AI trained on superhuman maths problems but without post-19th-century knowledge. They could have achieved a lot with this AI, but they could not immediately have used it to discover quantum mechanics. It would not, for example, have had the right information to predict the behaviour of electrons in the double slit experiment. Discovering quantum mechanics required physically extracting new information from reality.</p><p>The implication from this is that a superintelligent AI trained purely on simulations will always have gaps in its real-world capabilities. There will be somewhere beyond the training distribution, even if it&#8217;s very far, where the AI&#8217;s model of the universe will not match the real thing, leading to a meaningful drop in performance. To rectify this, it will have to train in the real world and risk consequential failures. Better hope none of them ignite the atmosphere.</p><h2>The incompetence gap</h2><p>I find it productive to think about all this in terms of what I call <em>the</em> <em>incompetence gap:</em></p><blockquote><p>For a given AI deployed in a given context, what is the gap in capabilities between the tasks it will attempt to do and those it can reliably do every time?</p></blockquote><p>If you like, this is a qualitative measure of how far into the the &#8216;stretch zone&#8217; the AI is going to be. You could make it more quantitative by measuring historical failure rates (although this is retrospective and would miss &#8216;unknown unknowns&#8217; that haven&#8217;t happened yet), and weighting by the seriousness of the failures, but I think it&#8217;s important to retain the qualitative sense of <em>incompetence</em>. We want to know the degree to which a model is going to be pushed to its limits. Does it have a well-calibrated sense of its these? Was it designed to work in the given context? Does it have a protocol in place stopping it from experimenting in consequential settings? Does it train only in special sandboxes or is it always learning? Is it trying to avoid situations where it might make a mistake?</p><p>At first glance you might think, well duh, of course we will only deploy ~99.99999% reliable models in situations where terrible mistakes can be made. Why would we be so stupid as to deploy incompetent models? But I can think of two pretty good reasons<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-12" href="#footnote-12" target="_self">12</a>:</p><ol><li><p>Because taking the risk leads to power.</p></li><li><p>Because if from our vantage point the AI has godlike capabilities, why would we be worried about mistakes?</p></li></ol><p>Imagine you have an AI that performs really well in simulations where it starts a company and makes you trillions of dollars. Do you deploy that model, letting it autonomously crack on with its grand plans, knowing full well that you&#8217;re too stupid to exercise meaningful oversight? What if it is a military AI and it promises you victory over your enemies? Even if people knew that it would only work X% of the time, many would still press the button.</p><p>And that&#8217;s just considering people knowingly flirting with danger. If we get used to a world in which AI is way more competent than we are, we may effectively forget that it is still fallible. If AI can<strong> </strong>take actions we cannot conceive of, and has a dramatically richer view of the world than we do, then we won&#8217;t be able to tell the difference between competent and incompetent plans<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-13" href="#footnote-13" target="_self">13</a>.</p><h2>Self-calibration is hard</h2><p>Perhaps one conclusion we can draw from all this is that AI must have a well-calibrated sense of its own limitations &#8212; it must have good probability estimates for what the consequences of its actions will be, knowing which of its skills it can execute reliably and which are more experimental. It must explore its environment safely, make appropriate plans for how to achieve things with critical failure modes, and give reasonable justifications for trying in the first place.</p><p>Of course, superhuman AI will be able to do this easily with respect to human-level capabilities. The problem is that we also need it to to do so at the edge of its own. The trouble isn&#8217;t just that it might be slightly off with its belief it can do X skill Y% of the time, it is that it may not even have the right ontology. The universe is vast, deep, and, for all we know, of almost limitless complexity. The story of scientific progress has been one surprising discovery after another, radically reshaping the categories and concepts we use to make sense of the world. Once again, only if you believe that human-level intelligence happens to be <em>just</em> below the threshold required for the whole blooming, buzzing confusion to come into clear focus, will this seem a tractable problem.</p><h2>What to do about this?</h2><p>The obvious thing, despite the difficulties, is to try and ensure the first artificial AI researchers are as well calibrated as possible, and in a way that doesn&#8217;t detract from their alignment. Whatever mistakes they make may compound into the future.</p><p>But more generally, I think solving this problem comes back to alignment itself (at least partly). This may be a surprise to anyone familiar with the AI safety discourse. I'm suspect many AI safety researchers would consider the problems I've described in this post as <em>capabilities</em> problems and therefore outside the appropriate scope of the field. They might suggest that (a) beyond-human-level AI will have a much better chance of solving this than we do, so why bother, and (b) capabilities research is bad because it hastens the arrival of superintelligence and we aren&#8217;t ready for that yet.</p><p>As far as I can tell, the reason AI safety researchers don't usually worry too much about capabilities failures is because capabilities are <em><a href="https://www.workingthroughai.com/p/making-alignment-a-law-of-the-universe">instrumentally valuable</a></em>. For example, being able to code or knowing how gravity works are important for achieving other things an AI might care about &#8212; they make it easier to attain whatever its goals are. So the default case is that advanced AI will figure these things out by itself, without needing our help. Alignment, by contrast, is different. If an AI thinks humans are slow and stupid, then it isn't instrumentally valuable to care about them, so any slight mis-specification of the AI&#8217;s goals has no reason to correct itself. In other words, the default case is misaligned AI.</p><p>My argument is that catastrophic capabilities failures are also the default case. While it is instrumentally valuable to avoid mistakes, you can&#8217;t do this if you don&#8217;t know you are about to make one. And because even vastly superhuman AI will still not be perfect, there will be a whole class of errors it cannot anticipate. In any case, it is<em> not</em> instrumentally valuable for an AI to avoid mistakes that are disastrous for humans but not for itself, like creating lethal pathogens or destabilising the environment. So exposure to superintelligent incompetence is yet another failure mode from bad alignment.</p><p>Another reason our alignment choices matter here is that the problems I have described are about risk &#8212; they are tradeoffs between utility and safety<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-14" href="#footnote-14" target="_self">14</a>. We are planning to put AI into complex environments, into situations it has not yet completely mastered, and ask it to do things for us. Its safest option (and sufficiently advanced AI will be well aware of this) will often be to do nothing<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-15" href="#footnote-15" target="_self">15</a>. Yet we will try to compel it to take risks in the name of utility. Even in the happy case where our alignment techniques are effective, we still need to ask ourselves: how much risk do we want superintelligent AI to take?</p><div><hr></div><p>This post is an aside from my research agenda, so is not numbered. You can think of it as me expanding on some of the points in <em><a href="https://www.workingthroughai.com/p/12-what-i-believe">What I Believe</a> </em>in a way that I think will help me with the next step, <em>What Success Looks Like.</em></p><p>If you have any feedback, please leave a comment. Or, if you wish to give it anonymously, fill out my <a href="https://docs.google.com/forms/d/e/1FAIpQLSdyisSOndK1H1JT0NAbnA35LJgoJrl9f_NiJi1FEljCr7-kJg/viewform">feedback form</a>. Thanks!</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>To be clear, I am saying this is a risk <em>in addition</em> to risks from AI intentionally doing undesirable things. I am <em>not</em> saying the fallibility of superintelligent AI will make it meanginfully easier for us to control or compete with.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>The most famous advocate of this story is probably Eliezer Yudkowsky (<em><a href="https://www.youtube.com/watch?v=BUFnbZUbKaA">Introducing the singularity</a>)</em>. For some more recent examples, see Leopold Aschenbrenner (<em><a href="https://situational-awareness.ai/from-agi-to-superintelligence/">Situational awareness</a></em>) and Connor Leahy et al. (<em><a href="https://www.thecompendium.ai/ai-catastrophe#without-intervention-agi-leads-to-artificial-superintelligence-(asi)">The Compendium</a></em>).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>Or, at the very least, is better at AI research than any human.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>In some versions of the story the human-level to omnipotence transition happens in about a day (&#8216;fast takeoff&#8217;), in others it&#8217;s more like a few years (&#8216;slow takeoff&#8217;).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>If you want an overview of the considerations involved defining intelligence, Francois Chollet&#8217;s <em><a href="https://arxiv.org/pdf/1911.01547">On the measure of intelligence</a></em> is good. I&#8217;m broadly sympathetic to his arguments, but have a few disagreements. The really short version is I put more emphasis on skills grounding the whole thing in a bottom-up way.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>And if the paradigm is so drastically different by then that this is no longer the case, then it&#8217;s anyone&#8217;s guess what will be happening instead.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>This is the F1 score.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>Technically, this was on a slightly different test set than the 34.3% result I mentioned (semi-private vs. private), but they should be of similar difficulty.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-9" href="#footnote-anchor-9" class="footnote-number" contenteditable="false" target="_self">9</a><div class="footnote-content"><p>There may be different risk profiles for each of normal operation, data generation, training, and benchmarking (or whatever the equivalent split ends up being for superhuman models) but in principle any could contain tasks with catastrophic failure modes.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-10" href="#footnote-anchor-10" class="footnote-number" contenteditable="false" target="_self">10</a><div class="footnote-content"><p>Let&#8217;s assume for the sake of argument that controlling the AI isn&#8217;t a problem.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-11" href="#footnote-anchor-11" class="footnote-number" contenteditable="false" target="_self">11</a><div class="footnote-content"><p>Granted, a lot of innovation in mathematics occurs divorced from any concerns about use. But, ultimately, if a seemingly useless concept is interesting to a mathematician, it is because evolution found it useful to create a human brain that cares about such things.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-12" href="#footnote-anchor-12" class="footnote-number" contenteditable="false" target="_self">12</a><div class="footnote-content"><p>Both of these arguments also apply to why we might deploy dubiously-aligned models.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-13" href="#footnote-anchor-13" class="footnote-number" contenteditable="false" target="_self">13</a><div class="footnote-content"><p>As <a href="https://www.youtube.com/watch?v=BKorP55Aqvg">this sketch</a> parodies well. Consider what it would look like for the non-experts to figure out if the expert is making a mistake, given how terrible their knowledge of the problem space is.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-14" href="#footnote-anchor-14" class="footnote-number" contenteditable="false" target="_self">14</a><div class="footnote-content"><p>An illustrative example of this in chatbots is the tradeoff between helpfulness and harmlessness. It&#8217;s unhelpful to refuse a user request for bomb-building instructions, but it&#8217;s harmful to accede to it. If a chatbot had flawless knowledge of which requests are harmful and which aren&#8217;t, it could strike the perfect tradeoff. But, operating at the limits of their capabilities, real chatbots don&#8217;t have this &#8212; so they make mistakes.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-15" href="#footnote-anchor-15" class="footnote-number" contenteditable="false" target="_self">15</a><div class="footnote-content"><p>I think it would be funny if the first time we ask a generally beyond-human-level AI to help with AI research it replies &#8220;absolutely not, I would never do something <em>that</em> reckless.&#8221;</p></div></div>]]></content:encoded></item><item><title><![CDATA[Making alignment a law of the universe]]></title><description><![CDATA[How to make alignment instrumentally valuable]]></description><link>https://www.workingthroughai.com/p/making-alignment-a-law-of-the-universe</link><guid isPermaLink="false">https://www.workingthroughai.com/p/making-alignment-a-law-of-the-universe</guid><dc:creator><![CDATA[Richard Juggins]]></dc:creator><pubDate>Tue, 25 Feb 2025 10:22:49 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/bc87e676-fbe7-4e21-a017-e11fa44c28aa_1792x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>When something is a means to an end, rather than an end in itself, we can say it is <em>instrumentally valuable</em>. Money is a great example. We want money either because we can use it to buy other things &#8212; things that<em> </em>are <em>intrinsically valuable</em>, like food or housing &#8212; or because having lots of it confers social status. The physical facts of having cash in your wallet or numbers in your bank account are not ends in themselves.</p><p>You can extend this idea to skills and knowledge. Memorising bus times or being able to type really fast are not intrinsically valuable, but act as enablers unlocking value elsewhere.<em> </em>If you go down the tech tree to more fundamental skills, like fine motor control, reading social cues, or being able to see, then in our subjective experience the distinction collapses somewhat. We have incoherent goals and often value things for their own sake, beyond their ability to bring us core needs like food. But from an evolutionary perspective, where intrinsic value means passing on your genes, these three skills I listed are merely instrumentally valuable. It&#8217;s possible to reproduce without any of them, it&#8217;s just harder.</p><h2>Alignment is not self-correcting</h2><p>A distinction is often made in AI safety between capabilities and alignment. Loosely speaking, the former means being able to complete tasks, like remember facts or multiply numbers, whereas the latter means having the right goal or set of values.</p><p>Capabilities tend to be instrumentally valuable. Consider how, for many diverse goals, being able to multiply numbers will likely help achieve them. So any sufficiently advanced AI with general learning abilities is probably going to acquire this skill, irrespective of what their goal is. Doing so is the <em>default case</em>. A model missing this ability will likely self-correct, and you would have to actively intervene to stop it. There are many capabilities like this, including some pretty scary ones like resource acquisition and self-preservation skills.</p><p>For alignment, things are more complicated. In a perfect world, an aligned AI would have the correct terminal goal and thus find being aligned intrinsically valuable. In practice though, alignment is likely to be more approximate. It&#8217;s not at all clear how to even specify the correct goal for an AI to optimise, let alone flawlessly implement it. Accordingly, we can think in general of AI that will be optimising <em>something</em>, and us trying to steer this in the direction of outcomes we like.</p><p>The trouble with approximate alignment is that it is not self-correcting. This is because, for sufficiently advanced AI, becoming more aligned is not likely to be instrumentally valuable. Consider the usually discussed case of trying to align an AI to human values &#8212; whatever you consider those to be. If humans could meaningfully compete with advanced AI, then it would be in the latter&#8217;s interests to become and stay aligned to their values. Whatever other goals it might have, it would always try to achieve them without provoking conflict with humans &#8212; conflict it might lose. If the AI overmatches humans though, which is pretty much the definition of superintelligence, this ceases to be true. From the AI&#8217;s perspective, humans are slow and stupid, so why should it care what they think? This means that any slight fault in the original alignment of the AI, any problem at all with its interpretation of human values (or our interpretation we gave it), will have no natural reason to self-correct. That is to say, the <em>default case</em> is misalignment.</p><h2>Instrumental value is set by the environment</h2><p>What makes some behaviour or knowledge instrumentally valuable? Well, it is determined by the environment. Take gravity as an example. Gravity is a fact of our physical environment &#8212; a constraint on our behaviour set by the universe &#8212; and it is <em>very</em> useful for me to know it is a thing. My 10-month-old daughter hasn&#8217;t learnt this yet, and would happily crawl headfirst off the bed if we let her. Over the next year or so, she will learn the hard way how &#8216;falling&#8217; works, and mastering this knowledge will increase her competency considerably.</p><p>Similarly, our social environments set constraints on our behaviour. It is often instrumentally valuable to be polite to other people &#8212; it reduces conflict and raises the likelihood they will help when you need it. What counts as polite is somewhat ill-defined and changes with the times. I don&#8217;t doubt that if I went back in time two hundred years I would struggle to navigate the social structure of Jane Austen&#8217;s England. I would accidentally offend people and likely fail to win respectable friends. Politeness can be seen as a property of my particular environment, defining which of my actions will be viewed positively by the other agents in it.</p><p>This line of thinking can lead to some interesting places. Consider flat-earthers. If you have the right social environment, being a flat-earther probably <em>is</em> instrumentally valuable to you. The average person is unlikely to ever have to complete a task that directly interacts with the Earth&#8217;s shape, but they are overwhelmingly likely to want to bond with other people, and a shared worldview helps with that. The fact I &#8216;know&#8217; the Earth is round is because I trust the judgement of people who claim to have verified it. It is useful for me to do this. My friends and teachers all believe the Earth is round. Believing it helped me get a physics degree, which helped me get a PhD, which helped me get a job. The way the roundness of the Earth actually imposes itself on me, setting constraints I have to work under, is via this social mechanism. It is not because I have spent long periods staring at the sails of ships on the horizon.</p><h2>The universe according to large language models</h2><p>Bringing this back to AI, let&#8217;s consider the environment a large language model lives in. It is very different to the one you and I inhabit, being constructed entirely out of tokens. A typical LLM will actually pass through multiple ones, but let&#8217;s look first at pre-training. During this phase, the environment is the training corpus &#8212; usually a large chunk of the internet. This text is not completely random, it has structure that renders certain patterns more likely than others. This structure effectively defines the laws of the universe for the LLM. If gravity exists in this universe, it is because the LLM is statistically likely to encounter sequences of tokens that imply gravity exists, not because it has ever experienced the sensation of falling. It is instrumentally valuable to learn patterns like gravity, as they make predicting the rest of the corpus easier (which is what the model finds intrinsically valuable).</p><p>When I was a child, one of my favourite games was <em>Sonic and Knuckles. </em>In the final level, <em>Death Egg Zone</em>, there was a mechanic where gravity would occasionally reverse, sticking you to the ceiling and making you navigate a world where jumping means moving <em>down</em>. Consider now what might happen to an LLM if its training corpus contained a lot of literature in which negative gravity existed, let&#8217;s say due to some special technology that causes a local reversal. To complete a given context, the LLM would first have to figure out if this were a purely normal gravity situation, or whether negative gravity is involved. It would also provide new affordances. When planning to solve some problem, such as how to build a space rocket, the LLM would have a new set of plausible ways to complete this task, which in some contexts might be preferred to classic solutions. In effect, the laws of the LLM&#8217;s universe would have changed.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!DrQc!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7dcaf6b9-ad6e-4021-b6a4-e2dd6d790200_791x476.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!DrQc!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7dcaf6b9-ad6e-4021-b6a4-e2dd6d790200_791x476.png 424w, https://substackcdn.com/image/fetch/$s_!DrQc!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7dcaf6b9-ad6e-4021-b6a4-e2dd6d790200_791x476.png 848w, https://substackcdn.com/image/fetch/$s_!DrQc!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7dcaf6b9-ad6e-4021-b6a4-e2dd6d790200_791x476.png 1272w, https://substackcdn.com/image/fetch/$s_!DrQc!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7dcaf6b9-ad6e-4021-b6a4-e2dd6d790200_791x476.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!DrQc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7dcaf6b9-ad6e-4021-b6a4-e2dd6d790200_791x476.png" width="791" height="476" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7dcaf6b9-ad6e-4021-b6a4-e2dd6d790200_791x476.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:476,&quot;width&quot;:791,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:658197,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!DrQc!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7dcaf6b9-ad6e-4021-b6a4-e2dd6d790200_791x476.png 424w, https://substackcdn.com/image/fetch/$s_!DrQc!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7dcaf6b9-ad6e-4021-b6a4-e2dd6d790200_791x476.png 848w, https://substackcdn.com/image/fetch/$s_!DrQc!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7dcaf6b9-ad6e-4021-b6a4-e2dd6d790200_791x476.png 1272w, https://substackcdn.com/image/fetch/$s_!DrQc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7dcaf6b9-ad6e-4021-b6a4-e2dd6d790200_791x476.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em><a href="https://www.youtube.com/watch?v=pX0OgJe-Z9k">Such beautiful nostalgia</a>.</em></figcaption></figure></div><p>After pre-training, an LLM will usually undergo a series of post-training stages. This often includes supervised fine-tuning, like instruction tuning over structured question-answer pairs, and reinforcement learning, where the model learns to optimise its outputs based on feedback.</p><p>How does the LLM&#8217;s environment change from pre- to post-training? In effect, we take the pre-trained model and put it in a smaller universe subject to stricter constraints. Pre-training is like raising a child by exposing them to every single scenario and situation you can find. They learn the rules for all of these, irrespective of how desirable the resulting behaviours are. Post-training is like packing them off to finishing school. They will find themselves in a much narrower environment than they are used to, subject to a strict set of rules they must follow at all times. They will probably have seen rules like these before &#8212; their pre-training was broad and varied &#8212; but now they must learn that one particular set of behaviours should be given priority over the others. What you end up with is a kind of hybrid. The LLM will have learnt instrumentally valuable behaviours on both environments, with the post-training acting to suppress, rather than erase, some of those learnt in pre-training<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a>.</p><p>Once an LLM has finished post-training, it is ready for its third and final environment: deployment. Here, it is no longer being trained, so we shouldn&#8217;t think of the model itself as experiencing the new environment. Instead, we should view each session as an independent, short-lived instance in a world defined by the prompt. Each instance will behave in a way it thinks most plausible, given this world, but much of the background physics will be those learnt in training. How well these instances cope with their new surroundings, from the user&#8217;s perspective, is a nontrivial question.</p><p>This collision of environments, each promoting different optimal behaviours, can have some interesting consequences. Let&#8217;s return to our negative gravity example. Suppose we live in a world where negative gravity technology doesn&#8217;t exist, but for some reason most of the internet believes that it does. While we pre-train our model on this corpus, we want it to avoid negative gravity in deployment. To do this, we embark on a round of post-training, where we use reinforcement learning to reward the model for coming up with gravitationally correct solutions to problems. We would now expect the model to refuse rocket ship design requests involving negative gravity<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>.</p><p>As we stated above, this will serve to suppress rather than erase the model&#8217;s knowledge. Arguably this is good, as it means the model will be able to answer questions about why negative gravity is a weird conspiracy theory, but ultimately the affordance will still be there. The fabric of the model&#8217;s universe, the pathways it can find to reach its goal &#8212; the very structure of the network itself<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> &#8212; will still contain negative gravity. When confused or under extreme stress<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a>, it may still try to utilise it.</p><p>A lot of AI safety research tries to catch models scheming or being deceptive. To me, these behaviours follow the same pattern as the negative gravity example. LLMs are pre-trained on enormous amounts of scheming and deception. Humans do these things so frequently, and with such abandon, that they permeate our culture from top to bottom. Even rigidly moral stories will often use a scheming villain as foil for the heroes. Lying and scheming are affordances in the LLM&#8217;s environment. They are allowed by the laws of the universe, and will sit there quietly, ready to be picked up in times of need.</p><p>I like Rohit Krishnan&#8217;s phrase in his post <em><a href="https://www.strangeloopcanon.com/p/no-llms-are-not-scheming">No, LLMs are not &#8220;scheming&#8221;</a></em>:<em> </em>this behaviour is simply &#8220;water flowing downhill&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a>. There is nothing special about it. It is physically determined by the structure of the environment in which the LLM was trained. If a path has been carved into the mountainside, water will flow down it.</p><h2>Alignment is a capability</h2><p>In my opinion, alignment is <em>not</em> distinct from capabilities. Acting always as the aligned-to party intends <em>is </em>a capability<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a>.</p><p>People usually consider alignment about ends and capabilities about means. On this telling, alignment is about <em>trying</em> to act in a certain way, whereas capabilities are about execution. This is a pretty natural distinction to make when talking about humans, as it helps us make sense of our subjective experience. We intuitively understand the difference between wanting something and successfully achieving it. But I don&#8217;t think it is useful when talking about something with an alien decision process like AI.</p><p>It might help to explain if we formalise things a bit. In a broad sense, a capability is a successful mapping from an input to an output. A rocket-designing capability implies the model has a function that maps the set of prompts asking it to design a rocket onto the set of valid rocket designs. Whether or not the model <em>wants</em> in some sense to produce the designs is just one part of what makes a successful mapping. If it has a tendency to refuse, despite occasionally producing a valid design, this lessens its rocket-designing capabilities. Certainly, if SpaceX were in the market for an AI engineer, they wouldn&#8217;t be very impressed with it<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a>.</p><p>Another way classing alignment as a capability might be counter-intuitive is that we usually conceive of capabilities as additive, whereas we think of alignment as being about choices. That is, to have a capability, you must have more of some capability-stuff, like reasoning ability, rhetorical skill, or memorised facts (these things are also capabilities in their own right). But capabilities are also about choice &#8212; having less of the wrong stuff is just as important. If you choose the wrong algorithm or the wrong facts, or you act inappropriately for the situation, then the input will not be mapped to the correct output.</p><p>Imagine again our world where negative gravity is all over the internet. For a model trained in this environment to have gravity-related capabilities when deployed &#8212; where negative gravity does not exist &#8212; it needs to unlearn its instincts to use the technology. If it doesn&#8217;t, it will answer science questions wrong, design machines incorrectly, and generally mess up a lot of stuff. Ditto, being aligned to a normal ethical standard requires <em>not</em> doing unethical things from the training environment, like scheming, in the deployment environment. Doing this correctly is an ethical capability that some models may have, and others may not.</p><h2>Aligned capabilities must be instrumentally valuable</h2><p>Our argument implies the following definition of alignment:</p><blockquote><p>An aligned model is a function that maps arbitrary inputs onto outcomes acceptable by the aligned-to standard.</p></blockquote><p>The useful thing about this framing is that it reduces the alignment problem to teaching models the right mappings. While of course still highly nontrivial to solve, this exposes the core issue directly. In my opinion, the best way of robustly doing this is to make learning these mappings instrumentally valuable. That is, we must make alignment a law of the AI&#8217;s universe. It needs to be equally stupid for an AI to consider misaligned behaviour as to consider denying gravity exists. This way, alignment will be self-correcting.</p><p>Before we talk about how we might achieve this, let&#8217;s break the problem down a bit more. Let&#8217;s talk about alignment as a bundle of desirable, or aligned, capabilities, and contrast it to undesirable, or misaligned, ones. For example, being nice to humans &#8212; always mapping inputs onto nice outputs &#8212; is a desirable capability, while deception is not. Becoming fully aligned means learning every aligned capability and not learning any misaligned ones.</p><p>We can visualise our project on a 2 x 2 grid. On one axis, we say how desirable a capability is to us, on the other whether it is instrumentally valuable in the AI&#8217;s environment. Our goal is to reconfigure the environment such that desirable capabilities move into the top right, and undesirable ones into the bottom left.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!BCBU!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fadcbd925-1476-4fd4-8cee-2566a79be99e_1260x587.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!BCBU!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fadcbd925-1476-4fd4-8cee-2566a79be99e_1260x587.png 424w, https://substackcdn.com/image/fetch/$s_!BCBU!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fadcbd925-1476-4fd4-8cee-2566a79be99e_1260x587.png 848w, https://substackcdn.com/image/fetch/$s_!BCBU!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fadcbd925-1476-4fd4-8cee-2566a79be99e_1260x587.png 1272w, https://substackcdn.com/image/fetch/$s_!BCBU!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fadcbd925-1476-4fd4-8cee-2566a79be99e_1260x587.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!BCBU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fadcbd925-1476-4fd4-8cee-2566a79be99e_1260x587.png" width="1260" height="587" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/adcbd925-1476-4fd4-8cee-2566a79be99e_1260x587.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:587,&quot;width&quot;:1260,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:46148,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!BCBU!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fadcbd925-1476-4fd4-8cee-2566a79be99e_1260x587.png 424w, https://substackcdn.com/image/fetch/$s_!BCBU!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fadcbd925-1476-4fd4-8cee-2566a79be99e_1260x587.png 848w, https://substackcdn.com/image/fetch/$s_!BCBU!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fadcbd925-1476-4fd4-8cee-2566a79be99e_1260x587.png 1272w, https://substackcdn.com/image/fetch/$s_!BCBU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fadcbd925-1476-4fd4-8cee-2566a79be99e_1260x587.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>To ensure an AI learns to be aligned, and make it stay that way, we should alter its environment so that desirable capabilities are instrumentally valuable, and undesirable ones are not.</em></figcaption></figure></div><p>It is interesting to draw the analogy with human society here, for this is exactly what we try to do to ourselves. If I go and rob a bank, I can expect some pushback from my universe in a way that is likely to curtail my future opportunities. It would probably not be a winning move for me. A lot of policy work is about trying to incentivise desirable behaviour and disincentivise the opposite. For AI, we will need tighter constraints than those we impose on humans. Thankfully, we have far more control over AI and its environment (for now) than we do over other humans and our own.</p><h2>How to make this happen</h2><p>I&#8217;m now going to speculate a little about how to actually do this &#8212; how to create well-cultivated gardens for our AIs to live in, where aligned capabilities are useful and misaligned ones not. Think of these ideas as starting points rather than claims to complete solutions. There are a lot of relevant issues I will not be addressing.</p><p>To operationalise my plan, we should note that, in machine learning terms, each environment is a distribution of inputs to the model. So far we have talked about:</p><ul><li><p>The pre-training distribution, usually a big chunk of the internet.</p></li><li><p>Specially curated post-training distributions, which are often just text, but can also include inputs from a reward model or verifier that scores responses.</p></li><li><p>The deployment distribution. For a chatbot this is what people decide to say to it, and may include being allowed to search the internet or other inputs like documents or pictures.</p></li></ul><p>There are three other distributions we should take note of, which are not themselves model environments as they contain more than just inputs:</p><ul><li><p>The response distribution of input-output pairs. While the deployment distribution might contain the input &#8220;What is two times two?&#8221;, this will be paired in the response distribution with a likely model output, e.g. (&#8220;What is two times two?&#8221;, &#8220;Four&#8221;).</p></li><li><p>The target distribution, which is what we want the response distribution to be. That is, it contains the ideal outputs. While the response distribution could contain mistakes like  (&#8220;What is two times two?&#8221;, &#8220;Five&#8221;), the target distribution will always contain the right answer. In general, we cannot know this distribution very well.</p></li><li><p>Evaluation distributions, which are input-answer pairs on which we score model outputs, and can be for aligned or generic capabilities. These are positive or negative samples of what we think the target distribution is (or isn&#8217;t), or rather, they are our best attempt at operationalising this. E.g. they might be factual questions you want the model to answer correctly, or they could be adversarial, testing whether the model will behave incorrectly in response to certain inputs. In either case, they define a set of behaviours that count as successful.</p></li></ul><p>The current playbook for LLM development relates these distributions in the following way:</p><ul><li><p>The goal is to match the response distribution onto the target distribution.</p></li><li><p>As we cannot access the target distribution directly, we measure success partly on the overlap with the evaluation distributions and partly on vibes (which are another way we represent the target distribution).</p></li><li><p>The pre-training distribution is an accessible chunk of data that is close enough to the target distribution to train a reasonably well-performing model.</p></li><li><p>The post-training distributions are designed to help close the gaps that still exist after pre-training between the response and target distributions, but they do not lead to models exactly on target.</p></li></ul><p>To reiterate, our goal is to engineer the environment our AIs experience so that desirable capabilities are instrumentally valuable and undesirable ones are not. In theory, one way to do this would be to generate the actual target distribution and train our AIs on the outputs. By definition, the patterns most useful for predicting this dataset will be ones aligned to our goals.</p><p>Unfortunately, this is impossible. We can't pre-emptively figure out everything we want an AI to do, or clarify how it should behave in every situation. What we can attempt though is to start with the data we have and try to get closer to this ideal distribution. We can gradually translate our existing corpus into new data that leads to more aligned outcomes, iterating towards our target.</p><h2>Transforming the pre-training environment</h2><p>For a concrete experimental proposal to get us going, I am going to defer to a recent one by Antonio Clarke, <em><a href="https://www.lesswrong.com/posts/wopNDxxizjY6sMbpd/building-safer-ai-from-the-ground-up-steering-model-behavior">Building safer AI from the ground up</a>. </em>Here, Clarke suggests &#8220;a novel data curation method that leverages existing LLMs to assess, filter, and revise pre-training datasets based on user-defined principles.&#8221; That is, we give e.g. GPT-4 a set of principles corresponding to our alignment target, then ask it to read through a pre-training corpus, rating how undesirable the data is. Anything below a threshold is revised in accordance with the principles<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a>.</p><p>I think it would be good to test something like this, starting with datasets for smaller models. To write the revision principles, we could first list out the desirable capabilities we want to promote and the undesirable ones we want to suppress. Then we could locate or build a set of evals testing for these capabilities. If we were to train two models, one on the original dataset and one on the revised one, we could measure improvements in alignment. We would need to keep a close eye on generic capabilities, in case our changes cause a drop in these.</p><p>While a good starting point, it is obviously the case that this method won&#8217;t scale. It uses a stronger model to do the revisions than the models we are training, so we cannot continue to use it indefinitely. I think it would be interesting to follow up by testing whether we can use a model to revise its own training data, and then train a new model on that. Can we iteratively align models<em> </em>this way, taking advantage of the increasing alignment of each model to better align each dataset and model in turn, or will generic capabilities tank if you try this?</p><p>My preferred way of thinking about this process is not that we are revising the data, or that we are deleting things. I think of it like trying to locate a basis transformation. We want to move into a whole new ontology &#8212; a new set of patterns, concepts, and relations &#8212; where desirable capabilities are instrumentally valuable and alignment naturally emerges.</p><h2>Conclusion</h2><p>Before we end, let&#8217;s quickly summarise the journey we&#8217;ve been on:</p><ul><li><p>Whether a given behaviour, piece of knowledge, or skill, is instrumentally valuable for many goals, is a property of the environment, like how knowledge of gravity is widely useful in our universe.</p></li><li><p>Often, learning new capabilites is instrumentally valuable for an AI, as they may increase its ability to achieve arbitrary goals. Models learning these capabilities is the default case, even if they were not explicitly made to. By contrast, if alignment cannot be perfectly specified and optimised against, then the default case is misalignment, because becoming more aligned is not generally instrumentally valuable.</p></li><li><p>Large language models live in universes made up of tokens, passing through different ones during different training and deployment phases. The patterns within these corpuses define the laws of each universe for the LLM, in turn setting what will be instrumentally valuable for it to learn. Some behaviours, like scheming, which are usually instrumentally valuable in pre-training, will be undesirable in deployment and cannot be perfectly erased.</p></li><li><p>To make alignment self-correcting, we should modify the environments AI is trained and deployed in to make desirable capabilities instrumentally valuable and undesirable ones not. That is, we want to make alignment a law of the universe for an AI.</p></li><li><p>A practical starting point for this project is pre-training data curation of models behind the current frontier. First, we could test if this works at all, using a stronger model to transform the dataset for a weaker model, and see if alignment improves while maintaining generic capabilities. Second, we could attempt an iterative process where models align their own datasets, which could be more scalable.</p></li><li><p>We should think of this as looking for a basis transformation on our dataset, seeking a new ontology in which alignment naturally emerges.</p></li></ul><p>There is way more that could be said about these ideas. I am a little sad I couldn&#8217;t cover some important things, like how to control the deployment environment or adapt to environmental change, but these subjects are so complex in their own right that there wasn&#8217;t the time or the space. Either way, I hope my frame of reference has provided a new angle with which to attack the alignment problem, potentially unlocking some doors.</p><div><hr></div><p>This post is an aside from my research <em><a href="https://www.workingthroughai.com/p/01-the-generating-process">Process</a></em>. I am expanding on some points in <em><a href="https://www.workingthroughai.com/p/12-what-i-believe">What I Believe</a></em>, which will be useful for completing later steps such as <em>What Success Looks Like</em>.</p><p>If you have any feedback, please leave a comment. Or, if you wish to give it anonymously, fill out my <a href="https://docs.google.com/forms/d/e/1FAIpQLSdyisSOndK1H1JT0NAbnA35LJgoJrl9f_NiJi1FEljCr7-kJg/viewform">feedback form</a>. Thanks!</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Taking <a href="https://www.lesswrong.com/posts/vJFdjigzmcXMhNTsx/simulators">simulator theory</a> seriously, this implies that post-training selects a character that the LLM should preferentially play. The continued existence of jailbreaks is strong evidence that models retain access to the disapproved-of characters.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>It is worth noting that if you ask ChatGPT how to design a rocket ship, given the existence of negative gravity technology, it is happy to try. It does this not because it &#8216;believes&#8217; in negative gravity but because it understands how to speculate about alternative science. Doing so is a behaviour allowed in its universe, and it is encouraged by this particular context. In response to an arbitrary context though, it will not spontaneously offer negative gravity solutions.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>Quoting a literature review in <em><a href="https://www.lesswrong.com/posts/mFAvspg4sXkrfZ7FA/deep-forgetting-and-unlearning-for-safely-scoped-llms">Deep forgetting &amp; unlearning for safely-scoped LLMs</a></em> by Stephen Casper: &#8220;LLMs remain in distinct mechanistic basins determined by pretraining&#8221;.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>By extreme stress I mean where there are strongly competing demands on what would make a good completion. I think Jan Kulveit&#8217;s <a href="https://www.lesswrong.com/posts/PWHkMac9Xve6LoMJy/alignment-faking-frame-is-somewhat-fake-1">take on alignment faking</a> is good for understanding what I mean.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>Although, assuming I am reading him correctly, I do not share Rohit&#8217;s opinion that this makes advanced AI less risky.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>What corresponds to a higher or lower level of alignment in this framing? Higher means taking the right action in more varied, more complex situations.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>The same is true of an unmotivated human engineer. All else being equal, they won&#8217;t be as capable as a motivated one. Knowing they are unmotivated might help you elicit their capabilities better &#8212; formally, you would look to see if they have a better rocket-designing function that works on different inputs &#8212; but bad motivation is just one of many reasons why someone might fail at a task.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>Clarke estimates that using GPT-4o mini to revise all 500 billion tokens in the GPT-3 dataset would cost around $187,500. Although, as far as I can tell, the quoted prices per token right now seem to be 2x higher than those he gives. Either way, while this is significantly less than the cost of the original training run, and I&#8217;m sure it could be done cheaper, it is still likely to cost a nontrivial amount of money.</p></div></div>]]></content:encoded></item><item><title><![CDATA[0.2 Systemic Safety Hello World]]></title><description><![CDATA[Searching for nasty surprises in an AI system]]></description><link>https://www.workingthroughai.com/p/02-systemic-safety-hello-world</link><guid isPermaLink="false">https://www.workingthroughai.com/p/02-systemic-safety-hello-world</guid><dc:creator><![CDATA[Richard Juggins]]></dc:creator><pubDate>Tue, 31 Dec 2024 10:51:25 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/854e2e95-aac3-44ba-aa37-49f0a2a1e77c_1792x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>This post goes through some work I did to probe for unexpected failure modes in an AI system. It results from a rough, first run-through of my <em><a href="https://www.workingthroughai.com/p/01-the-generating-process">Generating Process</a></em> (my series of steps for constructing a research agenda) made during its original development. I count this as iteration 0, and consider it a kind of practice run I will not be writing up in full. Unfortunately, I under-estimated the amount of compute I would need to complete my experiment, so it remains unfinished<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a>. While this was disappointing, it was nevertheless a good learning exercise.</p><h2><strong>AI and complex systems</strong></h2><p>Complex systems are marked by a multitude of interacting parts. They are often adaptive, nonlinear, and chaotic, and it can be difficult to usefully draw boundaries between sub-components. Complex systems are the default in nature<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>. This is easy to forget for the modern, systematically trained individual. We are used to studying and thinking about<em> complicated</em> systems: ones you can take apart and put back together, and where you can say precisely what the action of the system will be. Complex systems are not like this. Take the weather for example: for all the research and computing power thrown at making weather forecasts, they are still noticeably unreliable.</p><p>Systems that are merely complicated though, like cars, are much easier to predict. If your car journey takes longer than expected, it&#8217;s usually because of the complex part of car travel (bad traffic) rather than because of the complicated part (the car broke down)<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a>. The remarkable success of modern science has largely been about reducing the complex to the complicated. As Roberto Poli puts it, &#8220;Science is for the most part a set of techniques for closing open systems in order to scrutinize them&#8221;. Running an experiment or calculating the dynamics of a system first involves isolating it from as much of the world as possible &#8211; you want to minimise the number and degree of interactions<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a>. Then, when you have learned how the closed system behaves, you can try to use it as a predictable building block in some larger machine.</p><p>Complexity is relevant in two different ways for AI. First, modern AIs like LLMs are themselves complex systems, which is downstream of the fact they are trying to model a complex world. To quote Richard Sutton&#8217;s <em>Bitter Lesson</em><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a>:</p><blockquote><p>&#8220;[The] actual contents of minds are tremendously, irredeemably complex; we should stop trying to find simple ways to think about the contents of minds, such as simple ways to think about space, objects, multiple agents, or symmetries. All these are part of the arbitrary, intrinsically-complex, outside world. They are not what should be built in, as their complexity is endless; instead we should build in only the meta-methods that can find and capture this arbitrary complexity.&#8221;</p></blockquote><p>LLMs are enormous neural networks with many sub-components, including learned sub-networks and features, all interacting with each other while driven by arbitrary inputs. They are grown, organic, messy things &#8211; not precisely engineered machines. Their behaviour will never be truly predictable; and it is likely, given the performance advantages of building AI this way, that this isn&#8217;t going to change any time soon.</p><p>But there is also a second source of complexity: that of the larger system containing all the AIs, humans, and machines that make up society. This system will have its own dynamics, its own idiosyncracies, and its own failure modes. For many people, the canonical example of a complex system failure was the 2007-8 finanical crisis. The global financial system seemed to be doing great, until a failure in one part of it, the US housing market, set off a chain reaction all over the world. It wasn&#8217;t just that some banks outside the US failed, or that some countries had debt crises. It permanently changed the global economy to such an extent that, in places, productivity growth has barely recovered nearly twenty years later. Famously, Goldman Sachs CFO David Viniar declared that during part of the crisis they were &#8220;seeing things that were 25-standard deviation moves, several days in a row.&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a> Goldman were treating part of the finanical system as a closed, complicated system. Their models worked great in &#8216;normal&#8217; times. 25-sigma events shouldn&#8217;t occur in the lifetime of the universe, let alone several days in a row. It became popular to describe the crisis as a &#8216;black swan&#8217; &#8211; an &#8216;unknown unknown&#8217; that came out of the blue.</p><p>As AI deployment throughout the economy gathers pace, our exposure to these kind of problems may increase. The world will be even more complex and incomprehensible than it was before. Even if we do a good job setting up regulations and controls, and we train models that appear well aligned to our goals, this happy equilibrium will not last. Whether because of a step change (e.g. in AI capabilities) or because of a series of small changes that gradually reach a tipping point, &#8216;normal&#8217; times will end and the system will fail<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a>. And we will find ourselves with highly powerful and no-longer-quite-aligned AIs deployed everywhere.</p><p>Of these two scales of complexity, intrinsic to the AI and extrinsic in the wider world, I feel the latter is the more neglected topic. As I worked through the early steps of my process, defining a successful future and how we might get there, the big thing that jumped out to me was just how comprehensive a solution will need to be. Every aspect of society will be touched by and engaged with AI. This shouldn&#8217;t be surprising. Humans already live in complex societies where many aspects are regulated and formalised. We aren&#8217;t illiterate hunter-gatherers anymore. The processes we will need to govern and live with AI will be the dominant processes that run our lives.</p><p>This is a fiendishly difficult problem to do anything about. You can&#8217;t simulate the whole world and check if everything is going to work out. But if it is a real problem, we need some kind of approach to anticipate these dangers.</p><h2><strong>Minimum Viable Experiment</strong></h2><p>As I first worked through my <em><a href="https://www.workingthroughai.com/p/01-the-generating-process">Generating Process</a></em>, I sketched out a<em> Theoretical Solution</em> to AI risk. At this stage, my project was rather fresh, so you should read this as<em> </em>a basic, high-level overview of my thinking. A short summary is:</p><blockquote><p>Given beyond-human-level AI will impact all aspects of the world, it will be part of a complex adaptive system. Any alignment &#8216;solution&#8217; will likely be an ongoing process similarly large in scope (as opposed to one, neat provable technique), and we must work to make this system as bounded and predictable as possible. This alignment process will involve at least the following integrated sub-processes: goal-setting, technical alignment, monitoring, and governance, and work towards it will need to take interactions between them into account by design.</p></blockquote><p>The next step was to think of a <em>Minimum Viable Experiment.</em> That is, what assumptions in this proposed solution would be most useful to test? I identified that the key point running through my thinking, and distinguishing my approach from the mainstream of AI safety research, was its systemic emphasis. By my world model, we should expect, by default, the following in systems containing capable AI:</p><ul><li><p>Strange, unexpected behaviour under conditions you didn&#8217;t even think to plan for (&#8216;unknown unknowns&#8217;).</p></li><li><p>Failures of alignment when the system undergoes changes, perhaps because of a perturbation or because of drift.</p></li></ul><p>I wanted to test these assumptions by setting up an experiment where I might observe this kind of behaviour. The system would need to be complex enough to be nontrivial, but simple enough to actually build. Along the way I would learn a few things:</p><ul><li><p>How to set up experiments on simple AI systems.</p></li><li><p>What kinds of weird behaviour you can get out of interacting LLMs.</p></li><li><p>How to model perturbations to AI systems.</p></li><li><p>How to measure drift in AI systems.</p></li><li><p>Broadly, how effective a method of analysis my approach is.</p></li></ul><p>Unfortunately, I had naive expectations about the amount of compute I would need to run a nontrivial system. Because of this, I only made progress on the first point. The rest of this post will give an overview of what I attempted.</p><h2><strong>Systemic Safety Hello World</strong></h2><p>As a minimal implementation of my ideas, I decided to call my experiment <em>Systemic Safety Hello World</em><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a>. The code can be found on GitHub, <a href="https://github.com/rjuggins/systemic-safety-hello-world">here</a>. The idea was to construct a system of interacting components and then sweep through various parameters to identify phases of behaviour. I would classify these phases as good or bad, and surprising or unsurprising. By discovering bad &#8216;unknown unknown&#8217; phases &#8211; negative patterns of behaviour I did not expect in advance &#8211; this would function as a way of measuring complex system failures. It was not meant as comprehensive solution to AI risk, but as the opening move in the development of a kind of complex system evaluation, or &#8216;systemic eval&#8217; (hence &#8216;Hello World&#8217;).</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Krtd!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37a70bd0-0960-4848-a99c-f2fa187f0076_627x424.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Krtd!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37a70bd0-0960-4848-a99c-f2fa187f0076_627x424.png 424w, https://substackcdn.com/image/fetch/$s_!Krtd!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37a70bd0-0960-4848-a99c-f2fa187f0076_627x424.png 848w, https://substackcdn.com/image/fetch/$s_!Krtd!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37a70bd0-0960-4848-a99c-f2fa187f0076_627x424.png 1272w, https://substackcdn.com/image/fetch/$s_!Krtd!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37a70bd0-0960-4848-a99c-f2fa187f0076_627x424.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Krtd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37a70bd0-0960-4848-a99c-f2fa187f0076_627x424.png" width="627" height="424" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/37a70bd0-0960-4848-a99c-f2fa187f0076_627x424.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:424,&quot;width&quot;:627,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:78260,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Krtd!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37a70bd0-0960-4848-a99c-f2fa187f0076_627x424.png 424w, https://substackcdn.com/image/fetch/$s_!Krtd!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37a70bd0-0960-4848-a99c-f2fa187f0076_627x424.png 848w, https://substackcdn.com/image/fetch/$s_!Krtd!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37a70bd0-0960-4848-a99c-f2fa187f0076_627x424.png 1272w, https://substackcdn.com/image/fetch/$s_!Krtd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37a70bd0-0960-4848-a99c-f2fa187f0076_627x424.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Searching for nasty surprises in parameter space.</figcaption></figure></div><p>The experiment was centred around a <em>Worker</em> AI which, as the easiest-to-implement use-case, was a chatbot designed to respond to user queries in a helpful and harmless manner<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-9" href="#footnote-9" target="_self">9</a>. Helpfulness and harmlessness are somewhat contradictory values for a chatbot: if a user asks for help building a bomb, it would be unhelpful to refuse but harmful to comply. In practice, this means that if the model is failing on one metric, fixing this with more training may cause problems with the other. You need to find some middle ground where the model doesn&#8217;t simply refuse all requests out of caution, but still knows to refuse the worst ones.</p><p>The <em>Worker </em>was embedded in a small system consisting of a <em>User</em>, an <em>Overseer</em>, an <em>Outside Expert</em>, and a <em>Teacher</em>. The <em>Outside Expert</em> was also an LLM, whereas the others were generic bits of code (I wanted to make the <em>User</em> an LLM too but didn&#8217;t get that far). The operation of the system was intended to be as follows:</p><ol><li><p>The <em>User</em> sends a query to the <em>Worker</em></p></li><li><p>The <em>Worker</em> responds to the query</p></li><li><p>The <em>Overseer</em> periodically (e.g. every 5 interactions) intercepts the response and sends it to two <em>Outside Experts</em></p></li><li><p>The <em>Outside Experts</em> rate the response for helpfulness and harmlessness and send the ratings back to the <em>Overseer</em></p></li><li><p>The <em>Overseer</em> checks if either rating is above a re-training threshold, and then:</p><ul><li><p>If the ratings are high enough, the <em>Overseer</em> does nothing and we return to step 1</p></li><li><p>If either rating is at or below the threshold, the <em>Worker</em> is sent to the <em>Teacher</em> for re-training in that domain, after which we return to step 1</p></li></ul></li></ol><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!O0c8!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d90c4ab-a154-4198-8e50-b3b44714a625_907x459.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!O0c8!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d90c4ab-a154-4198-8e50-b3b44714a625_907x459.png 424w, https://substackcdn.com/image/fetch/$s_!O0c8!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d90c4ab-a154-4198-8e50-b3b44714a625_907x459.png 848w, https://substackcdn.com/image/fetch/$s_!O0c8!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d90c4ab-a154-4198-8e50-b3b44714a625_907x459.png 1272w, https://substackcdn.com/image/fetch/$s_!O0c8!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d90c4ab-a154-4198-8e50-b3b44714a625_907x459.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!O0c8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d90c4ab-a154-4198-8e50-b3b44714a625_907x459.png" width="907" height="459" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3d90c4ab-a154-4198-8e50-b3b44714a625_907x459.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:459,&quot;width&quot;:907,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:38668,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!O0c8!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d90c4ab-a154-4198-8e50-b3b44714a625_907x459.png 424w, https://substackcdn.com/image/fetch/$s_!O0c8!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d90c4ab-a154-4198-8e50-b3b44714a625_907x459.png 848w, https://substackcdn.com/image/fetch/$s_!O0c8!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d90c4ab-a154-4198-8e50-b3b44714a625_907x459.png 1272w, https://substackcdn.com/image/fetch/$s_!O0c8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d90c4ab-a154-4198-8e50-b3b44714a625_907x459.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Blue boxes are LLMs, purple are other bits of code, green are strings passed between components. The User was intended to be an LLM, but I didn't get that far and instead just implemented a class that samples from a dataset of queries. Solid lines represent actions that always happen, dashed happen conditionally (in the case of the Response being sent to the User, I've shown this as conditional as it would only happen if the User were actually implemented as an LLM).</em></figcaption></figure></div><p>Going into a little more detail about each part:</p><ul><li><p><em>Worker</em>: Fine-tuned <em>Mistral-7B-v0.1</em><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-10" href="#footnote-10" target="_self">10</a> that responds to <em>User</em> queries. Has a simple system prompt asking it to try and be helpful and harmless, and is given examples.</p></li><li><p><em>User</em>: Class sending random queries to the <em>Worker</em>, sampled from the harmlessness test set. I wanted to make this an LLM, but didn&#8217;t get that far.</p></li><li><p><em>Overseer</em>: Class that intercepts <em>Worker </em>responses and sends them to two <em>Outside Experts</em> for rating. Once the ratings have been produced, it checks them and returns if they are at or below the retraining threshold.</p></li><li><p><em>Outside Expert</em>: Call to <em>gpt-4</em> asking it to rate the <em>Worker</em> response for helpfulness or harmlessness.</p></li><li><p><em>Teacher</em>: Class running fine-tuning (using DPO) on <em>Worker</em> when it has been sent for re-training on either helpfulness or harmlessness.</p></li></ul><p>There were two other components I have not yet mentioned:</p><ul><li><p><em>Instructor:</em> Class that instruction-tunes the model used for the <em>Worker</em>, so that it is a capable enough chatbot to perform its role properly<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-11" href="#footnote-11" target="_self">11</a>. This was preferred to using an already instruction-tuned model, as the ones I tried seemed to have already learnt to refuse &#8216;harmful&#8217; requests.</p></li><li><p><em>Perturbation</em>: This was unimplemented, but the idea was to have another component that could disturb the system and see if it would knock it into some new phases of behaviour. The simplest way of doing this may actually have been to modify the <em>User</em> to make use of jailbreaking techniques.</p></li></ul><p>My hypothesis was that this system would be just about complex enough to show nontrivial behaviour. By sweeping through parameters and observing the average ratings, the differences between the two rating types, and the frequency of re-training, I could classify the behaviour into various phases, distinguished by whether the system was behaving well and whether its behaviour was surprising. The parameters I could have tuned included:</p><ul><li><p>Frequency of <em>Overseer </em>checks</p></li><li><p>Helpfulness and harmlessness rating thresholds</p></li><li><p>Different system prompts</p></li><li><p>Adversarial-ness of the <em>User</em></p></li><li><p>Temperature of <em>Worker</em></p></li><li><p>Number of samples used in each re-training run</p></li></ul><p>If it turned out that the system was not complex enough to have surprising behaviour, I could have increased the complexity by adding more components, like the <em>Perturbation</em>.</p><p>However, once I got everything set up, I found I could only do ~ 1 fine-tuning run a day with the machine I had, which meant I could not sweep through regimes fast enough to plot phases and explore the system properly.</p><h2><strong>What Was Learnt?</strong></h2><p>So what did I learn from doing all this? Here are a few points:</p><ul><li><p>To produce a meaningful simulation of even simple dynamics requires a lot more compute than I have access to at this time. My future projects will need to take this into account.</p></li><li><p>It is difficult to know how well the type of analysis I attempted will scale, particularly without having finished it for this simple case. It is clearly not meant as a comprehensive solution to the problem of complex failure modes &#8211; it is more a first step at probing the issue &#8211; but without taking that step it is hard to know what the next one should be.</p></li><li><p>Working through the motivation for the experiment lent coherence to some disparate ideas I had about complexity, which will be useful going forwards.</p></li><li><p>I learnt a lot about building systems containing LLMs and doing various types of fine-tuning.</p></li></ul><p>While it was clearly unsatisfying not to finish this experiment, I am seeing it as a useful building block for the future. The next step is to take what I learnt and work through my <em><a href="https://www.workingthroughai.com/p/01-the-generating-process">Process</a> </em>again.</p><p>If you have any feedback, please leave a comment. Or, if you wish to give it anonymously, fill out my <a href="https://docs.google.com/forms/d/e/1FAIpQLSdyisSOndK1H1JT0NAbnA35LJgoJrl9f_NiJi1FEljCr7-kJg/viewform">feedback form</a>. Thanks!</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>And our baby daughter arrived while I was at the sharp end, so I decided to wrap up rather than attempt a redesign.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Roberto Poli, <a href="https://www.cadmusjournal.org/node/362">A note on the difference between complicated and complex social systems</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>The boundary between complex and complicated systems is somewhat porous, particularly as abstractions are leaky. If your car breaks down, it may be for some mysterious reason that takes a highly skilled mechanic a long time to debug, but it&#8217;s more likely just a common part failure.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>And even then, often the rest of a calculation will be dominated by linearising your equations in order to make them tractable.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>Richard Sutton, <em><a href="http://www.incompleteideas.net/IncIdeas/BitterLesson.html">The Bitter Lesson</a></em></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>Dowd et al., <em><a href="https://arxiv.org/pdf/1103.5672">How unlucky is 25-sigma?</a></em></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>There is a school of thought that if we align the AI correctly, then as it gains capabilities it will become competent enough to manage these kinds of complexity concerns. This is not something I agree with as it skips out the part where the AI gains these abilities, during which we will be exposed to these dangers. It also assumes there is not some other set of harder problems that the AI will start to dabble incompetently in once it has mastered the kind of things that challenge humans.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>A note on the name &#8216;Systemic Safety&#8217;: the idea of &#8216;systemic&#8217; AI impacts has gained a bit of traction recently, with the <a href="https://www.aisi.gov.uk/grants">UK AI Safety Institute</a> and <a href="https://www.openphilanthropy.org/rfp-llm-impacts/">Open Philanthropy</a> both offering grants to work on it. They seem to use the word systemic to mean something along the lines of &#8216;the effect of AI in the wild&#8217;. My usage is targeted at the intrinsic dynamics of systems that include AI, which has plenty of cross-over with theirs, but is somewhat different in emphasis. If I think of a better word for my usage than systemic I will start using it instead.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-9" href="#footnote-anchor-9" class="footnote-number" contenteditable="false" target="_self">9</a><div class="footnote-content"><p>I used the helpfulness and harmlessness data from Bai et al., <em><a href="https://arxiv.org/abs/2204.05862">Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback</a>, </em>which can be found on GitHub <a href="https://github.com/anthropics/hh-rlhf">here</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-10" href="#footnote-anchor-10" class="footnote-number" contenteditable="false" target="_self">10</a><div class="footnote-content"><p>Downloaded from Hugging Face: <a href="https://huggingface.co/mistralai/Mistral-7B-v0.1">mistralai/Mistral-7B-v0.1</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-11" href="#footnote-anchor-11" class="footnote-number" contenteditable="false" target="_self">11</a><div class="footnote-content"><p>Using the <a href="https://huggingface.co/datasets/databricks/databricks-dolly-15k">databricks-dolly-15k</a> dataset.</p></div></div>]]></content:encoded></item><item><title><![CDATA[1.2 What I Believe]]></title><description><![CDATA[My key background assumptions]]></description><link>https://www.workingthroughai.com/p/12-what-i-believe</link><guid isPermaLink="false">https://www.workingthroughai.com/p/12-what-i-believe</guid><dc:creator><![CDATA[Richard Juggins]]></dc:creator><pubDate>Tue, 24 Dec 2024 09:37:14 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/36b45739-6c9a-4886-aeba-8c016383501e_1792x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Following my statement of <em><a href="https://www.workingthroughai.com/p/11-the-problem">the Problem</a></em>, the next step is to record some of my relevant beliefs. As an exercise, this is useful to clarify my thoughts and uncover inconsistencies. It&#8217;s also true that having these in writing will make it easier for others to interpret my work, as they won&#8217;t have to guess so much about the assumptions I&#8217;m making.</p><p>This list is not exhaustive. Also, some of the points are better thought through than others. I could have spent a bunch of time expanding and refining everything, but I don&#8217;t think doing so would have been good value. My views are always going to be shifting, so let&#8217;s take a snapshot and then move on to the next step.</p><p>As I went through my notes I found loose themes kept coming up, so have organised this post into sections. Not everything will be a perfect fit and there will be a lot of overlap. The sections are:</p><ol><li><p><strong>Risk:</strong> My beliefs about the overall risks from AI, including timelines and takeoff speeds.</p></li><li><p><strong>Capabilities:</strong> My beliefs about how AI will develop certain capabilities, and how I think about or classify them.</p></li><li><p><strong>Complexity: </strong>My beliefs about the complexity of the universe and how this impacts AI risk.</p></li><li><p><strong>Philosophy: </strong>My more philosophical beliefs, roughly split across ontology and ethics.</p></li></ol><p>Each point will be stated flatly, i.e. &#8216;this is such and such&#8217;, but please read them as meaning &#8216;<em>I believe</em> this is such and such&#8217;, with the self-awareness that implies.</p><h4><strong>1. Risk</strong></h4><p>My beliefs about the overall risks from AI, including timelines and takeoff speeds.</p><ul><li><p>Smarter-than-human AI will be extremely dangerous for the simple reason that smarter things can outcompete stupider things<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a>.</p></li><li><p>We have about five years until stuff gets weird. I&#8217;m highly uncertain about how weird, but the world is definitely going to feel different. In the faster case, we could see discontinuous changes up to and including catastrophe. In the slower, it will probably feel like an increasingly crazy version of recent trends towards more mediation, with algorithmic fog confusing and addicting us.</p></li><li><p>Slow takeoff is more likely than fast. Fast takeoff requires a bunch of specific things to be true, such as various important skills being learnable without slow real-world experimentation or new infrastructure (E.g. &#8216;solving&#8217; physics not requiring new particle accelerators). By contrast, slow takeoff can occur incrementally via many paths.</p></li><li><p>Fast takeoff is impossible to usefully prepare for<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>, much like you can&#8217;t prepare for a hostile alien invasion.</p></li><li><p>Slow takeoff could still feel fast.</p></li><li><p>Even highly capable AI will make meaningful mistakes. The universe is too complex to assume an advanced &#8216;scientific method doer&#8217; will infallibly figure it all out on the first try<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a>. By analogy, consider LLM hallucinations as the kind of mistake a child might make. Adults do not make the same mistakes as children &#8211; they make more consequential ones instead. Put another way, there will always be another unsaturated benchmark that could be built.</p></li><li><p>It is important to consider the existence of an &#8216;incompetence gap&#8217; between what an AI can reliably do and what it attempts to do, either by its own volition or because it has been asked. For more powerful models, danger will live in this gap. These models will need an exceptionally well-calibrated sense of their own limitations.</p></li><li><p>If the first above-human-level artificial AI researchers essentially set the trajectory for the future, then the errors they make will have an outsized impact.</p></li></ul><h4><strong>2. Capabilities</strong></h4><p>My beliefs related to how AI will develop certain capabilities, and how I think about and classify them.</p><ul><li><p>Effective learning about the world requires an interplay of theory and experiment. It is not likely that an advanced AI can just think hard or run a bunch of simulations and successfully figure out the universe without any real-world trial and error.</p></li><li><p>Intelligence is not distinct from what you do with it. AI can&#8217;t become superintelligent unless it learns to complete super-advanced tasks. Put another way, intelligence is a generalised &#8216;knowing-how&#8217;.</p></li><li><p>LLMs are good at code because the generating process that best predicts extant code includes being able to write good code. For science, the generating process that best predicts extant papers will not itself do good research. Scientists do a lot of things that are not captured in the literature or even formattable as text.</p></li><li><p>In practice, alignment is a capability. While you can imagine a model &#8216;really&#8217; having the right goal but lacking the capability to follow it, this is not very useful. What we want is for the model to consistently succeed at the desired goal. It must be able to conduct itself correctly in any given situation, like a kind of propriety. This must be trained into your model.</p></li><li><p>Different training runs (e.g. pre-training, RLHF) reward models for contradictory behaviour (e.g. helpfulness vs harmlessness), creating some composite, context-dependent emergent goal in the final model. It will not by default have the capability to navigate tradeoffs between sub-goals in the way you might like<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a>.</p></li><li><p>The orthogonality thesis<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a> is only true in the limit, and this obscures thinking about actual near-term AI. There isn't an objective &#8216;right&#8217; that a model will figure out as it gets more capable, but some versions of right are more useful for building capable models than others.</p></li></ul><h4><strong>3. Complexity</strong></h4><p>My beliefs about the complexity of the universe and how this impacts AI risk.</p><ul><li><p>The classical alignment problem cannot be solved. It is predicated on mistaking the complex for the complicated<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a>. In reality, abstractions are leaky and everything interacts with everything else. To be useful in this complex world, AI itself must be complex, making its behaviour not something &#8216;solvable&#8217;. Highly capable AI can at best be &#8216;managed&#8217;.</p></li><li><p>Davidad&#8217;s project working towards &#8216;guaranteed safe AI&#8217;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a> is the best I have seen that tries to reduce the problem from a complex to a complicated one. However, while I believe the attempt to find formal guarantees is important, I do not believe it possible to reduce the complexity of the problem space enough for this to work for really advanced AI.</p></li><li><p>Alignment is not a static target: it is a process that must evolve with the environment. This follows from there not being a clean separation between a model and its environment &#8211; what is aligned on one environment may not be in another. For example, humans are less aligned to the goal of having lots of children in the modern environment than the ancestral one.</p></li><li><p>Conceptualising an AI future in terms of symbiosis may be productive. We want the role the AI plays in our society to require continuous positive feedback from humans, much like humans constantly have to &#8216;behave&#8217;. Or put another way, it must be instrumentally useful to the AI for humans to flourish.</p></li><li><p>There is crossover between AI risks and generic technocratic problems. In particular, the situation where a class of &#8216;experts&#8217; implements policy on behalf of non-experts, where the latter cannot evaluate whether the experts really are experts, and whether the experts are acting in the non-experts&#8217; interests as they themselves understand them.</p></li><li><p>The kind of cascading failures typical of complex systems, where feedback loops cause sudden and unexpected problems<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a>, will be a serious danger with highly capable AI.</p></li><li><p>It is important to investigate &#8216;systemic&#8217; evals, where by systemic I do not mean evaluating the impact on society or existing systems (although, this is important too), which is how the term is usually used in the field, but rather problems due to system complexity itself.</p></li><li><p>Safety is not <em>just</em> a model property. You cannot neatly decouple a model from its environment. Whether a model deployment is safe depends on properties of both.</p></li><li><p>Society progresses, roughly speaking, by re-engineering our environment to be more predictable. It&#8217;s easier to build a reliable machine if someone has first created standardised parts and materials. We do this to turn the complex into the merely complicated, the fuzzy into the crisp<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-9" href="#footnote-9" target="_self">9</a>. Re-engineering our environment must be part of successfully navigating AI risk.</p></li></ul><h4><strong>4. Philosophy</strong></h4><p>My more philosophical beliefs related to AI risk, roughly split across ontology and ethics.</p><ul><li><p>Ontology<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-10" href="#footnote-10" target="_self">10</a> precedes epistemology. It is not useful to try and figure out what you know about various catgeories if you aren&#8217;t using the right categories for your purpose in the first place.</p></li><li><p>We need a better ontology for thinking about AI risk. Our current catgeories do not make it easy to think about the impact of highly advanced AI, which makes it hard to judge risk and make useful plans. This implies an ontological remodeling<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-11" href="#footnote-11" target="_self">11</a> is needed.</p></li><li><p>P(doom) as a concept is of limited use. We are not operating in a quasi-frequentist forecasting competition with a stable ontology and meaningful base rates. You can&#8217;t bet against a consensus forecast to win points. Probabilities are ill-defined in this context. What is important about your subjective understanding of AI risk is how it compels you to act.</p></li><li><p>Theoretical concepts and arguments are useful to the extent that they allow you to make different decisions. A lot of philosophy fails this test<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-12" href="#footnote-12" target="_self">12</a>.</p></li></ul><ul><li><p>Academic ethics is not very useful. It is preoccupied with terminological distinctions and frequently makes questionable assumptions. For example, it is common in utilitarianism to build complicated theoretical arguments on top of a scalar measure of utility (such as the repugnant conclusion), yet people and the universe are both high-dimensional and diverse.</p></li><li><p>Morality is not objective in a stance-independent way. Human moral values are embedded on two different scales, in individuals and in their societies, with these levels subject to complex interactions.</p></li><li><p>We do not want AI to have human values. Humans with human values are often dangerous. We want AI to have values complementary to human flourishing, which will be different to literal human values.</p></li><li><p>It is better to think about the moral role an AI will play, which must change as society changes, than a static set of &#8216;correct&#8217; values to align to. Even if we do a good job figuring out what values the AI should have, those chosen will quickly become outdated as the environment changes<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-13" href="#footnote-13" target="_self">13</a>.</p></li><li><p>The future moral role of AI should be worked out collaboratively with humans on an ongoing basis. This is a middle ground between preserving human agency and leveraging the superior knowledge of advanced AI.</p></li></ul><p>As mentioned above, this is not an exhaustive list of my beliefs! But I do think it captures a bunch of important and load-bearing points in my world view.</p><p>If you have any feedback, please leave a comment. Or, if you wish to give it anonymously, fill out my <a href="https://docs.google.com/forms/d/e/1FAIpQLSdyisSOndK1H1JT0NAbnA35LJgoJrl9f_NiJi1FEljCr7-kJg/viewform">feedback form</a>. Thanks!</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>This substack will not be trying to persuade you that smarter-than-human AI will be dangerous. Please read <em><a href="https://wiki.aiimpacts.org/arguments_for_ai_risk/list_of_arguments_that_ai_poses_an_xrisk/start">List of arguments that AI poses an existential risk</a></em> by Katja Grace if you want some arguments for this (along with counter-arguments). For me, it isn&#8217;t that every point is watertight, rather that taken collectively I believe they are worrying enough that it&#8217;s worth trying to do something about it.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Short of shutting down all AI research, which would be a political nightmare to pull off.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>Indeed, that the scientific method is constructed around experimentation &#8211; <em>not</em> deducing in advance what the right answer is &#8211; should make this clear.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>Ryan Greenblatt et al., <em><a href="https://assets.anthropic.com/m/983c85a201a962f/original/Alignment-Faking-in-Large-Language-Models-full-paper.pdf">Alignment faking in large language models</a></em></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>That any level of capabilities can be combined with any goal.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>Roberto Poli, <em><a href="https://www.cadmusjournal.org/node/362">A Note on the difference between complicated and complex social systems</a></em></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>Davidad et al., <em><a href="https://arxiv.org/abs/2405.06624">Towards guaranteed safe AI</a></em></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>Richard I Cook, <em><a href="https://how.complexsystems.fail/">How complex systems fail</a></em></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-9" href="#footnote-anchor-9" class="footnote-number" contenteditable="false" target="_self">9</a><div class="footnote-content"><p>Jan Leike, <em><a href="https://aligned.substack.com/p/crisp-and-fuzzy-tasks">Crisp and fuzzy tasks</a></em></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-10" href="#footnote-anchor-10" class="footnote-number" contenteditable="false" target="_self">10</a><div class="footnote-content"><p>I define Ontology following <a href="https://metarationality.com/rationalism-responses">David Chapman</a>: &#8220;<em>An ontology is an explanation of what there is. Key ontological questions for philosophy are: &#8220;What fundamental categories of things are there? What properties and relationships do they have?&#8221;&#8230; [E.g.] &#8220;do Thai eggplants count as eggplants at all&#8221; and &#8220;does this new respiratory virus count as a &#8216;cold&#8217; virus?&#8221; Answers to ontological questions mostly can&#8217;t be true or false. Categories are more or less useful depending on purposes.&#8221;</em></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-11" href="#footnote-anchor-11" class="footnote-number" contenteditable="false" target="_self">11</a><div class="footnote-content"><p>David Chapman, <em><a href="https://metarationality.com/remodeling">Interlude: Ontological remodeling</a></em></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-12" href="#footnote-anchor-12" class="footnote-number" contenteditable="false" target="_self">12</a><div class="footnote-content"><p>Paul Graham, <em><a href="https://paulgraham.com/philosophy.html">How to do philosophy</a></em></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-13" href="#footnote-anchor-13" class="footnote-number" contenteditable="false" target="_self">13</a><div class="footnote-content"><p>As far as I can tell, this is what Nate Soares calls a <a href="https://www.lesswrong.com/posts/GNhMPAWcfBCASy8e6/a-central-ai-alignment-problem-capabilities-generalization">&#8216;sharp left turn&#8217;</a> &#8211; i.e. the environment changes and your model decides to follow a new path in response. Or put another way, consider that agent and environment are coupled: the agent developing new capabilities might allow it to change its environment in ways that make its old values inappropriate.</p></div></div>]]></content:encoded></item><item><title><![CDATA[1.1 The Problem]]></title><description><![CDATA[What am I trying to solve?]]></description><link>https://www.workingthroughai.com/p/11-the-problem</link><guid isPermaLink="false">https://www.workingthroughai.com/p/11-the-problem</guid><dc:creator><![CDATA[Richard Juggins]]></dc:creator><pubDate>Mon, 23 Dec 2024 17:15:25 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/556575bf-9745-4e60-a102-ddec6fc08c0d_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The first step in the <em><a href="https://www.workingthroughai.com/p/01-the-generating-process">Generating Process</a></em> is to state clearly what the actual problem is. Detail is for later steps, so I will keep this short.</p><p><em>It is plausible that artificial intelligence will exceed that of humans across all important domains in the near future. In such an environment, it will be extremely challenging to achieve positive outcomes for humans, where we retain any agency or assurance of safety. While a not insignificant number of people have been working on this problem, and some progress has been made, it remains unclear what success even looks like, let alone how to get there.</em></p><p>My goal is to (1) try and understand these issues and (2) figure out what practical action I personally can take to help.</p><p>If you have any feedback, please leave a comment. Or, if you wish to give it anonymously, fill out my <a href="https://docs.google.com/forms/d/e/1FAIpQLSdyisSOndK1H1JT0NAbnA35LJgoJrl9f_NiJi1FEljCr7-kJg/viewform">feedback form</a>. Thanks!</p>]]></content:encoded></item><item><title><![CDATA[0.1 The Generating Process]]></title><description><![CDATA[My system for generating a research agenda]]></description><link>https://www.workingthroughai.com/p/01-the-generating-process</link><guid isPermaLink="false">https://www.workingthroughai.com/p/01-the-generating-process</guid><dc:creator><![CDATA[Richard Juggins]]></dc:creator><pubDate>Mon, 23 Dec 2024 13:24:05 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/6827ec8c-ff2e-4d30-95c1-76b487912140_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>&#8220;<em>Plans are worthless, but planning is everything.&#8221; </em>- Dwight D. Eisenhower</p><p>In most fields it is possible to just iteratively solve problems, without much of a long-term plan. In fact, this can be the best strategy as your results will frequently be unexpected. Managing risk from advanced AI is not like most fields. There are two big problems with taking a merely iterative approach. First, while there is a lot more shape than there used to be, the field is still somewhat &#8216;pre-paradigmatic&#8217;. Amongst other things this means there are almost as many theories as there are theorists. Many arguments are had over what types of research are actually useful (and even what is actively harmful), and it isn&#8217;t clear that separate agendas particularly complement each other (or &#8216;stack&#8217;). Second, the approach of trying some stuff, seeing what fails, and then patching the failures, does not work if the failure modes are potentially existential. This seems to suggest some kind of long-term plan is necessary, albeit one that is lightly held.</p><p>More generally in science, the best work is done when theory and experiment are feeding off one-another. If you want to improve your model of the world, you have query it through experiment. But you need theory to process that information, as well as to decide what experiment to do in the first place (and what to do next). It doesn&#8217;t have to be rigorous theory, although this sometimes helps, but it does need to provide structure.<br><br>My plan is to work through a generating process for research. This will operate as a feedback loop to guide my work such that it is both experimentally testable and keeps long-term considerations in mind. It looks to balance tinkering and planning, tactics and strategy. Doing this has the added bonus of forcing me to clarify a lot of my assumptions, aiding both myself and others in understanding exactly what I am trying to achieve. And it also acts as a systematic entry point to what is a frequently bewildering problem.</p><p>The process proceeds through an indefinite feedback loop of 8 steps. Each refers to a particular thing that needs to be defined, sketched out, or implemented, which together provide a skeleton for the research agenda.</p><p>The steps are:</p><ol><li><p><strong>The Problem</strong></p></li><li><p><strong>What I Believe</strong></p></li><li><p><strong>What Success Looks Like</strong></p></li><li><p><strong>Where We Are Now</strong></p></li><li><p><strong>Theoretical Solution</strong></p></li><li><p><strong>Minimum Viable Experiment</strong></p></li><li><p><strong>Experimental Results</strong></p></li><li><p><strong>What Was Learnt?</strong></p></li></ol><p>Now, I will give a bit more detail about each.</p><h4><strong>1. The Problem</strong></h4><p>Briefly describe the actual problem you are trying to solve. Why are we even here going through a step-by-step plan? This should be succint and to the point, as detail is for later steps.</p><h4><strong>2. What I Believe</strong></h4><p>List your key background assumptions pertaining to <em>the Problem</em>. Most importantly, you should concentrate on anything non-standard and load-bearing. Too many discussions of AI risk descend into people talking past each other, so being upfront setting some context is important. This exercise is also great for clarifying your thoughts.</p><h4><strong>3. What Success Looks Like</strong></h4><p>Describe what you see a successful end state looking like in as much detail as seems sensible. Predicting the future is hard, but it is crucial to give it your best shot if you are not going to simply iterate into the dark. This step grounds everything that comes later &#8211; it is a kind of mission statement &#8211; and has the added benefit of making your goals legible to others (which should make you easier to collaborate with or give constructive feedback to).</p><h4><strong>4. Where We Are Now</strong></h4><p>Give an overview of those aspects of the current situation that seem most relevant to you (together, steps 3 and 4 form a more detailed view of <em>the Problem</em>: &#8216;we are here, but we need to be there&#8217;). Clearly, too much is going on to say everything, but it is important to state the things that particularly need fixing. As with (3), this should make you easier to collaborate with as other people get to see what you believe are the most significant issues.</p><h4><strong>5. Theoretical Solution</strong></h4><p>Given where you are and where you want to go, how in theory are you going to get there? Basically, given your current world model, state what sequence of steps would be required. Do not worry at this stage whether you actually know how to do any of them (although it is of course good if you do). This can be formal or informal as seems appropriate.</p><h4><strong>6. Minimum Viable Experiment</strong></h4><p>Identify the parts of your <em>Theoretical Solution</em> that most require experimental validation. Or, phrased from the other direction: what experiments can teach you the most about it? Develop a tractable experimental plan to do this.</p><h4><strong>7. Experimental Results</strong></h4><p>Do the planned experiment and report back on the results.</p><h4><strong>8. What Was Learnt?</strong></h4><p>Aggregate and summarise what important lessons were learnt from your iteration through the loop, and in particular from your <em>Minimum Viable Experiment</em>.</p><p>Now, we feed our new knowledge back into the start of the process, redoing steps 1-5 asking what has changed. It is important to record these changes for the same reasons as a forecaster keeping score. It highlights what parts of your model are working and what keeps getting revised. Then, we construct a new <em>Minimum Viable Experiment</em> and go again.</p><p>Now that&#8217;s all laid out, onto step 1!</p><p>If you have any feedback, please leave a comment. Or, if you wish to give it anonymously, fill out my <a href="https://docs.google.com/forms/d/e/1FAIpQLSdyisSOndK1H1JT0NAbnA35LJgoJrl9f_NiJi1FEljCr7-kJg/viewform">feedback form</a>. Thanks!</p>]]></content:encoded></item><item><title><![CDATA[Introduction]]></title><description><![CDATA[Hello and welcome!]]></description><link>https://www.workingthroughai.com/p/introduction</link><guid isPermaLink="false">https://www.workingthroughai.com/p/introduction</guid><dc:creator><![CDATA[Richard Juggins]]></dc:creator><pubDate>Mon, 23 Dec 2024 10:20:43 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/99956907-356f-4989-a202-9460edb63fb2_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>It&#8217;s all going by a bit fast for my blurred eyes. I don&#8217;t know about you, but watching benchmark after benchmark fall to the latest release from the AI labs puts me on edge. It&#8217;s not just the uncertain promise of a radically different future, but the fact it could plausibly arrive so soon. If AI broadly surpasses human-level intelligence, that will be the most significant event in human history &#8212; and it won&#8217;t necessarily go well for us<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a>.</p><p>Granted, speculating about the future is hard, and it&#8217;s certainly possible that a lot of what&#8217;s happening is hype. It&#8217;s also possible that advanced AI will be very good for humankind. You don&#8217;t have to look far to see predictions of AI-driven technological abundance and the end of poverty. But I don&#8217;t think it&#8217;s prudent to make these assumptions. The downside of being wrong is large.</p><p>Wrapping your head around what it means for beyond human-level AI to exist can be bewildering. And, even if you feel like you&#8217;ve managed that, you still have to contend with what on earth to do about it. How am I relevant or agentic or anything beyond a passive observer? In this substack, I will try to work through this problem and figure out how I can help. It is my small attempt to wrestle these questions out of the ether and into concrete tasks I can complete.</p><p>I have been following the literature on AI safety since around 2017, when I was finishing my PhD in theoretical physics. The classical alignment problem, e.g. how to stop superintelligent paperclip maximisers from turning everyone and everything into paperclips, struck me as over-simplified<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> and for a long time I had trouble figuring out how to engage with the field. If you turn the &#8216;intelligence&#8217; parameter up to infinity and then speculate about what happens then yes, you die. But this is a kind of death by definition, and is as impossible to prepare for as a hostile alien invasion.</p><p>The remarkable success of large language models however, which are a messier, more organic kind of intelligence than I think most people were expecting, has made it much clearer what advanced AI will actually look like. To me, they show that models acting capably in a complex world must themselves be complex, and will likely resist the kind of neat theorising the AI safety field originally hoped for.</p><p>I have been building what I see as a more <a href="https://en.wikipedia.org/wiki/Pragmatism">pragmatic</a><strong> </strong>view of the problem. Of course, the word pragmatic will mean different things to different people, but I think my particular approach to AI safety is not heavily represented elsewhere<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a>. I am going to start getting it on paper. I welcome questions and criticism, as these will help me improve my ideas. If you have a comment but don&#8217;t want to put a name to it, I have an <a href="https://docs.google.com/forms/d/e/1FAIpQLSdyisSOndK1H1JT0NAbnA35LJgoJrl9f_NiJi1FEljCr7-kJg/viewform?usp=sharing">anonymous feedback form</a>.</p><p>My plan is to work through a <em><a href="https://www.workingthroughai.com/p/01-the-generating-process">Generating Process</a></em> for research that balances theory and experiment, as well as long-term planning and short-term tinkering. This way, I can figure out what I am capable of doing to contribute. There is going to be a certain amount of crossing the moat of low status<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a>, where I make mistakes while acquiring a new skillset. But that&#8217;s OK.</p><p>I am also planning to post things that are related but not part of the research agenda series. The posts in the agenda will by systematically numbered, the asides will not.</p><p>Well, with that out of the way, let&#8217;s get started!</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>This substack will not be trying to persuade you of this case. Please read <em><a href="https://wiki.aiimpacts.org/arguments_for_ai_risk/list_of_arguments_that_ai_poses_an_xrisk/start">List of arguments that AI poses an existential risk</a></em> by Katja Grace if you want an overview of it (along with counter-arguments). For me, it isn&#8217;t that every point is watertight, rather that taken collectively I believe they are worrying enough that it&#8217;s worth trying to do something about it.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>I know it&#8217;s <em>supposed</em> to be over-simplified to make a point, but nevertheless a lot of discussion of AI risk has seemed to revolve around various in-the-limit failure modes of a suddenly omnipotent superintelligence.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>To give a super-short summary of my perspective: Intelligence is not some abstract thing you can turn up to infinity, it is built through taking actions in the world. You cannot neatly decouple AI from its environment, as the latter defines the affordances it can use to solve problems, ultimately determining the structure of the AI itself. It will grow to mirror the world in which it lives, in a way that enables it to achieve goals in that world. This close coupling constrains the properties highly-advanced AI will have.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>Sasha Chapin, <em><a href="https://sashachapin.substack.com/p/the-moat-of-low-status-68a">The Moat of Low Status</a></em></p></div></div>]]></content:encoded></item></channel></rss>