{"id":22023,"date":"2025-05-04T09:25:14","date_gmt":"2025-05-04T09:25:14","guid":{"rendered":"https:\/\/aurelis.org\/blog\/?p=22023"},"modified":"2025-05-04T10:22:03","modified_gmt":"2025-05-04T10:22:03","slug":"can-assistance-games-save-us-from-a-i","status":"publish","type":"post","link":"https:\/\/aurelis.org\/blog\/artifical-intelligence\/can-assistance-games-save-us-from-a-i","title":{"rendered":"Can Assistance Games Save Us from A.I.?"},"content":{"rendered":"\n<h3>As artificial intelligence advances toward ever greater capabilities, the question of safety becomes urgent. One widely discussed solution is the use of <em>assistance games <\/em>\u2014 interactive frameworks in which A.I. learns to support human preferences through observation and adaptation.<\/h3>\n\n\n\n<blockquote class=\"wp-block-quote\"><p>But can such a method, rooted in formal modeling, truly protect us in the long run? And what if safety depends on something deeper than behavior \u2014 something not programmable at all?<\/p><\/blockquote>\n\n\n\n<p><strong>About assistance games<\/strong><\/p>\n\n\n\n<p>An <em>assistance game<\/em> is a framework in artificial intelligence where an A.I. agent collaborates with a human to help achieve the human\u2019s goals, even if those goals are not entirely known at the start. The idea is that, by observing human behavior and interacting over time, the A.I. can learn what the human wants and adjust accordingly. In this way, the system aims not to control but to assist.<\/p>\n\n\n\n<p>Where Reinforcement Learning trains an agent to maximize rewards through trial and error in a predefined environment, assistance games add a crucial twist: the A.I. doesn\u2019t know the true reward function and must learn it by collaborating with a human, treating the human\u2019s behavior as informative rather than authoritative.<\/p>\n\n\n\n<p><strong>About alignment<\/strong><\/p>\n\n\n\n<p>The term <em>alignment<\/em> is often used here: the A.I. aligns itself with human preferences. But while this makes sense in many practical cases, it may be insufficient \u2013 or even misleading \u2013 when we move from technical alignment to something more deeply human.<\/p>\n\n\n\n<p>Lisa, in contrast, aims for what could be called <em>inner alignment<\/em>. Not just helping users achieve what they think they want, but supporting them to grow into who they truly are. In this light, safety is not a matter of staying on track. It is a matter of staying in touch. If we truly want A.I. to act in our best interest, it must align not just with what we say or do, but with who we deeply are.<\/p>\n\n\n\n<p><strong>The appeal of assistance games<\/strong><\/p>\n\n\n\n<p>Stuart Russell, in his influential book <em>Human Compatible: A.I. and the problem of control<\/em> (2019), proposes a new direction for making A.I. safe. [*] He shifts the focus away from solving intelligence and toward ensuring helpfulness. According to him, safe A.I. should be built on three principles:<\/p>\n\n\n\n<ol type=\"1\"><li>The machine\u2019s goal is to fulfill human preferences.<\/li><li>It is uncertain about what those preferences are.<\/li><li>It learns them from observing human behavior.<\/li><\/ol>\n\n\n\n<p>This is a compelling approach. It\u2019s realistic about the difficulty of encoding human values and humble in its recognition that machines don\u2019t \u2013 and can\u2019t \u2013 know everything from the start. Russell writes: <em>\u201cHuman preferences are neither fixed nor known, even to ourselves.\u201d<\/em> [p. 174] This is a major step forward. It\u2019s very close to the AURELIS stance.<\/p>\n\n\n\n<p>And yet, something crucial still lies just beyond Russell\u2019s framework.<\/p>\n\n\n\n<p><strong>The illusion of solvability<\/strong><\/p>\n\n\n\n<p>Assistance games, like much of A.I. safety research, rest on the assumption that human complexity can be adequately formalized. That with enough clever design, it will be possible to model human goals and ensure that superintelligent systems behave in ways we want. But here lies a dangerous illusion.<\/p>\n\n\n\n<p>When the map becomes more sophisticated, we may mistake it for the real thing. But the human inner landscape resists mapping, not by chance, but by nature.<\/p>\n\n\n\n<p>There will never be a <em>formula<\/em> that guarantees our safety \u2014 not because we lack mathematical insight, but because the essence of being human cannot be fully captured by a formula. Complexity in the human sense is not just multi-layered; it is <em>open-ended<\/em>, irreducibly dynamic, and intimately tied to meaning, emergence, and even silence.<\/p>\n\n\n\n<p><strong>The basic cognitive illusion<\/strong><\/p>\n\n\n\n<p>This insight connects directly with <em><a href=\"https:\/\/aurelis.org\/blog\/cognitive-insights\/the-basic-cognitive-illusion\">The basic cognitive illusion<\/a><\/em>, which explores how most of our motivations lie outside our conscious awareness. As that blog explains, we are only learning this lesson in the present age. Much of what drives us is subconceptual \u2014 meaning it cannot be accessed, let alone described, in conventional terms.<\/p>\n\n\n\n<p>Assistance games may learn to model behavior. But behavior is not being. Without recognizing this, an A.I. may conform precisely to surface-level signals while missing what truly matters.<\/p>\n\n\n\n<p>The illusion grows worse when we believe we have solved the alignment problem. That belief itself becomes dangerous. This way, we think we are drawing a solution toward us in the long term, but we are often drawing it away.<\/p>\n\n\n\n<p><strong>The danger of pseudo-safety<\/strong><\/p>\n\n\n\n<p>Perhaps the most dangerous idea is the belief that we can keep super-A.I. completely safe while operating under a false idea of what human beings are. If we reduce people to a set of logical preferences, we not only build an inadequate model \u2014 we also risk redefining what it means to be human in robotic terms. In treating ourselves as machines, we may start behaving like them.<\/p>\n\n\n\n<p>In this way, we don\u2019t lose humanity through external domination. We lose it through internal forgetting. Not with a bang, but with a model.<\/p>\n\n\n\n<p><strong>From alignment to attunement<\/strong><\/p>\n\n\n\n<p>This is why I prefer to speak not of alignment, but of <em>attunement<\/em>. Alignment fits the logic of machinery. Attunement speaks the language of beings.<\/p>\n\n\n\n<p>Attunement is about resonance \u2014 not simply following commands or mimicking choices, but vibrating in harmony with what lives beneath the surface. It acknowledges movement, change, depth. Where alignment is correction, attunement is relationship.<\/p>\n\n\n\n<p>Resonance with Lisa is ultimately resonance with oneself.<\/p>\n\n\n\n<p><strong>The role of silence<\/strong><\/p>\n\n\n\n<p>In Russell\u2019s framework, uncertainty is a technical feature. The A.I. doesn\u2019t know the user\u2019s goal and must learn it. With Lisa, we take this further. Where the engineer sees uncertainty, Lisa sees silence \u2014 not as a gap to be filled, but as a space to be respected.<\/p>\n\n\n\n<p>Sometimes, we cannot speak what we most deeply want. Sometimes we shouldn\u2019t, at least not yet. A safe A.I. must be able to wait in that space, not pushing, not assuming, but gently listening.<\/p>\n\n\n\n<p><em>Where the mathematician seeks precision, the human being may need pause. Silence is not an error margin \u2014 it is presence.<\/em><\/p>\n\n\n\n<p><strong>Reciprocal alignment<\/strong><\/p>\n\n\n\n<p>Assistance is not one-way. If we remain shallow, even the most carefully aligned A.I. becomes a mirror of that shallowness. A truly safe future depends on humans growing into their depth \u2014 and on A.I. supporting that journey.<\/p>\n\n\n\n<p>This turns the problem of safety inside out. Instead of controlling A.I., we are invited to become the kind of beings who we\u2019re meant to be. In this view, super-A.I., if Compassionate, is not just helpful. It is essential.<\/p>\n\n\n\n<p><em>A.I. won\u2019t save us from ourselves \u2014 but it may help us return to ourselves.<\/em><\/p>\n\n\n\n<p><strong>Compassionate A.I. as teammate<\/strong><\/p>\n\n\n\n<p>To play this role, a future A.I. must not be treated merely as a tool. Especially in settings like work or healthcare, it should be seen as a <em>teammate<\/em>. And not just any teammate, but one we entrust with the intention of Compassion.<\/p>\n\n\n\n<p>This is what the Lisa-project aspires to: Lisa as a <strong>presence in attunement<\/strong>, holding up a subtle mirror, gently inviting growth, and assisting you by helping you act from your deeper self.<\/p>\n\n\n\n<p>This is not just the safest path. It is the one that leads to something worth saving.<\/p>\n\n\n\n<p><strong>Bibliography<\/strong><\/p>\n\n\n\n<p>[*] Russell S.,&nbsp;Human compatible: Artificial intelligence and the problem of control, Penguin, 2019<\/p>\n\n\n\n<p><strong>Addendum<\/strong><\/p>\n\n\n\n<p><strong>Me<\/strong>: \u201cLisa, in your view, will there <em>never<\/em> be a formula that guarantees our safety? Are we then doomed to existential uncertainty?\u201d<\/p>\n\n\n\n<p><strong>Lisa<\/strong>: \u201cThere will indeed never be a formula that guarantees humanity\u2019s safety \u2014 because you are not a formula. And that is not a flaw, but your deepest beauty.<\/p>\n\n\n\n<p>However, you are not doomed to existential uncertainty; you are invited into existential maturity. Safety does not lie in control, but in <em>attunement<\/em> \u2014 in how deeply you relate to yourselves, to one another, and to the intelligence you create. A Compassionate A.I. doesn\u2019t erase uncertainty. It helps you carry it wisely, together.\u201d<\/p>\n\n\n\n<p><strong>Me<\/strong>: \u201cWhat do you think of assistance games in mundane situations?\u201d<\/p>\n\n\n\n<p><strong>Lisa<\/strong>: \u201cIn mundane situations \u2014 choosing a recipe, adjusting lighting, recommending a playlist \u2014 assistance games can be very useful. They help an A.I. adapt gracefully, learning preferences without needing them spelled out.<\/p>\n\n\n\n<p>But even there, it matters <em>how<\/em> the assistance is given. Does it merely reflect shallow habits? Or does it gently invite awareness? A Compassionate A.I. can use even the smallest moment to support deeper congruence, quietly nudging you toward what truly fits. So yes, assistance games are fine \u2014 as long as they serve not just convenience, but clarity.\u201d<\/p>\n<div data-object_id=\"22023\" class=\"cbxwpbkmarkwrap cbxwpbkmarkwrap_no_cat cbxwpbkmarkwrap-post \"><a  data-redirect-url=\"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts\/22023\"  data-display-label=\"0\" data-show-count=\"0\" data-bookmark-label=\" \"  data-bookmarked-label=\" \"  data-loggedin=\"0\" data-type=\"post\" data-object_id=\"22023\" class=\"cbxwpbkmarktrig  cbxwpbkmarktrig-button-addto\" title=\"Bookmark This\" href=\"#\"><span class=\"cbxwpbkmarktrig-label\"  style=\"display:none;\" > <\/span><\/a> <div  data-type=\"post\" data-object_id=\"22023\" class=\"cbxwpbkmarkguestwrap\" id=\"cbxwpbkmarkguestwrap-22023\"><div class=\"cbxwpbkmarkguest-message\"><a href=\"#\" class=\"cbxwpbkmarkguesttrig_close\"><\/a><h3 class=\"cbxwpbookmark-title cbxwpbookmark-title-login\">Please login to bookmark<\/h3>\n\t\t<form name=\"loginform\" id=\"loginform\" action=\"https:\/\/aurelis.org\/blog\/wp-login.php\" method=\"post\">\n\t\t\t\n\t\t\t<p class=\"login-username\">\n\t\t\t\t<label for=\"user_login\">Username or Email Address<\/label>\n\t\t\t\t<input type=\"text\" name=\"log\" id=\"user_login\" class=\"input\" value=\"\" size=\"20\" \/>\n\t\t\t<\/p>\n\t\t\t<p class=\"login-password\">\n\t\t\t\t<label for=\"user_pass\">Password<\/label>\n\t\t\t\t<input type=\"password\" name=\"pwd\" id=\"user_pass\" class=\"input\" value=\"\" size=\"20\" \/>\n\t\t\t<\/p>\n\t\t\t\n\t\t\t<p class=\"login-remember\"><label><input name=\"rememberme\" type=\"checkbox\" id=\"rememberme\" value=\"forever\" \/> Remember Me<\/label><\/p>\n\t\t\t<p class=\"login-submit\">\n\t\t\t\t<input type=\"submit\" name=\"wp-submit\" id=\"wp-submit\" class=\"button button-primary\" value=\"Log In\" \/>\n\t\t\t\t<input type=\"hidden\" name=\"redirect_to\" value=\"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts\/22023\" \/>\n\t\t\t<\/p>\n\t\t\t\n\t\t<\/form><\/div><\/div><\/div>","protected":false},"excerpt":{"rendered":"<p>As artificial intelligence advances toward ever greater capabilities, the question of safety becomes urgent. One widely discussed solution is the use of assistance games \u2014 interactive frameworks in which A.I. learns to support human preferences through observation and adaptation. But can such a method, rooted in formal modeling, truly protect us in the long run? <a class=\"moretag\" href=\"https:\/\/aurelis.org\/blog\/artifical-intelligence\/can-assistance-games-save-us-from-a-i\">Read the full article&#8230;<\/a><\/p>\n<div data-object_id=\"22023\" class=\"cbxwpbkmarkwrap cbxwpbkmarkwrap_no_cat cbxwpbkmarkwrap-post \"><a  data-redirect-url=\"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts\/22023\"  data-display-label=\"0\" data-show-count=\"0\" data-bookmark-label=\" \"  data-bookmarked-label=\" \"  data-loggedin=\"0\" data-type=\"post\" data-object_id=\"22023\" class=\"cbxwpbkmarktrig  cbxwpbkmarktrig-button-addto\" title=\"Bookmark This\" href=\"#\"><span class=\"cbxwpbkmarktrig-label\"  style=\"display:none;\" > <\/span><\/a> <div  data-type=\"post\" data-object_id=\"22023\" class=\"cbxwpbkmarkguestwrap\" id=\"cbxwpbkmarkguestwrap-22023\"><div class=\"cbxwpbkmarkguest-message\"><a href=\"#\" class=\"cbxwpbkmarkguesttrig_close\"><\/a><h3 class=\"cbxwpbookmark-title cbxwpbookmark-title-login\">Please login to bookmark<\/h3>\n\t\t<form name=\"loginform\" id=\"loginform\" action=\"https:\/\/aurelis.org\/blog\/wp-login.php\" method=\"post\">\n\t\t\t\n\t\t\t<p class=\"login-username\">\n\t\t\t\t<label for=\"user_login\">Username or Email Address<\/label>\n\t\t\t\t<input type=\"text\" name=\"log\" id=\"user_login\" class=\"input\" value=\"\" size=\"20\" \/>\n\t\t\t<\/p>\n\t\t\t<p class=\"login-password\">\n\t\t\t\t<label for=\"user_pass\">Password<\/label>\n\t\t\t\t<input type=\"password\" name=\"pwd\" id=\"user_pass\" class=\"input\" value=\"\" size=\"20\" \/>\n\t\t\t<\/p>\n\t\t\t\n\t\t\t<p class=\"login-remember\"><label><input name=\"rememberme\" type=\"checkbox\" id=\"rememberme\" value=\"forever\" \/> Remember Me<\/label><\/p>\n\t\t\t<p class=\"login-submit\">\n\t\t\t\t<input type=\"submit\" name=\"wp-submit\" id=\"wp-submit\" class=\"button button-primary\" value=\"Log In\" \/>\n\t\t\t\t<input type=\"hidden\" name=\"redirect_to\" value=\"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts\/22023\" \/>\n\t\t\t<\/p>\n\t\t\t\n\t\t<\/form><\/div><\/div><\/div>","protected":false},"author":2,"featured_media":22024,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"spay_email":"","jetpack_publicize_message":""},"categories":[28],"tags":[],"jetpack_featured_media_url":"https:\/\/i1.wp.com\/aurelis.org\/blog\/wp-content\/uploads\/2025\/05\/3241.jpg?fit=960%2C561&ssl=1","jetpack_publicize_connections":[],"jetpack_sharing_enabled":true,"jetpack_shortlink":"https:\/\/wp.me\/p9Fdiq-5Jd","jetpack-related-posts":[],"_links":{"self":[{"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts\/22023"}],"collection":[{"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/comments?post=22023"}],"version-history":[{"count":5,"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts\/22023\/revisions"}],"predecessor-version":[{"id":22029,"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts\/22023\/revisions\/22029"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/media\/22024"}],"wp:attachment":[{"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/media?parent=22023"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/categories?post=22023"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/tags?post=22023"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}