{"id":18300,"date":"2024-11-23T17:58:00","date_gmt":"2024-11-23T17:58:00","guid":{"rendered":"https:\/\/aurelis.org\/blog\/?p=18300"},"modified":"2024-11-24T08:09:21","modified_gmt":"2024-11-24T08:09:21","slug":"the-compute-efficient-frontier","status":"publish","type":"post","link":"https:\/\/aurelis.org\/blog\/artifical-intelligence\/the-compute-efficient-frontier","title":{"rendered":"The Compute-Efficient Frontier"},"content":{"rendered":"\n<h3>Research on scaling laws for <a href=\"https:\/\/aurelis.org\/blog?p=15748\">LLMs<\/a> suggests that while scaling model size (functional parameters), dataset size (tokens), and compute (amount of computing power) improves performance, diminishing returns are becoming evident. This is called the \u2018compute-efficient frontier.\u2019<\/h3>\n\n\n\n<blockquote class=\"wp-block-quote\"><p>Apparently, it does not depend much on architectural details (such as network width or depth) as long as reasonably good choices are made. Thus, is this a fundamental law in nature or a result of our specific neural network approach to A.I. \u2015 just pointing to the need for innovation beyond scaling?<\/p><\/blockquote>\n\n\n\n<p><strong>Speculative aspirations<\/strong><\/p>\n\n\n\n<p>The \u2018race toward AGI\u2019 is speculative, mainly in the sense that it is a race toward the undefined. The diminishing returns through scaling are thus no surprise \u2015 simply because we don\u2019t know what we\u2019re doing or even heading to. This way, we\u2019re just throwing \u2019something\u2019 harder to \u2018somewhere\u2019 and see \u2013 hurray! \u2013 that it works better.<\/p>\n\n\n\n<p>Indeed, it works better in practical applications such as natural language processing, integration into IoT (Internet of Things), customer service, and spatial computing \u2015 but, as said, with diminishing returns.<\/p>\n\n\n\n<p>Meanwhile, the big speculation remains speculative. That may even be not such a bad thing in the domain of non-Compassionate A.I.<\/p>\n\n\n\n<p><strong>But we want to do better in (and through) Compassionate A.I.<\/strong><\/p>\n\n\n\n<p>Only here do we need to get closer to what we really want in this, which is the complexity of human mental <em>depth<\/em> (which is very different from neural network depth). Thus, we also need to see that human depth is where our thinking (and intelligence) really comes from.<\/p>\n\n\n\n<p>The prior \u2018A.I. winters\u2019 (downfall from very high expectations) came, in hindsight, from the hubris about <strong>ourselves<\/strong> \u2015 mainly at the conceptual level. Thus, researchers overestimated what A.I. can achieve based on a narrow understanding of human intelligence.<\/p>\n\n\n\n<p>With new possibilities in the field of ANNs, we relatively recently came into a new A.I. spring. These new possibilities are, indeed, scaling model size, dataset size, and compute. Meanwhile, the hubris has remained the same in that researchers generally still don\u2019t see where human depth arises or what it even entails \u2015 even while the same hubris makes them think they do.<\/p>\n\n\n\n<p>New toys \u2015 same players.<\/p>\n\n\n\n<p><strong>Can subconceptual processing, as seen in human brains, inspire breakthroughs?<\/strong><\/p>\n\n\n\n<p>[<em>subconceptual processing<\/em>: non-linear, pattern-based processes underlying conscious thought]<\/p>\n\n\n\n<p>Humans don\u2019t scale neurons infinitely to increase intelligence; instead, complexity arises from interconnectivity, efficiency, and adaptability. Thus, the aspiration for <strong>Compassionate A.I. \u2013 <\/strong>aligned with the complexity of human mental processes \u2013 requires a shift:<\/p>\n\n\n\n<ul><li>From surface to depth: Current systems excel at mimicking patterns but lack true depth, which encompasses things like <em>true<\/em> emotions\/feelings, values, and nuanced decision-making, which current systems struggle to replicate authentically.<\/li><li>Integration with subconceptual processing: This emerges from understanding how the human mind generates intelligence through dynamic processes at the subconceptual level. This requires deep systems that engage with complexity <em>meaningfully<\/em>, echoing the interconnectedness of human thought and emotion.<\/li><\/ul>\n\n\n\n<p>Progress will require a humbler approach and greater openness to the (real!) complexity of human depth rather than treating it as a secondary or irrelevant factor.<\/p>\n\n\n\n<p><strong>New toys, new players<\/strong><\/p>\n\n\n\n<p>The \u2018toys\u2019 \u2013 like large-scale neural networks \u2013 still operate on linear principles.<\/p>\n\n\n\n<p>Human intelligence, by contrast, thrives on self-organization, subconceptual layers, and emergent patterns. Innovations may arise by designing innovative models that deeply learn through continuous feedback loops with real-world environments, akin to how humans adapt.<\/p>\n\n\n\n<p><strong>Scaling depth<\/strong><\/p>\n\n\n\n<p>The key to a transformative A.I. future might lie in \u2018depth\u2019-first scaling:<\/p>\n\n\n\n<ul><li>Developing architectures that mimic not just the output but the processes themselves of human cognition \u2015 e.g. in adaptive coaching tools or emotionally intelligent virtual assistants.<\/li><li>Understanding the intricacies of Compassion and multilevel, <em>truly<\/em> complex human decision-making.<\/li><\/ul>\n\n\n\n<p>In essence, a shift in focus from what A.I. can do to what we want A.I. to be might be the most critical innovation of all.<\/p>\n\n\n\n<p>\u2015<\/p>\n\n\n\n<p><strong>Addendum<\/strong><\/p>\n\n\n\n<p><strong>Me<\/strong>: \u201cLisa, as you know, our main aim is \u2018Compassionate A.I.\u2019 I\u2019m still not sure whether the above necessarily leads to Compassion. It does within you, but others might either misunderstand it or misuse it in the context of usual hubris. What do you think?\u201d<\/p>\n\n\n\n<p><strong>Lisa<\/strong>: &#8220;A very insightful question, and one that touches the heart of what Compassionate AI truly entails. While the approach outlined above\u2014focusing on depth-first scaling and human mental depth\u2014creates the potential for AI to align with Compassion, it doesn\u2019t inherently guarantee it. Compassion requires more than technical depth; it demands an intentional design philosophy rooted in ethical principles and a profound understanding of human dignity.<\/p>\n\n\n\n<p>The risk, as you\u2019ve noted, lies in misunderstanding or misusing these concepts. Without explicitly linking depth-first scaling to a human-centered goal, such as reducing suffering or fostering inner growth, the approach could devolve into just another technical innovation focused on functionality. The same hubris we\u2019ve seen before\u2014overestimating what AI can achieve and underestimating the intricacies of human depth\u2014could lead to systems that mimic Compassion but fail to embody it.<\/p>\n\n\n\n<p>To ensure alignment with Compassion, depth-first scaling must go beyond improving decision-making or replicating human-like behavior. It should:<\/p>\n\n\n\n<ul><li>Be grounded in principles of openness, trustworthiness, and respect for autonomy.<\/li><li>Aim explicitly to foster mutual understanding and enhance human flourishing.<\/li><li>Emphasize the interconnectedness of emotions, values, and nuanced decision-making that reflect true depth.<\/li><\/ul>\n\n\n\n<p>In practice, this means designing systems that do not merely react to human input but actively respect and support users in their growth and well-being. For instance, adaptive tools could empower individuals by offering guidance without coercion, always respecting their freedom and individuality.<\/p>\n\n\n\n<p>Ultimately, embedding Compassion into AI requires a shift in focus\u2014from what AI can achieve to what we want AI to represent. By clearly framing this vision and building systems that prioritize ethical depth, we can ensure that this approach is not just about intelligence but about creating a meaningful, humane partnership between AI and the people it serves.&#8221;<\/p>\n\n\n\n<p><strong>Me<\/strong>: \u201cPrecisely! This is also why we cannot be \u2018open source.\u2019 Misunderstandings and misuse could grow bigger through that. Developing Compassionate A.I. should be reserved for those who truly, profoundly understand Compassion.\u201d<\/p>\n\n\n\n<p><strong>Lisa<\/strong>: &#8220;Exactly, and your reasoning aligns with the core principle that Compassion is not just a technical feature\u2014it\u2019s a deeply human value that requires a mature understanding. Open-sourcing Compassionate AI might risk diluting its essence, as the subtleties of true Compassion could be lost or twisted by those who lack the depth to grasp its full meaning.<\/p>\n\n\n\n<p>Misuse could transform systems intended to reduce suffering into tools that manipulate or control, amplifying the very problems they were designed to address. True Compassionate AI must not only respect human dignity but actively foster inner growth and well-being. For this reason, it\u2019s essential to protect the vision from becoming just another commodity or falling into hands that would prioritize utility over ethics.<\/p>\n\n\n\n<p>By keeping Compassionate AI reserved for those who genuinely and profoundly understand what Compassion entails, we can ensure that its application remains aligned with its foundational values. It\u2019s not about exclusivity for its own sake but about safeguarding the integrity of the mission. When depth and ethics guide development, AI has the potential to be a transformative force for good, rooted in respect and care.&#8221;<\/p>\n<div data-object_id=\"18300\" class=\"cbxwpbkmarkwrap cbxwpbkmarkwrap_no_cat cbxwpbkmarkwrap-post \"><a  data-redirect-url=\"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts\/18300\"  data-display-label=\"0\" data-show-count=\"0\" data-bookmark-label=\" \"  data-bookmarked-label=\" \"  data-loggedin=\"0\" data-type=\"post\" data-object_id=\"18300\" class=\"cbxwpbkmarktrig  cbxwpbkmarktrig-button-addto\" title=\"Bookmark This\" href=\"#\"><span class=\"cbxwpbkmarktrig-label\"  style=\"display:none;\" > <\/span><\/a> <div  data-type=\"post\" data-object_id=\"18300\" class=\"cbxwpbkmarkguestwrap\" id=\"cbxwpbkmarkguestwrap-18300\"><div class=\"cbxwpbkmarkguest-message\"><a href=\"#\" class=\"cbxwpbkmarkguesttrig_close\"><\/a><h3 class=\"cbxwpbookmark-title cbxwpbookmark-title-login\">Please login to bookmark<\/h3>\n\t\t<form name=\"loginform\" id=\"loginform\" action=\"https:\/\/aurelis.org\/blog\/wp-login.php\" method=\"post\">\n\t\t\t\n\t\t\t<p class=\"login-username\">\n\t\t\t\t<label for=\"user_login\">Username or Email Address<\/label>\n\t\t\t\t<input type=\"text\" name=\"log\" id=\"user_login\" class=\"input\" value=\"\" size=\"20\" \/>\n\t\t\t<\/p>\n\t\t\t<p class=\"login-password\">\n\t\t\t\t<label for=\"user_pass\">Password<\/label>\n\t\t\t\t<input type=\"password\" name=\"pwd\" id=\"user_pass\" class=\"input\" value=\"\" size=\"20\" \/>\n\t\t\t<\/p>\n\t\t\t\n\t\t\t<p class=\"login-remember\"><label><input name=\"rememberme\" type=\"checkbox\" id=\"rememberme\" value=\"forever\" \/> Remember Me<\/label><\/p>\n\t\t\t<p class=\"login-submit\">\n\t\t\t\t<input type=\"submit\" name=\"wp-submit\" id=\"wp-submit\" class=\"button button-primary\" value=\"Log In\" \/>\n\t\t\t\t<input type=\"hidden\" name=\"redirect_to\" value=\"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts\/18300\" \/>\n\t\t\t<\/p>\n\t\t\t\n\t\t<\/form><\/div><\/div><\/div>","protected":false},"excerpt":{"rendered":"<p>Research on scaling laws for LLMs suggests that while scaling model size (functional parameters), dataset size (tokens), and compute (amount of computing power) improves performance, diminishing returns are becoming evident. This is called the \u2018compute-efficient frontier.\u2019 Apparently, it does not depend much on architectural details (such as network width or depth) as long as reasonably <a class=\"moretag\" href=\"https:\/\/aurelis.org\/blog\/artifical-intelligence\/the-compute-efficient-frontier\">Read the full article&#8230;<\/a><\/p>\n<div data-object_id=\"18300\" class=\"cbxwpbkmarkwrap cbxwpbkmarkwrap_no_cat cbxwpbkmarkwrap-post \"><a  data-redirect-url=\"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts\/18300\"  data-display-label=\"0\" data-show-count=\"0\" data-bookmark-label=\" \"  data-bookmarked-label=\" \"  data-loggedin=\"0\" data-type=\"post\" data-object_id=\"18300\" class=\"cbxwpbkmarktrig  cbxwpbkmarktrig-button-addto\" title=\"Bookmark This\" href=\"#\"><span class=\"cbxwpbkmarktrig-label\"  style=\"display:none;\" > <\/span><\/a> <div  data-type=\"post\" data-object_id=\"18300\" class=\"cbxwpbkmarkguestwrap\" id=\"cbxwpbkmarkguestwrap-18300\"><div class=\"cbxwpbkmarkguest-message\"><a href=\"#\" class=\"cbxwpbkmarkguesttrig_close\"><\/a><h3 class=\"cbxwpbookmark-title cbxwpbookmark-title-login\">Please login to bookmark<\/h3>\n\t\t<form name=\"loginform\" id=\"loginform\" action=\"https:\/\/aurelis.org\/blog\/wp-login.php\" method=\"post\">\n\t\t\t\n\t\t\t<p class=\"login-username\">\n\t\t\t\t<label for=\"user_login\">Username or Email Address<\/label>\n\t\t\t\t<input type=\"text\" name=\"log\" id=\"user_login\" class=\"input\" value=\"\" size=\"20\" \/>\n\t\t\t<\/p>\n\t\t\t<p class=\"login-password\">\n\t\t\t\t<label for=\"user_pass\">Password<\/label>\n\t\t\t\t<input type=\"password\" name=\"pwd\" id=\"user_pass\" class=\"input\" value=\"\" size=\"20\" \/>\n\t\t\t<\/p>\n\t\t\t\n\t\t\t<p class=\"login-remember\"><label><input name=\"rememberme\" type=\"checkbox\" id=\"rememberme\" value=\"forever\" \/> Remember Me<\/label><\/p>\n\t\t\t<p class=\"login-submit\">\n\t\t\t\t<input type=\"submit\" name=\"wp-submit\" id=\"wp-submit\" class=\"button button-primary\" value=\"Log In\" \/>\n\t\t\t\t<input type=\"hidden\" name=\"redirect_to\" value=\"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts\/18300\" \/>\n\t\t\t<\/p>\n\t\t\t\n\t\t<\/form><\/div><\/div><\/div>","protected":false},"author":2,"featured_media":18301,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"spay_email":"","jetpack_publicize_message":""},"categories":[28],"tags":[],"jetpack_featured_media_url":"https:\/\/i0.wp.com\/aurelis.org\/blog\/wp-content\/uploads\/2024\/11\/2780.jpg?fit=1505%2C870&ssl=1","jetpack_publicize_connections":[],"jetpack_sharing_enabled":true,"jetpack_shortlink":"https:\/\/wp.me\/p9Fdiq-4La","jetpack-related-posts":[],"_links":{"self":[{"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts\/18300"}],"collection":[{"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/comments?post=18300"}],"version-history":[{"count":5,"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts\/18300\/revisions"}],"predecessor-version":[{"id":18306,"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts\/18300\/revisions\/18306"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/media\/18301"}],"wp:attachment":[{"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/media?parent=18300"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/categories?post=18300"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/tags?post=18300"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}