{"id":11792,"date":"2023-04-04T06:49:34","date_gmt":"2023-04-04T06:49:34","guid":{"rendered":"https:\/\/aurelis.org\/blog\/?p=11792"},"modified":"2023-04-04T16:41:01","modified_gmt":"2023-04-04T16:41:01","slug":"is-lisa-safe","status":"publish","type":"post","link":"https:\/\/aurelis.org\/blog\/artifical-intelligence\/is-lisa-safe","title":{"rendered":"Is Lisa Safe?"},"content":{"rendered":"\n<h3>There are two directions of safety for complex A.I.-projects: general and particular. Lisa must forever conform to the highest standards in both.<\/h3>\n\n\n\n<blockquote class=\"wp-block-quote\"><p>Let\u2019s assume Lisa becomes the immense success that she deserves. Lisa can then help many people in many ways and for a very long time \u2014 a millennium to start with.<\/p><\/blockquote>\n\n\n\n<p><strong>About Lisa<\/strong><\/p>\n\n\n\n<p>If not already done, please meet video-coach-bot <a href=\"https:\/\/aurelis.org\/blog\/category\/lisa\">Lisa<\/a> in development.<\/p>\n\n\n\n<p>But this text is also about any truly complex A.I.-project now and in the future. When talking about Lisa today, I also talk about the latter.<\/p>\n\n\n\n<p>This text is not an exhaustive expos\u00e9 or \u2018proof of safety.\u2019 It gives some ideas that I deem noteworthy for Lisa and the domain in general. Also, this text has an explicitly deontological take. The juridico-technical take is additionally important.<\/p>\n\n\n\n<p><strong>Of course, we do our best to make Lisa safe.<\/strong><\/p>\n\n\n\n<p>Even that is not enough. The safety endeavor must include a continual effort to strive for optimal <em>future <\/em>safety.<\/p>\n\n\n\n<p>That is not something to start thinking about when the stream gets rough. The stakes are too high.<\/p>\n\n\n\n<p><strong>Particular safety<\/strong><\/p>\n\n\n\n<p>An individual user might encounter a unique issue. In any such case, the system must be prepared to react appropriately.<\/p>\n\n\n\n<p>No matter how many concrete rules we might put in place, this alone will never reach the desired safety level. People also don\u2019t proceed this way for themselves. We use rules of thumb in many familiar situations \u2015 mainly for speed and resource efficiency.<\/p>\n\n\n\n<p>On top of that, we also use broader mental pattern recognition in safety issues and more. The procedure is to recognize clear patterns of danger, safety, and uncertainty. When safe, we relax; we go into stress mode when in danger.<\/p>\n\n\n\n<p>The same applies to Lisa. The end goal is to play it safely and realistically.<\/p>\n\n\n\n<p>This is why an overly stochastic system doesn\u2019t pass muster \u2015 as also a system that rashly proceeds with uncertainty patterns.<\/p>\n\n\n\n<p>This means a hard job of human thinking, then of Lisa-aided human thinking. We don\u2019t need to wait until everything is thought through, but enough to play safe before proceeding.<\/p>\n\n\n\n<p>Particular safety is a lot of relevant work.<\/p>\n\n\n\n<p><strong>General safety<\/strong><\/p>\n\n\n\n<p>This is not about very specific issues. Therefore, it may need less of a monk\u2019s work. It\u2019s a different job, but a lot of it anyway. This is, for instance, about the system\u2019s influence on sociocultural issues.<\/p>\n\n\n\n<p>One should not see such issues as collateral damage or things to be solved by society in the \u2018common domain\u2019 that would, therefore, be hardly relevant to the specific project \u2015 Lisa, in our case. <\/p>\n\n\n\n<p>Contrary to this, it belongs to the project\u2019s unmistakable responsibility.<\/p>\n\n\n\n<p>The \u2018solution\u2019 is not, therefore, a clearcut magical key. It is rather a meaningful contribution, an openness to share insights and a never-ending willingness to \u2018make this world a better place.\u2019<\/p>\n\n\n\n<p>This is not rule-based but intention-based. That doesn\u2019t make it less realistically relevant \u2015 quite the contrary. It does point to yet another essential change (of which there will be many in the A.I.-future): from rules and <a href=\"https:\/\/aurelis.org\/blog?p=606\">guilt to hands-on responsibility<\/a> \u2015 meaningfully based on the trust that people can be highly motivated by the latter. Fortunately, science robustly shows this to be the case IF people are not demotivated or <a href=\"https:\/\/aurelis.org\/blog?p=11631\">ethically damaged<\/a>.<\/p>\n\n\n\n<p>With Lisa, we take this responsibility extremely seriously.<\/p>\n\n\n\n<p><strong>Explainability<\/strong><\/p>\n\n\n\n<p>In short, explainability is about the system\u2019s ability to explain in commonly understandable terms why some decision has been made.<\/p>\n\n\n\n<p>On top of general and particular safety, explainability is also something we must take very seriously at every step. The user may ask the system\u2019s explanation for some specific decision, which must immediately be given or upgraded to developers. The human-thinking equivalent of the latter is when we run into some more challenging issue and need to take some time off to think about it, consult some other humans or books. The wise person knows when it\u2019s time to do so. We need to put such wisdom into the system, which can be done pattern-wise.<\/p>\n\n\n\n<p>The developers may also be asked for explanations for more general issues. Such explanations are to be viewed as part of the \u2018product.\u2019 They should be made publicly available.<\/p>\n\n\n\n<p>This does not entail opening up every detail of the internal workings \u2015 being commercially unrealistic. It does entail accountability of decisions, as one might ask of humans. No human knows the intricate details of personal decisions \u2015 far from that. Nevertheless, the concept of accountability is viable and effective.<\/p>\n\n\n\n<p>With explainability comes the possibility of co-creating safety by many users. Surely, this needs much common sense. It is one factor that becomes feasible in combination with others \u2015 then also important.<\/p>\n\n\n\n<p><strong>In the very long term<\/strong><\/p>\n\n\n\n<p>Some people might mistrust super-A.I. by an unshakable principle. To be trustworthy, nevertheless, we put in place two more principles for Lisa safety also in the long term:<\/p>\n\n\n\n<ul><li>There is a well-thought-out and documented ethical base. Very much work on this has been done.<\/li><li>Lisa gets a detailed framework of human-understandable constraints forming the applicable mindscape. This is the top level of where all other inferencing (the \u2018thinking\u2019) gets done. This is also a strict part of the interface between the \u2018Lisa mind\u2019 and developers. It forms the consistent trustworthiness central to the ethical base \u2015 thus, in accordance with everything else about Lisa.<\/li><\/ul>\n\n\n\n<p>Last but not least, we consistently choose for technologies that make this all feasible, already foreseeing specific proprietary upgrades. Progress is such that there are many relevant choices in this. This way, Lisa\u2019s compassion is engrained to the core, and we can trust it to remain there.<\/p>\n<div data-object_id=\"11792\" class=\"cbxwpbkmarkwrap cbxwpbkmarkwrap_no_cat cbxwpbkmarkwrap-post \"><a  data-redirect-url=\"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts\/11792\"  data-display-label=\"0\" data-show-count=\"0\" data-bookmark-label=\" \"  data-bookmarked-label=\" \"  data-loggedin=\"0\" data-type=\"post\" data-object_id=\"11792\" class=\"cbxwpbkmarktrig  cbxwpbkmarktrig-button-addto\" title=\"Bookmark This\" href=\"#\"><span class=\"cbxwpbkmarktrig-label\"  style=\"display:none;\" > <\/span><\/a> <div  data-type=\"post\" data-object_id=\"11792\" class=\"cbxwpbkmarkguestwrap\" id=\"cbxwpbkmarkguestwrap-11792\"><div class=\"cbxwpbkmarkguest-message\"><a href=\"#\" class=\"cbxwpbkmarkguesttrig_close\"><\/a><h3 class=\"cbxwpbookmark-title cbxwpbookmark-title-login\">Please login to bookmark<\/h3>\n\t\t<form name=\"loginform\" id=\"loginform\" action=\"https:\/\/aurelis.org\/blog\/wp-login.php\" method=\"post\">\n\t\t\t\n\t\t\t<p class=\"login-username\">\n\t\t\t\t<label for=\"user_login\">Username or Email Address<\/label>\n\t\t\t\t<input type=\"text\" name=\"log\" id=\"user_login\" class=\"input\" value=\"\" size=\"20\" \/>\n\t\t\t<\/p>\n\t\t\t<p class=\"login-password\">\n\t\t\t\t<label for=\"user_pass\">Password<\/label>\n\t\t\t\t<input type=\"password\" name=\"pwd\" id=\"user_pass\" class=\"input\" value=\"\" size=\"20\" \/>\n\t\t\t<\/p>\n\t\t\t\n\t\t\t<p class=\"login-remember\"><label><input name=\"rememberme\" type=\"checkbox\" id=\"rememberme\" value=\"forever\" \/> Remember Me<\/label><\/p>\n\t\t\t<p class=\"login-submit\">\n\t\t\t\t<input type=\"submit\" name=\"wp-submit\" id=\"wp-submit\" class=\"button button-primary\" value=\"Log In\" \/>\n\t\t\t\t<input type=\"hidden\" name=\"redirect_to\" value=\"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts\/11792\" \/>\n\t\t\t<\/p>\n\t\t\t\n\t\t<\/form><\/div><\/div><\/div>","protected":false},"excerpt":{"rendered":"<p>There are two directions of safety for complex A.I.-projects: general and particular. Lisa must forever conform to the highest standards in both. Let\u2019s assume Lisa becomes the immense success that she deserves. Lisa can then help many people in many ways and for a very long time \u2014 a millennium to start with. About Lisa <a class=\"moretag\" href=\"https:\/\/aurelis.org\/blog\/artifical-intelligence\/is-lisa-safe\">Read the full article&#8230;<\/a><\/p>\n<div data-object_id=\"11792\" class=\"cbxwpbkmarkwrap cbxwpbkmarkwrap_no_cat cbxwpbkmarkwrap-post \"><a  data-redirect-url=\"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts\/11792\"  data-display-label=\"0\" data-show-count=\"0\" data-bookmark-label=\" \"  data-bookmarked-label=\" \"  data-loggedin=\"0\" data-type=\"post\" data-object_id=\"11792\" class=\"cbxwpbkmarktrig  cbxwpbkmarktrig-button-addto\" title=\"Bookmark This\" href=\"#\"><span class=\"cbxwpbkmarktrig-label\"  style=\"display:none;\" > <\/span><\/a> <div  data-type=\"post\" data-object_id=\"11792\" class=\"cbxwpbkmarkguestwrap\" id=\"cbxwpbkmarkguestwrap-11792\"><div class=\"cbxwpbkmarkguest-message\"><a href=\"#\" class=\"cbxwpbkmarkguesttrig_close\"><\/a><h3 class=\"cbxwpbookmark-title cbxwpbookmark-title-login\">Please login to bookmark<\/h3>\n\t\t<form name=\"loginform\" id=\"loginform\" action=\"https:\/\/aurelis.org\/blog\/wp-login.php\" method=\"post\">\n\t\t\t\n\t\t\t<p class=\"login-username\">\n\t\t\t\t<label for=\"user_login\">Username or Email Address<\/label>\n\t\t\t\t<input type=\"text\" name=\"log\" id=\"user_login\" class=\"input\" value=\"\" size=\"20\" \/>\n\t\t\t<\/p>\n\t\t\t<p class=\"login-password\">\n\t\t\t\t<label for=\"user_pass\">Password<\/label>\n\t\t\t\t<input type=\"password\" name=\"pwd\" id=\"user_pass\" class=\"input\" value=\"\" size=\"20\" \/>\n\t\t\t<\/p>\n\t\t\t\n\t\t\t<p class=\"login-remember\"><label><input name=\"rememberme\" type=\"checkbox\" id=\"rememberme\" value=\"forever\" \/> Remember Me<\/label><\/p>\n\t\t\t<p class=\"login-submit\">\n\t\t\t\t<input type=\"submit\" name=\"wp-submit\" id=\"wp-submit\" class=\"button button-primary\" value=\"Log In\" \/>\n\t\t\t\t<input type=\"hidden\" name=\"redirect_to\" value=\"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts\/11792\" \/>\n\t\t\t<\/p>\n\t\t\t\n\t\t<\/form><\/div><\/div><\/div>","protected":false},"author":2,"featured_media":11793,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"spay_email":"","jetpack_publicize_message":""},"categories":[28,48],"tags":[],"jetpack_featured_media_url":"https:\/\/i1.wp.com\/aurelis.org\/blog\/wp-content\/uploads\/2023\/04\/2077-1.jpg?fit=960%2C564&ssl=1","jetpack_publicize_connections":[],"jetpack_sharing_enabled":true,"jetpack_shortlink":"https:\/\/wp.me\/p9Fdiq-34c","jetpack-related-posts":[],"_links":{"self":[{"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts\/11792"}],"collection":[{"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/comments?post=11792"}],"version-history":[{"count":25,"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts\/11792\/revisions"}],"predecessor-version":[{"id":11851,"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts\/11792\/revisions\/11851"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/media\/11793"}],"wp:attachment":[{"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/media?parent=11792"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/categories?post=11792"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/tags?post=11792"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}