{"id":12526,"date":"2023-05-29T14:32:00","date_gmt":"2023-05-29T14:32:00","guid":{"rendered":"https:\/\/aurelis.org\/blog\/?p=12526"},"modified":"2023-05-31T12:29:52","modified_gmt":"2023-05-31T12:29:52","slug":"patterns-and-rewards-in-learning","status":"publish","type":"post","link":"https:\/\/aurelis.org\/blog\/artifical-intelligence\/patterns-and-rewards-in-learning","title":{"rendered":"Patterns + Rewards in A.I."},"content":{"rendered":"\n<h3>Human-inspired Pattern Recognition and Completion (PRC) may significantly heighten the efficiency of Reinforcement Learning (RL) \u2014 also in A.I.<\/h3>\n\n\n\n<p>See for PRC: <a href=\"https:\/\/aurelis.org\/blog?p=5746\">The Brain as a Predictor<\/a><\/p>\n\n\n\n<p>See for RL: <a href=\"https:\/\/aurelis.org\/blog?p=12134\">Why Reinforcement Learning is Special<\/a><\/p>\n\n\n\n<p><strong>Mutually reinforcing<\/strong><\/p>\n\n\n\n<p>PRC shows valid directions and tentatively also realizes them. RL consolidates\/reinforces the best directions and attenuates the lesser ones.<\/p>\n\n\n\n<p>Without RL, PRC may go forward pretty slowly, like a person who can learn only in small steps from what he already knows \u2014 crawling, never jumping.<\/p>\n\n\n\n<p>Without PRC, RL may go forward pretty stochastically, like a person in a dark room searching for the light switch without any guidance. PRC then provides a glow through which the person may guess probable directions at least. He may find the switch after a limited search. Next time and in a different room, he can make use of the glow even better. To an inexperienced observer, he may seem to find the switch miraculously easily.<\/p>\n\n\n\n<p><strong>RL within PRC<\/strong><\/p>\n\n\n\n<p>This may take away some confusion from the reader who is already thinking beyond the above at this point.<\/p>\n\n\n\n<p>Indeed, the recognized and completed pattern may incorporate the reward itself. It\u2019s a semantic choice. However, this shows how things can overlap, and through their overlap may lead to new possibilities.<\/p>\n\n\n\n<p><strong>The lesson from humans<\/strong><\/p>\n\n\n\n<p>The combination enables humans to learn from few examples, as even children do spontaneously. It\u2019s a significant part of the way the brain works.<\/p>\n\n\n\n<p>It can be used for the same in A.I. Through PRC and rewards, the system can learn where to evolve toward in a way similar to humans. This makes the need for many rewards and smooth rewarding less stringent, as this is a substantial bottleneck in present-day A.I.<\/p>\n\n\n\n<p>In humans, PRC is realized in a specifically human way that is intrinsically related to the human medium. Probably most crucial in this are our countless <a href=\"https:\/\/aurelis.org\/blog?p=4629\">mental-neuronal patterns<\/a>. These enable a flexible and performant kind of PRC, albeit a very fuzzy one.<\/p>\n\n\n\n<p>Probably so fuzzy that we should never indulge this in a non-Compassionate vein of super-A.I. Yet I fear that we are precisely closing into that, unfortunately.<\/p>\n\n\n\n<p><strong>At the same time, a boon for Compassion<\/strong><\/p>\n\n\n\n<p>Broadly overlapping patterns are pointers to <a href=\"https:\/\/aurelis.org\/blog?p=6819\">Compassion, basically<\/a>. They enable intra-brain intuition and our natural kind of conceptual intelligence. They also enable our natural urge for social thinking. Within us, this goes together to a large degree, making us &#8216;social animals&#8217; to whom ethics frequently means a big deal. Our intelligence has essentially been developed socially over a very long time.<\/p>\n\n\n\n<p>The same provides hope for a <a href=\"https:\/\/aurelis.org\/blog?p=2819\">Journey Towards Compassionate A.I.<\/a> if that gets based on likewise principles even when from an utterly different background. An example of this is being realized in <a href=\"https:\/\/aurelis.org\/blog\/category\/lisa\">Lisa<\/a>.<\/p>\n\n\n\n<p>We may be very optimistic if we don\u2019t blow it.<\/p>\n<div data-object_id=\"12526\" class=\"cbxwpbkmarkwrap cbxwpbkmarkwrap_no_cat cbxwpbkmarkwrap-post \"><a  data-redirect-url=\"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts\/12526\"  data-display-label=\"0\" data-show-count=\"0\" data-bookmark-label=\" \"  data-bookmarked-label=\" \"  data-loggedin=\"0\" data-type=\"post\" data-object_id=\"12526\" class=\"cbxwpbkmarktrig  cbxwpbkmarktrig-button-addto\" title=\"Bookmark This\" href=\"#\"><span class=\"cbxwpbkmarktrig-label\"  style=\"display:none;\" > <\/span><\/a> <div  data-type=\"post\" data-object_id=\"12526\" class=\"cbxwpbkmarkguestwrap\" id=\"cbxwpbkmarkguestwrap-12526\"><div class=\"cbxwpbkmarkguest-message\"><a href=\"#\" class=\"cbxwpbkmarkguesttrig_close\"><\/a><h3 class=\"cbxwpbookmark-title cbxwpbookmark-title-login\">Please login to bookmark<\/h3>\n\t\t<form name=\"loginform\" id=\"loginform\" action=\"https:\/\/aurelis.org\/blog\/wp-login.php\" method=\"post\">\n\t\t\t\n\t\t\t<p class=\"login-username\">\n\t\t\t\t<label for=\"user_login\">Username or Email Address<\/label>\n\t\t\t\t<input type=\"text\" name=\"log\" id=\"user_login\" class=\"input\" value=\"\" size=\"20\" \/>\n\t\t\t<\/p>\n\t\t\t<p class=\"login-password\">\n\t\t\t\t<label for=\"user_pass\">Password<\/label>\n\t\t\t\t<input type=\"password\" name=\"pwd\" id=\"user_pass\" class=\"input\" value=\"\" size=\"20\" \/>\n\t\t\t<\/p>\n\t\t\t\n\t\t\t<p class=\"login-remember\"><label><input name=\"rememberme\" type=\"checkbox\" id=\"rememberme\" value=\"forever\" \/> Remember Me<\/label><\/p>\n\t\t\t<p class=\"login-submit\">\n\t\t\t\t<input type=\"submit\" name=\"wp-submit\" id=\"wp-submit\" class=\"button button-primary\" value=\"Log In\" \/>\n\t\t\t\t<input type=\"hidden\" name=\"redirect_to\" value=\"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts\/12526\" \/>\n\t\t\t<\/p>\n\t\t\t\n\t\t<\/form><\/div><\/div><\/div>","protected":false},"excerpt":{"rendered":"<p>Human-inspired Pattern Recognition and Completion (PRC) may significantly heighten the efficiency of Reinforcement Learning (RL) \u2014 also in A.I. See for PRC: The Brain as a Predictor See for RL: Why Reinforcement Learning is Special Mutually reinforcing PRC shows valid directions and tentatively also realizes them. RL consolidates\/reinforces the best directions and attenuates the lesser <a class=\"moretag\" href=\"https:\/\/aurelis.org\/blog\/artifical-intelligence\/patterns-and-rewards-in-learning\">Read the full article&#8230;<\/a><\/p>\n<div data-object_id=\"12526\" class=\"cbxwpbkmarkwrap cbxwpbkmarkwrap_no_cat cbxwpbkmarkwrap-post \"><a  data-redirect-url=\"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts\/12526\"  data-display-label=\"0\" data-show-count=\"0\" data-bookmark-label=\" \"  data-bookmarked-label=\" \"  data-loggedin=\"0\" data-type=\"post\" data-object_id=\"12526\" class=\"cbxwpbkmarktrig  cbxwpbkmarktrig-button-addto\" title=\"Bookmark This\" href=\"#\"><span class=\"cbxwpbkmarktrig-label\"  style=\"display:none;\" > <\/span><\/a> <div  data-type=\"post\" data-object_id=\"12526\" class=\"cbxwpbkmarkguestwrap\" id=\"cbxwpbkmarkguestwrap-12526\"><div class=\"cbxwpbkmarkguest-message\"><a href=\"#\" class=\"cbxwpbkmarkguesttrig_close\"><\/a><h3 class=\"cbxwpbookmark-title cbxwpbookmark-title-login\">Please login to bookmark<\/h3>\n\t\t<form name=\"loginform\" id=\"loginform\" action=\"https:\/\/aurelis.org\/blog\/wp-login.php\" method=\"post\">\n\t\t\t\n\t\t\t<p class=\"login-username\">\n\t\t\t\t<label for=\"user_login\">Username or Email Address<\/label>\n\t\t\t\t<input type=\"text\" name=\"log\" id=\"user_login\" class=\"input\" value=\"\" size=\"20\" \/>\n\t\t\t<\/p>\n\t\t\t<p class=\"login-password\">\n\t\t\t\t<label for=\"user_pass\">Password<\/label>\n\t\t\t\t<input type=\"password\" name=\"pwd\" id=\"user_pass\" class=\"input\" value=\"\" size=\"20\" \/>\n\t\t\t<\/p>\n\t\t\t\n\t\t\t<p class=\"login-remember\"><label><input name=\"rememberme\" type=\"checkbox\" id=\"rememberme\" value=\"forever\" \/> Remember Me<\/label><\/p>\n\t\t\t<p class=\"login-submit\">\n\t\t\t\t<input type=\"submit\" name=\"wp-submit\" id=\"wp-submit\" class=\"button button-primary\" value=\"Log In\" \/>\n\t\t\t\t<input type=\"hidden\" name=\"redirect_to\" value=\"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts\/12526\" \/>\n\t\t\t<\/p>\n\t\t\t\n\t\t<\/form><\/div><\/div><\/div>","protected":false},"author":2,"featured_media":12554,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"spay_email":"","jetpack_publicize_message":""},"categories":[28],"tags":[],"jetpack_featured_media_url":"https:\/\/i0.wp.com\/aurelis.org\/blog\/wp-content\/uploads\/2023\/05\/2130b.jpg?fit=960%2C560&ssl=1","jetpack_publicize_connections":[],"jetpack_sharing_enabled":true,"jetpack_shortlink":"https:\/\/wp.me\/p9Fdiq-3g2","jetpack-related-posts":[],"_links":{"self":[{"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts\/12526"}],"collection":[{"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/comments?post=12526"}],"version-history":[{"count":24,"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts\/12526\/revisions"}],"predecessor-version":[{"id":12556,"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/posts\/12526\/revisions\/12556"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/media\/12554"}],"wp:attachment":[{"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/media?parent=12526"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/categories?post=12526"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/aurelis.org\/blog\/wp-json\/wp\/v2\/tags?post=12526"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}