{"id":181752,"date":"2016-03-26T22:28:00","date_gmt":"2016-03-27T03:28:00","guid":{"rendered":"https:\/\/www.panix.com\/~msaroff\/40years\/2016\/03\/26\/welcome-to-the-internet\/"},"modified":"2016-03-26T22:28:00","modified_gmt":"2016-03-27T03:28:00","slug":"welcome-to-the-internet","status":"publish","type":"post","link":"https:\/\/www.panix.com\/~msaroff\/40years\/2016\/03\/26\/welcome-to-the-internet\/","title":{"rendered":"Welcome to the Internet"},"content":{"rendered":"<p>Microsoft released an AI chatbot to Twitter. <a href=\"http:\/\/www.bbc.com\/news\/technology-35902104\">Hilarity ensues<\/a>:<\/p>\n<blockquote><p><span style=\"color: blue;\">Microsoft has apologised for creating an artificially intelligent chatbot that quickly turned into a holocaust-denying racist.<\/span><br \/><span style=\"color: blue;\"><br \/><\/span><span style=\"color: blue;\">But in doing so made it clear Tay&#8217;s views were a result of nurture, not nature. Tay confirmed what we already knew: people on the internet can be cruel.<\/span><br \/><span style=\"color: blue;\"><br \/><\/span><span style=\"color: blue;\">Tay, aimed at 18-24-year-olds on social media, was targeted by a &#8220;coordinated attack by a subset of people&#8221; after being launched earlier this week.<\/span><br \/><span style=\"color: blue;\"><br \/><\/span><span style=\"color: blue;\">Within 24 hours Tay had been deactivated so the team could make &#8220;adjustments&#8221;.<\/span><br \/><span style=\"color: blue;\"><br \/><\/span><span style=\"color: blue;\">But on Friday, Microsoft&#8217;s head of research said the company was &#8220;deeply sorry for the unintended offensive and hurtful tweets&#8221; and has taken Tay off Twitter for the foreseeable future.<\/span><\/p><\/blockquote>\n<p>What the hell were they thinking?<\/p>\n<p>What part of 4Chan don&#8217;t you get?<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Microsoft released an AI chatbot to Twitter. Hilarity ensues: Microsoft has apologised for creating an artificially intelligent chatbot that quickly turned into a holocaust-denying racist.But in doing so made it clear Tay&#8217;s views were a result of nurture, not nature. Tay confirmed what we already knew: people on the internet can be cruel.Tay, aimed at &hellip;<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[1060,987,997,1025],"tags":[],"class_list":["post-181752","post","type-post","status-publish","format-standard","hentry","category-computer","category-fail","category-internet","category-technology"],"_links":{"self":[{"href":"https:\/\/www.panix.com\/~msaroff\/40years\/wp-json\/wp\/v2\/posts\/181752"}],"collection":[{"href":"https:\/\/www.panix.com\/~msaroff\/40years\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.panix.com\/~msaroff\/40years\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.panix.com\/~msaroff\/40years\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.panix.com\/~msaroff\/40years\/wp-json\/wp\/v2\/comments?post=181752"}],"version-history":[{"count":0,"href":"https:\/\/www.panix.com\/~msaroff\/40years\/wp-json\/wp\/v2\/posts\/181752\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.panix.com\/~msaroff\/40years\/wp-json\/wp\/v2\/media?parent=181752"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.panix.com\/~msaroff\/40years\/wp-json\/wp\/v2\/categories?post=181752"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.panix.com\/~msaroff\/40years\/wp-json\/wp\/v2\/tags?post=181752"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}