{"id":1126,"date":"2017-11-10T15:04:42","date_gmt":"2017-11-10T20:04:42","guid":{"rendered":"http:\/\/www.jsylvest.com\/blog\/?p=1126"},"modified":"2017-11-22T15:24:59","modified_gmt":"2017-11-22T20:24:59","slug":"ais-one-trick-pony-has-a-hell-of-a-trick","status":"publish","type":"post","link":"https:\/\/www.jsylvest.com\/blog\/2017\/11\/ais-one-trick-pony-has-a-hell-of-a-trick\/","title":{"rendered":"AI's \"one trick pony\" has a hell of a trick"},"content":{"rendered":"<p>The MIT Technology Review has a recent article by James Somers about error backpropagation, \"<a href=\"https:\/\/www.technologyreview.com\/s\/608911\/is-ai-riding-a-one-trick-pony\/?utm_campaign=Data%2BElixir&amp;utm_medium=email&amp;utm_source=Data_Elixir_151\">Is AI Riding a One-Trick Pony?<\/a>\" Overall, I agree with the message in the article. We need to keep thinking of new paradigms because the SotA right now is very useful, but not <em>correct<\/em> in any rigorous way. However, as much as I agree with the thesis, I think\u00a0Somers oversells it, especially in the beginning of the piece. For instance, the introductory segment concludes:<\/p>\n<blockquote><p>When you boil it down, AI today is deep learning, and deep learning is backprop \u2014 which is amazing, considering that backprop is more than 30 years old. It\u2019s worth understanding how that happened\u2014how a technique could lie in wait for so long and then cause such an explosion \u2014 because once you understand the story of backprop, you\u2019ll start to understand the current moment in AI, and in particular the fact that maybe we\u2019re not actually at the beginning of a revolution. Maybe we\u2019re at the end of one.<\/p><\/blockquote>\n<p>That's a bit like saying \"When you boil it down, flight is airfoils, and airfoils are Bernoulli's principle \u2014 which is amazing, considering that Bernoulli's principle is almost 300 years old.\" I totally endorse the idea that we ought to understand backprop; I've spent a lot of effort in the last couple of months organizing training for some of my firm's senior leadership on neural networks, and EBP\/gradient descent is the heart of my presentation. But I would be very, very careful about concluding that backprop is the entire show.<\/p>\n<p>Backprop was also not \"lying in wait.\" People were working on it since it was introduced in 1986. The problem was that '86 was the height of the 2nd AI winter, which lasted another decade. Just like people should understand backprop to understand contemporary AI, they should learn about the history of AI to understand contemporary AI. Just because no one outside of CS (and precious few people in CS, for that matter) paid any attention to neural networks before 2015 doesn't mean they were completely dormant,\u00a0only to spring up fully formed in some sort of intellectual Athenian birth.<\/p>\n<p>I really don't want to be in the position of defending backprop. I took the trouble to write a dissertation about non-backprop neural nets for a reason, after all. ((That reason being, roughly put, that we're pretty sure the brain is not using backprop, and it seems ill-advised to ignore the mechanisms employed by the most intelligent thing we are aware of.)) But I also don't want to be in the position of letting sloppy arguments against neural nets go unremarked. That road leads to people mischaracterizing Minksy and Papert, abandoning neural nets for generations, and putting us epochs behind where we might have been. ((Plus sloppy arguments should be eschewed on the basis of the sloppiness alone, irrespective of their consequences.))<\/p>\n<hr \/>\n<p><em>PS<\/em> This is also worth a rejoinder:<\/p>\n<blockquote><p>Big patterns of neural activity, if you\u2019re a mathematician, can be captured in a vector space, with each neuron\u2019s activity corresponding to a number, and each number to a coordinate of a really big vector. In Hinton\u2019s view, that\u2019s what thought is: a dance of vectors.<\/p><\/blockquote>\n<p>That's not what thought <em>is<\/em>, that's how thought can be <em>represented<\/em>. Planets are not vectors, but their orbits can be profitably described that way, because \"it behooves us to place the foundations of knowledge in mathematics.\" I'm sorry if that seems pedantic, but the distinction between a thing and its representation\u2014besides giving semioticians something to talk about\u2014underpins much of our interpretation of AI systems and cognitive science as well. Indeed, a huge chunk of data science work is figuring out the right representations. If you can get that, your problem is often largely solved. ((IIRC both Knuth and Torvalds have aphorisms to the effect that once you have chosen the correct data structures, the correct algorithms will naturally follow. I think AI and neuroscience are dealing with a lot of friction because we haven't been able to figure out the right representations\/data structures. When we do, the right learning algorithms will follow much more easily.))<\/p>\n<p><em>PPS<\/em> This, on the other hand, I agree with entirely:<\/p>\n<blockquote><p>Deep learning in some ways mimics what goes on in the human brain, but only in a shallow way. \u2026 What we know about intelligence is nothing against the vastness of what we still don\u2019t know.<\/p><\/blockquote>\n<p>What I fear is that people read that and conclude that artificial neural networks are built on a shallow foundation, so we should give up on them as being unreliable. A much better conclusion would be that we need to keep working and build better, deeper foundations.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>The MIT Technology Review has a recent article by James Somers about error backpropagation, \"Is AI Riding a One-Trick Pony?\" Overall, I agree with the message in the article. We need to keep thinking of new paradigms because the SotA &hellip; <a href=\"https:\/\/www.jsylvest.com\/blog\/2017\/11\/ais-one-trick-pony-has-a-hell-of-a-trick\/\">Continue reading <span class=\"meta-nav\">&rarr;<\/span><\/a><\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"jetpack_post_was_ever_published":false,"_jetpack_newsletter_access":"","_jetpack_dont_email_post_to_subs":false,"_jetpack_newsletter_tier_id":0,"_jetpack_memberships_contains_paywalled_content":false,"_jetpack_memberships_contains_paid_content":false,"footnotes":"","jetpack_publicize_message":"","jetpack_publicize_feature_enabled":true,"jetpack_social_post_already_shared":true,"jetpack_social_options":{"image_generator_settings":{"template":"highway","default_image_id":0,"font":"","enabled":false},"version":2}},"categories":[10],"tags":[24,31,3,39,34,40,38,15],"class_list":["post-1126","post","type-post","status-publish","format-standard","hentry","category-cs","tag-ai","tag-cogsci","tag-computer-science","tag-deep-learning","tag-machine-learning","tag-ml","tag-neural-nets","tag-technology","wpautop"],"jetpack_publicize_connections":[],"jetpack_featured_media_url":"","jetpack_sharing_enabled":true,"jetpack_shortlink":"https:\/\/wp.me\/p3sddF-ia","jetpack-related-posts":[{"id":1043,"url":"https:\/\/www.jsylvest.com\/blog\/2017\/04\/will-ai-steal-our-jobs\/","url_meta":{"origin":1126,"position":0},"title":"Will AI steal our jobs?","author":"jsylvest","date":"5 April 2017","format":false,"excerpt":"As an AI researcher, I think I am required to have an opinion about this. Here's what I have to say to the various tribes. AI-pessimists: please remember that the Luddites have been wrong about technology causing\u00a0economic cataclysm\u00a0every time so far. We're talking about several consecutive centuries of wrongness. ((I\u2026","rel":"","context":"In &quot;Business \/ Economics&quot;","block_context":{"text":"Business \/ Economics","link":"https:\/\/www.jsylvest.com\/blog\/category\/business-2\/"},"img":{"alt_text":"","src":"","width":0,"height":0},"classes":[]},{"id":1098,"url":"https:\/\/www.jsylvest.com\/blog\/2018\/07\/hume-on-knowledge\/","url_meta":{"origin":1126,"position":1},"title":"Hume on Knowledge","author":"jsylvest","date":"1 July 2018","format":"aside","excerpt":"All knowledge degenerates into probability. \u2014 David Hume, \"A Treatise on Human Nature,\" \u00a7IV.1","rel":"","context":"In &quot;Quotes&quot;","block_context":{"text":"Quotes","link":"https:\/\/www.jsylvest.com\/blog\/category\/quotes\/"},"img":{"alt_text":"","src":"","width":0,"height":0},"classes":[]},{"id":1144,"url":"https:\/\/www.jsylvest.com\/blog\/2017\/12\/malconv\/","url_meta":{"origin":1126,"position":2},"title":"MalConv: Lessons learned from Deep Learning on executables","author":"jsylvest","date":"8 December 2017","format":false,"excerpt":"I don't usually write up my technical work here, mostly because I spend enough hours as is doing technical writing. But a co-author, Jon Barker, recently wrote\u00a0a post on the NVIDIA Parallel For All blog about one of our papers on neural networks for detecting malware, so I thought I'd\u2026","rel":"","context":"In &quot;CS \/ Science \/ Tech \/ Coding&quot;","block_context":{"text":"CS \/ Science \/ Tech \/ Coding","link":"https:\/\/www.jsylvest.com\/blog\/category\/cs\/"},"img":{"alt_text":"The MalConv architecture","src":"https:\/\/i0.wp.com\/www.jsylvest.com\/blog\/wp-content\/uploads\/2017\/12\/malconv.png?resize=350%2C200","width":350,"height":200,"srcset":"https:\/\/i0.wp.com\/www.jsylvest.com\/blog\/wp-content\/uploads\/2017\/12\/malconv.png?resize=350%2C200 1x, https:\/\/i0.wp.com\/www.jsylvest.com\/blog\/wp-content\/uploads\/2017\/12\/malconv.png?resize=525%2C300 1.5x"},"classes":[]},{"id":1124,"url":"https:\/\/www.jsylvest.com\/blog\/2017\/10\/national-ai-strategy\/","url_meta":{"origin":1126,"position":3},"title":"National AI Strategy","author":"jsylvest","date":"9 October 2017","format":false,"excerpt":"Some of my co-workers published a sponsored piece in the Atlantic calling for a national AI strategy,\u00a0which was tied in to\u00a0some discussions at the\u00a0Washington Ideas event. I'm 100% on board with the US having a strategy, but I want to offer one caveat: \"comprehensive national strategies\" are susceptible to becoming\u2026","rel":"","context":"In &quot;Business \/ Economics&quot;","block_context":{"text":"Business \/ Economics","link":"https:\/\/www.jsylvest.com\/blog\/category\/business-2\/"},"img":{"alt_text":"","src":"","width":0,"height":0},"classes":[]},{"id":1188,"url":"https:\/\/www.jsylvest.com\/blog\/2018\/02\/aies-2018\/","url_meta":{"origin":1126,"position":4},"title":"AIES 2018","author":"jsylvest","date":"9 February 2018","format":false,"excerpt":"Last week I attended the first annual conference on AI, Ethics & Society where I presented some work on a Decision Tree\/Random Forest algorithm that makes decisions that are less biased or discriminatory. ((In the colloquial rather than technical sense)) You can read all the juicy details in our paper.\u2026","rel":"","context":"In &quot;CS \/ Science \/ Tech \/ Coding&quot;","block_context":{"text":"CS \/ Science \/ Tech \/ Coding","link":"https:\/\/www.jsylvest.com\/blog\/category\/cs\/"},"img":{"alt_text":"","src":"","width":0,"height":0},"classes":[]},{"id":403,"url":"https:\/\/www.jsylvest.com\/blog\/2013\/05\/pi\/","url_meta":{"origin":1126,"position":5},"title":"Pi","author":"jsylvest","date":"9 May 2013","format":false,"excerpt":"The Economist :: Babbage Blog :: Humble Pi The Raspberry Pi is the brainchild of a couple of computer scientists at Cambridge University. Back in 2006, they lamented the decline in programming skills among applicants for computer-science courses. ... Over the past ten years, computer-science students have gone from arriving\u2026","rel":"","context":"In &quot;CS \/ Science \/ Tech \/ Coding&quot;","block_context":{"text":"CS \/ Science \/ Tech \/ Coding","link":"https:\/\/www.jsylvest.com\/blog\/category\/cs\/"},"img":{"alt_text":"","src":"","width":0,"height":0},"classes":[]}],"_links":{"self":[{"href":"https:\/\/www.jsylvest.com\/blog\/wp-json\/wp\/v2\/posts\/1126","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.jsylvest.com\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.jsylvest.com\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.jsylvest.com\/blog\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.jsylvest.com\/blog\/wp-json\/wp\/v2\/comments?post=1126"}],"version-history":[{"count":7,"href":"https:\/\/www.jsylvest.com\/blog\/wp-json\/wp\/v2\/posts\/1126\/revisions"}],"predecessor-version":[{"id":1141,"href":"https:\/\/www.jsylvest.com\/blog\/wp-json\/wp\/v2\/posts\/1126\/revisions\/1141"}],"wp:attachment":[{"href":"https:\/\/www.jsylvest.com\/blog\/wp-json\/wp\/v2\/media?parent=1126"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.jsylvest.com\/blog\/wp-json\/wp\/v2\/categories?post=1126"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.jsylvest.com\/blog\/wp-json\/wp\/v2\/tags?post=1126"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}