{"id":7564,"date":"2016-02-07T07:47:52","date_gmt":"2016-02-07T07:47:52","guid":{"rendered":"http:\/\/revoscience.com\/en\/?p=7564"},"modified":"2016-02-07T07:48:23","modified_gmt":"2016-02-07T07:48:23","slug":"energy-friendly-chip-can-perform-powerful-artificial-intelligence-tasks","status":"publish","type":"post","link":"https:\/\/www.revoscience.com\/en\/energy-friendly-chip-can-perform-powerful-artificial-intelligence-tasks\/","title":{"rendered":"Energy-friendly chip can perform powerful artificial-intelligence tasks"},"content":{"rendered":"<p style=\"text-align: justify;\"><span style=\"color: #000000;\"><em><strong style=\"color: #222222;\">Advance could enable mobile devices to implement \u201cneural networks\u201d modeled on the human brain.<\/strong><\/em><\/span><\/p>\n<figure id=\"attachment_7566\" aria-describedby=\"caption-attachment-7566\" style=\"width: 617px\" class=\"wp-caption alignnone\"><a href=\"http:\/\/revoscience.com\/en\/wp-content\/uploads\/2016\/02\/MIT-Neural-Chip_0.jpg\" target=\"_blank\" rel=\"noopener\"><img loading=\"lazy\" decoding=\"async\" class=\" wp-image-7566\" src=\"http:\/\/revoscience.com\/en\/wp-content\/uploads\/2016\/02\/MIT-Neural-Chip_0.jpg\" alt=\"MIT researchers have designed a new chip to implement neural networks. It is 10 times as efficient as a mobile GPU, so it could enable mobile devices to run powerful artificial-intelligence algorithms locally, rather than uploading data to the Internet for processing. Image: MIT News\" width=\"617\" height=\"412\" title=\"\" srcset=\"https:\/\/www.revoscience.com\/en\/wp-content\/uploads\/2016\/02\/MIT-Neural-Chip_0.jpg 448w, https:\/\/www.revoscience.com\/en\/wp-content\/uploads\/2016\/02\/MIT-Neural-Chip_0-300x200.jpg 300w\" sizes=\"auto, (max-width: 617px) 100vw, 617px\" \/><\/a><figcaption id=\"caption-attachment-7566\" class=\"wp-caption-text\">MIT researchers have designed a new chip to implement neural networks. It is 10 times as efficient as a mobile GPU, so it could enable mobile devices to run powerful artificial-intelligence algorithms locally, rather than uploading data to the Internet for processing.<br \/>Image: MIT News<\/figcaption><\/figure>\n<p style=\"text-align: justify;\"><span style=\"color: #000000;\"><strong>CAMBRIDGE, Mass<\/strong>. &#8212;\u00a0In recent years, some of the most exciting advances in artificial intelligence have come courtesy of convolutional neural networks, large virtual networks of simple information-processing units, which are loosely modeled on the anatomy of the human brain.<\/span><\/p>\n<p style=\"text-align: justify;\"><span style=\"color: #000000;\">Neural networks are typically implemented using graphics processing units (GPUs), special-purpose graphics chips found in all computing devices with screens. A mobile GPU, of the type found in a cell phone, might have almost 200 cores, or processing units, making it well suited to simulating a network of distributed processors.<\/span><\/p>\n<p style=\"text-align: justify;\"><span style=\"color: #000000;\">At the International Solid State Circuits Conference in San Francisco, MIT researchers presented a new chip designed specifically to implement neural networks. It is 10 times as efficient as a mobile GPU, so it could enable mobile devices to run powerful artificial-intelligence algorithms locally, rather than uploading data to the Internet for processing.<\/span><\/p>\n<p style=\"text-align: justify;\"><span style=\"color: #000000;\">Neural nets were widely studied in the early days of artificial-intelligence research, but by the 1970s, they\u2019d fallen out of favor. In the past decade, however, they\u2019ve enjoyed a revival, under the name \u201cdeep learning.\u201d<\/span><\/p>\n<p style=\"text-align: justify;\">[pullquote]The key to Eyeriss\u2019s efficiency is to minimize the frequency with which cores need to exchange data with distant memory banks, an operation that consumes a good deal of time and energy.[\/pullquote]<\/p>\n<p style=\"text-align: justify;\"><span style=\"color: #000000;\">\u201cDeep learning is useful for many applications, such as object recognition, speech, face detection,\u201d says Vivienne Sze, an assistant professor of electrical engineering at MIT whose group developed the new chip. \u201cRight now, the networks are pretty complex and are mostly run on high-power GPUs. You can imagine that if you can bring that functionality to your cell phone or embedded devices, you could still operate even if you don\u2019t have a Wi-Fi connection. You might also want to process locally for privacy reasons. Processing it on your phone also avoids any transmission latency, so that you can react much faster for certain applications.\u201d<\/span><\/p>\n<p style=\"text-align: justify;\"><span style=\"color: #000000;\">The new chip, which the researchers dubbed \u201cEyeriss,\u201d could also help usher in the \u201cInternet of things\u201d \u2014 the idea that vehicles, appliances, civil-engineering structures, manufacturing equipment, and even livestock would have sensors that report information directly to networked servers, aiding with maintenance and task coordination. With powerful artificial-intelligence algorithms on board, networked devices could make important decisions locally, entrusting only their conclusions, rather than raw personal data, to the Internet. And, of course, onboard neural networks would be useful to battery-powered autonomous robots.<\/span><\/p>\n<p style=\"text-align: justify;\"><span style=\"color: #000000;\"><strong>Division of labor<\/strong><\/span><\/p>\n<p style=\"text-align: justify;\"><span style=\"color: #000000;\">A neural network is typically organized into layers, and each layer contains a large number of processing nodes. Data come in and are divided up among the nodes in the bottom layer. Each node manipulates the data it receives and passes the results on to nodes in the next layer, which manipulate the data they receive and pass on the results, and so on. The output of the final layer yields the solution to some computational problem.<\/span><\/p>\n<p style=\"text-align: justify;\"><span style=\"color: #000000;\">In a convolutional neural net, many nodes in each layer process the same data in different ways. The networks can thus swell to enormous proportions. Although they outperform more conventional algorithms on many visual-processing tasks, they require much greater computational resources.<\/span><\/p>\n<p style=\"text-align: justify;\"><span style=\"color: #000000;\">The particular manipulations performed by each node in a neural net are the result of a training process, in which the network tries to find correlations between raw data and labels applied to it by human annotators. With a chip like the one developed by the MIT researchers, a trained network could simply be exported to a mobile device.<\/span><\/p>\n<p style=\"text-align: justify;\"><span style=\"color: #000000;\">This application imposes design constraints on the researchers. On one hand, the way to lower the chip\u2019s power consumption and increase its efficiency is to make each processing unit as simple as possible; on the other hand, the chip has to be flexible enough to implement different types of networks tailored to different tasks.<\/span><\/p>\n<p style=\"text-align: justify;\"><span style=\"color: #000000;\">Sze and her colleagues \u2014 Yu-Hsin Chen, a graduate student in electrical engineering and computer science and first author on the conference paper; Joel Emer, a professor of the practice in MIT\u2019s Department of Electrical Engineering and Computer Science, and a senior distinguished research scientist at the chip manufacturer NVidia, and, with Sze, one of the project\u2019s two principal investigators; and Tushar Krishna, who was a postdoc with the Singapore-MIT Alliance for Research and Technology when the work was done and is now an assistant professor of computer and electrical engineering at Georgia Tech \u2014 settled on a chip with 168 cores, roughly as many as a mobile GPU has.<\/span><\/p>\n<p style=\"text-align: justify;\"><span style=\"color: #000000;\"><strong>Act locally<\/strong><\/span><\/p>\n<p style=\"text-align: justify;\"><span style=\"color: #000000;\">The key to Eyeriss\u2019s efficiency is to minimize the frequency with which cores need to exchange data with distant memory banks, an operation that consumes a good deal of time and energy. Whereas many of the cores in a GPU share a single, large memory bank, each of the Eyeriss cores has its own memory. Moreover, the chip has a circuit that compresses data before sending it to individual cores.<\/span><\/p>\n<p style=\"text-align: justify;\"><span style=\"color: #000000;\">Each core is also able to communicate directly with its immediate neighbors, so that if they need to share data, they don\u2019t have to route it through main memory. This is essential in a convolutional neural network, in which so many nodes are processing the same data.<\/span><\/p>\n<p style=\"text-align: justify;\"><span style=\"color: #000000;\">The final key to the chip\u2019s efficiency is special-purpose circuitry that allocates tasks across cores. In its local memory, a core needs to store not only the data manipulated by the nodes it\u2019s simulating but data describing the nodes themselves. The allocation circuit can be reconfigured for different types of networks, automatically distributing both types of data across cores in a way that maximizes the amount of work that each of them can do before fetching more data from main memory.<\/span><\/p>\n<p style=\"text-align: justify;\"><span style=\"color: #000000;\">At the conference, the MIT researchers used Eyeriss to implement a neural network that performs an image-recognition task, the first time that a state-of-the-art neural network has been demonstrated on a custom chip.<\/span><\/p>\n","protected":false},"excerpt":{"rendered":"<p>At the International Solid State Circuits Conference in San Francisco, MIT researchers presented a new chip designed specifically to implement neural networks.<\/p>\n","protected":false},"author":6,"featured_media":7566,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[14,28],"tags":[],"class_list":["post-7564","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-innovation","category-techbiz"],"featured_image_urls":{"full":["https:\/\/www.revoscience.com\/en\/wp-content\/uploads\/2016\/02\/MIT-Neural-Chip_0.jpg",448,299,false],"thumbnail":["https:\/\/www.revoscience.com\/en\/wp-content\/uploads\/2016\/02\/MIT-Neural-Chip_0-150x150.jpg",150,150,true],"medium":["https:\/\/www.revoscience.com\/en\/wp-content\/uploads\/2016\/02\/MIT-Neural-Chip_0-300x200.jpg",300,200,true],"medium_large":["https:\/\/www.revoscience.com\/en\/wp-content\/uploads\/2016\/02\/MIT-Neural-Chip_0.jpg",448,299,false],"large":["https:\/\/www.revoscience.com\/en\/wp-content\/uploads\/2016\/02\/MIT-Neural-Chip_0.jpg",448,299,false],"1536x1536":["https:\/\/www.revoscience.com\/en\/wp-content\/uploads\/2016\/02\/MIT-Neural-Chip_0.jpg",448,299,false],"2048x2048":["https:\/\/www.revoscience.com\/en\/wp-content\/uploads\/2016\/02\/MIT-Neural-Chip_0.jpg",448,299,false],"ultp_layout_landscape_large":["https:\/\/www.revoscience.com\/en\/wp-content\/uploads\/2016\/02\/MIT-Neural-Chip_0.jpg",448,299,false],"ultp_layout_landscape":["https:\/\/www.revoscience.com\/en\/wp-content\/uploads\/2016\/02\/MIT-Neural-Chip_0.jpg",448,299,false],"ultp_layout_portrait":["https:\/\/www.revoscience.com\/en\/wp-content\/uploads\/2016\/02\/MIT-Neural-Chip_0.jpg",448,299,false],"ultp_layout_square":["https:\/\/www.revoscience.com\/en\/wp-content\/uploads\/2016\/02\/MIT-Neural-Chip_0.jpg",448,299,false],"newspaper-x-single-post":["https:\/\/www.revoscience.com\/en\/wp-content\/uploads\/2016\/02\/MIT-Neural-Chip_0.jpg",448,299,false],"newspaper-x-recent-post-big":["https:\/\/www.revoscience.com\/en\/wp-content\/uploads\/2016\/02\/MIT-Neural-Chip_0.jpg",448,299,false],"newspaper-x-recent-post-list-image":["https:\/\/www.revoscience.com\/en\/wp-content\/uploads\/2016\/02\/MIT-Neural-Chip_0.jpg",95,63,false],"web-stories-poster-portrait":["https:\/\/www.revoscience.com\/en\/wp-content\/uploads\/2016\/02\/MIT-Neural-Chip_0.jpg",448,299,false],"web-stories-publisher-logo":["https:\/\/www.revoscience.com\/en\/wp-content\/uploads\/2016\/02\/MIT-Neural-Chip_0.jpg",96,64,false],"web-stories-thumbnail":["https:\/\/www.revoscience.com\/en\/wp-content\/uploads\/2016\/02\/MIT-Neural-Chip_0.jpg",150,100,false]},"author_info":{"info":["Amrita Tuladhar"]},"category_info":"<a href=\"https:\/\/www.revoscience.com\/en\/category\/innovation\/\" rel=\"category tag\">Innovation<\/a> <a href=\"https:\/\/www.revoscience.com\/en\/category\/techbiz\/\" rel=\"category tag\">Tech<\/a>","tag_info":"Tech","comment_count":"0","_links":{"self":[{"href":"https:\/\/www.revoscience.com\/en\/wp-json\/wp\/v2\/posts\/7564","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.revoscience.com\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.revoscience.com\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.revoscience.com\/en\/wp-json\/wp\/v2\/users\/6"}],"replies":[{"embeddable":true,"href":"https:\/\/www.revoscience.com\/en\/wp-json\/wp\/v2\/comments?post=7564"}],"version-history":[{"count":0,"href":"https:\/\/www.revoscience.com\/en\/wp-json\/wp\/v2\/posts\/7564\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.revoscience.com\/en\/wp-json\/wp\/v2\/media\/7566"}],"wp:attachment":[{"href":"https:\/\/www.revoscience.com\/en\/wp-json\/wp\/v2\/media?parent=7564"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.revoscience.com\/en\/wp-json\/wp\/v2\/categories?post=7564"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.revoscience.com\/en\/wp-json\/wp\/v2\/tags?post=7564"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}