{"id":1310,"date":"2020-09-17T09:50:58","date_gmt":"2020-09-17T13:50:58","guid":{"rendered":"https:\/\/www.ece.ncsu.edu\/?p=237965"},"modified":"2020-09-17T09:50:58","modified_gmt":"2020-09-17T13:50:58","slug":"new-data-processing-module-makes-deep-neural-networks-smarter","status":"publish","type":"post","link":"https:\/\/my.ece.ncsu.edu\/communications\/2020\/new-data-processing-module-makes-deep-neural-networks-smarter\/","title":{"rendered":"New Data Processing Module Makes Deep Neural Networks Smarter"},"content":{"rendered":"<p><img decoding=\"async\" width=\"1024\" height=\"576\" src=\"https:\/\/ece.ncsu.edu\/wp-content\/uploads\/2020\/09\/Tianfu-Wu-AN-2020-HEADER-1024x576.jpg\" class=\"attachment-large size-large wp-post-image\" alt=\"\" loading=\"lazy\" srcset=\"https:\/\/ece.ncsu.edu\/wp-content\/uploads\/2020\/09\/Tianfu-Wu-AN-2020-HEADER-980x551.jpg 980w, https:\/\/ece.ncsu.edu\/wp-content\/uploads\/2020\/09\/Tianfu-Wu-AN-2020-HEADER-480x270.jpg 480w\" sizes=\"auto, (min-width: 0px) and (max-width: 480px) 480px, (min-width: 481px) and (max-width: 980px) 980px, (min-width: 981px) 1024px, 100vw\" \/><\/p>\n<p>Artificial intelligence researchers at North Carolina State University have improved the performance of deep neural networks by combining feature normalization and feature attention modules into a single module that they call attentive normalization (AN). The hybrid module improves the accuracy of the system significantly while using negligible extra computational power.<\/p>\n<p>\u201cFeature normalization is a crucial element of training deep neural networks, and feature attention is equally important for helping networks highlight which features learned from raw data are most important for accomplishing a given task,\u201d says <a href=\"https:\/\/www.ece.ncsu.edu\/people\/twu19\/\">Tianfu Wu<\/a>, the corresponding author of a paper on the work and an assistant professor of electrical and computer engineering at NC\u00a0State. \u201cBut they have mostly been treated separately. We found that combining them made them more efficient and effective.\u201d<\/p>\n<p>To test their AN module, the researchers plugged it into four of the most widely used neural network architectures: ResNets, DenseNets, MobileNetsV2, and AOGNets. They then tested the networks against two industry-standard benchmarks: the ImageNet-1000 classification benchmark and the MS-COCO 2017 object detection and instance segmentation benchmark.<\/p>\n<p>\u201cWe found that AN improved performance for all four architectures on both benchmarks,\u201d Wu says. \u201cFor example, top-1 accuracy in the ImageNet-1000 improved by between 0.5% and 2.7%. And Average Precision (AP) accuracy increased by up to 1.8% for bounding box and 2.2% for semantic mask in MS-COCO.<\/p>\n<p>\u201cAnother advantage of AN is that it facilitates better transfer learning between different domains,\u201d Wu says. \u201cFor example, from image classification in ImageNet to object detection and semantic segmentation in MS-COCO. This is illustrated by the performance improvement in the MS-COCO benchmark, which was obtained by fine-tuning ImageNet-pre-trained deep neural networks in MS-COCO, a common workflow in state-of-the-art computer vision.<\/p>\n<p>\u201cWe have released the source code and hope our AN will lead to better integrative design of deep neural networks.\u201d<\/p>\n<p>The paper, \u201c<a href=\"https:\/\/www.ecva.net\/papers\/eccv_2020\/papers_ECCV\/papers\/123620069.pdf\"  rel=\"noreferrer noopener\">Attentive Normalization<\/a>,\u201d was presented at the European Conference on Computer Vision (ECCV), which was held online Aug. 23-28. The paper was co-authored by Xilai Li, a recent Ph.D. graduate from NC\u00a0State; and by Wei Sun, a Ph.D. student at NC\u00a0State. The work was done with support from the National Science Foundation, under grants 1909644, 1822477, and 2013451; and by the U.S. Army Research Office, under grant W911NF1810295.<\/p>\n","protected":false},"excerpt":{"rendered":"<p><img decoding=\"async\" width=\"1024\" height=\"576\" src=\"https:\/\/ece.ncsu.edu\/wp-content\/uploads\/2020\/09\/Tianfu-Wu-AN-2020-HEADER-1024x576.jpg\" class=\"attachment-large size-large wp-post-image\" alt=\"\" loading=\"lazy\" srcset=\"https:\/\/ece.ncsu.edu\/wp-content\/uploads\/2020\/09\/Tianfu-Wu-AN-2020-HEADER-980x551.jpg 980w, https:\/\/ece.ncsu.edu\/wp-content\/uploads\/2020\/09\/Tianfu-Wu-AN-2020-HEADER-480x270.jpg 480w\" sizes=\"auto, (min-width: 0px) and (max-width: 480px) 480px, (min-width: 481px) and (max-width: 980px) 980px, (min-width: 981px) 1024px, 100vw\">AI researchers from NC State ECE\u2014including Tianfu Wu\u2014have improved the performance of deep neural networks by combining feature normalization and feature attention into a single module.<\/p>\n","protected":false},"author":2,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"_acf_changed":false,"ncst_dynamicHeaderBlockName":"","ncst_dynamicHeaderData":"","ncst_content_audit_freq":"","ncst_content_audit_date":"","ncst_content_audit_display":false,"ncst_backToTopFlag":"","footnotes":""},"categories":[180],"tags":[],"class_list":["post-1310","post","type-post","status-publish","format-standard","hentry","category-research"],"displayCategory":null,"acf":{"ncst_posts_meta_modified_date":null},"_links":{"self":[{"href":"https:\/\/my.ece.ncsu.edu\/communications\/wp-json\/wp\/v2\/posts\/1310","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/my.ece.ncsu.edu\/communications\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/my.ece.ncsu.edu\/communications\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/my.ece.ncsu.edu\/communications\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/my.ece.ncsu.edu\/communications\/wp-json\/wp\/v2\/comments?post=1310"}],"version-history":[{"count":2,"href":"https:\/\/my.ece.ncsu.edu\/communications\/wp-json\/wp\/v2\/posts\/1310\/revisions"}],"predecessor-version":[{"id":2546,"href":"https:\/\/my.ece.ncsu.edu\/communications\/wp-json\/wp\/v2\/posts\/1310\/revisions\/2546"}],"wp:attachment":[{"href":"https:\/\/my.ece.ncsu.edu\/communications\/wp-json\/wp\/v2\/media?parent=1310"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/my.ece.ncsu.edu\/communications\/wp-json\/wp\/v2\/categories?post=1310"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/my.ece.ncsu.edu\/communications\/wp-json\/wp\/v2\/tags?post=1310"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}