{"id":291,"date":"2021-04-27T16:27:12","date_gmt":"2021-04-27T07:27:12","guid":{"rendered":"https:\/\/nanotechnet.t.u-tokyo.ac.jp\/kairo\/?p=291"},"modified":"2021-04-27T16:27:12","modified_gmt":"2021-04-27T07:27:12","slug":"%e7%9b%b8%e6%be%a4%e7%a0%94%e7%a9%b6%e5%ae%a4_en","status":"publish","type":"post","link":"https:\/\/nanotechnet.t.u-tokyo.ac.jp\/kairo\/?p=291","title":{"rendered":"\u76f8\u6fa4\u7814\u7a76\u5ba4_en"},"content":{"rendered":"\n<h2 class=\"wp-block-heading\">Multimedia, computer vision, image processing  :Kiyoharu Aizawa<\/h2>\n\n\n\n<h3 class=\"wp-block-heading\">Fundamentals of Recognition and Learning, Open World.<\/h3>\n\n\n\n<div class=\"wp-block-columns is-layout-flex wp-container-core-columns-is-layout-1 wp-block-columns-is-layout-flex\">\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<p>Deep Learning accurately works for closed dataset containing large number of data per class. However, in reality, unknown classes and new classes with small amount of data frequently appear. We are investigating identification and recognition techniques for such open world situation. The specific topics are effective methodology for noisy training data, out-of-distribution detection, positive-unlabeled learning, open-set data learning, uncertainty estimation etc. We also work on scence character recognition and show that SOTA can be achieved with about 1\/100th of the real images as opposed to existing approaches that use huge amounts of synthetic characters.<\/p>\n<\/div>\n\n\n\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"946\" height=\"408\" src=\"https:\/\/nanotechnet.t.u-tokyo.ac.jp\/kairo\/wp-content\/uploads\/2021\/04\/aizawa_fig1.png\" alt=\"\" class=\"wp-image-285\" srcset=\"https:\/\/nanotechnet.t.u-tokyo.ac.jp\/kairo\/wp-content\/uploads\/2021\/04\/aizawa_fig1.png 946w, https:\/\/nanotechnet.t.u-tokyo.ac.jp\/kairo\/wp-content\/uploads\/2021\/04\/aizawa_fig1-300x129.png 300w, https:\/\/nanotechnet.t.u-tokyo.ac.jp\/kairo\/wp-content\/uploads\/2021\/04\/aizawa_fig1-768x331.png 768w\" sizes=\"(max-width: 946px) 100vw, 946px\" \/><\/figure>\n<\/div>\n<\/div>\n\n\n\n<h3 class=\"wp-block-heading\">360\u030a Image Processing\u30fb3D\uff0cMovie Map.<\/h3>\n\n\n\n<div class=\"wp-block-columns is-layout-flex wp-container-core-columns-is-layout-2 wp-block-columns-is-layout-flex\">\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<p>We are investigating 360\u030a image processing. Specifically, we build \u201cmovie map\u201d for walkers to explore in a city. Using 360\u030a street videos, we work on many different research issues such as hyperlaspe 360\u030avideo, 360\u030aimage object detection, accurate SLAM, intersection detection, RoI detection, real-time route view generation based on user input, building database of automatically segmented video sections etc. We prototyped our first version MovieMap by which we can freely explore in a certain area in a city.<\/p>\n<\/div>\n\n\n\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"624\" height=\"470\" src=\"https:\/\/nanotechnet.t.u-tokyo.ac.jp\/kairo\/wp-content\/uploads\/2021\/04\/aizawa_fig2.png\" alt=\"\" class=\"wp-image-286\" srcset=\"https:\/\/nanotechnet.t.u-tokyo.ac.jp\/kairo\/wp-content\/uploads\/2021\/04\/aizawa_fig2.png 624w, https:\/\/nanotechnet.t.u-tokyo.ac.jp\/kairo\/wp-content\/uploads\/2021\/04\/aizawa_fig2-300x226.png 300w\" sizes=\"(max-width: 624px) 100vw, 624px\" \/><\/figure>\n<\/div>\n<\/div>\n\n\n\n<h3 class=\"wp-block-heading\">Life Logging, Food Computing<\/h3>\n\n\n\n<div class=\"wp-block-columns is-layout-flex wp-container-core-columns-is-layout-3 wp-block-columns-is-layout-flex\">\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<p>We have been pioneering life logging technology. Starting from generic purpose lifelogging, we now pursue specific purpose lifelogging. We focus on research on capture and analysis of our daily food logs (FoodLog), Using the app we developed, food records we captured exceeds 10 million. We are investigating various processing of FoodLog data, such as personalized food recognition, recipe and food record multimodal analysis, prediction of healthy index, etc. We also built a new a new food logging tool, FoodLog Athl, and made it publicly available that supports communications between users and dietitians.<\/p>\n<\/div>\n\n\n\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"576\" height=\"658\" src=\"https:\/\/nanotechnet.t.u-tokyo.ac.jp\/kairo\/wp-content\/uploads\/2021\/04\/aizawa_fig3.png\" alt=\"\" class=\"wp-image-287\" srcset=\"https:\/\/nanotechnet.t.u-tokyo.ac.jp\/kairo\/wp-content\/uploads\/2021\/04\/aizawa_fig3.png 576w, https:\/\/nanotechnet.t.u-tokyo.ac.jp\/kairo\/wp-content\/uploads\/2021\/04\/aizawa_fig3-263x300.png 263w\" sizes=\"(max-width: 576px) 100vw, 576px\" \/><\/figure>\n<\/div>\n<\/div>\n\n\n\n<h3 class=\"wp-block-heading\">Manga, Comic Computing<\/h3>\n\n\n\n<div class=\"wp-block-columns is-layout-flex wp-container-core-columns-is-layout-4 wp-block-columns-is-layout-flex\">\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<p>Manga, our unique culture, is our research target, which has rarely been discussed in the field of image processing. We have built a world largest scale Manga dataset, Manga109, and investigate image processing techniques such as retrieval, segmentation, recognition, colorization etc.<\/p>\n<\/div>\n\n\n\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"499\" height=\"340\" src=\"https:\/\/nanotechnet.t.u-tokyo.ac.jp\/kairo\/wp-content\/uploads\/2021\/04\/aizawa_fig4.png\" alt=\"\" class=\"wp-image-288\" srcset=\"https:\/\/nanotechnet.t.u-tokyo.ac.jp\/kairo\/wp-content\/uploads\/2021\/04\/aizawa_fig4.png 499w, https:\/\/nanotechnet.t.u-tokyo.ac.jp\/kairo\/wp-content\/uploads\/2021\/04\/aizawa_fig4-300x204.png 300w\" sizes=\"(max-width: 499px) 100vw, 499px\" \/><\/figure>\n<\/div>\n<\/div>\n\n\n\n<h3 class=\"wp-block-heading\">Desing, Fonts<\/h3>\n\n\n\n<div class=\"wp-block-columns is-layout-flex wp-container-core-columns-is-layout-5 wp-block-columns-is-layout-flex\">\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<p>We investigate image technology for creation, retrieval of designs of fonts, products etc. We built Emotype &#8211; a mobile messenger expressing our emotions via different typo-graphics, social font search by multimodal inputs, font search across various languages, font generation using a small number of samples, design of bags etc.<\/p>\n<\/div>\n\n\n\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"982\" height=\"719\" src=\"https:\/\/nanotechnet.t.u-tokyo.ac.jp\/kairo\/wp-content\/uploads\/2021\/04\/aizawa_fig5.png\" alt=\"\" class=\"wp-image-289\" srcset=\"https:\/\/nanotechnet.t.u-tokyo.ac.jp\/kairo\/wp-content\/uploads\/2021\/04\/aizawa_fig5.png 982w, https:\/\/nanotechnet.t.u-tokyo.ac.jp\/kairo\/wp-content\/uploads\/2021\/04\/aizawa_fig5-300x220.png 300w, https:\/\/nanotechnet.t.u-tokyo.ac.jp\/kairo\/wp-content\/uploads\/2021\/04\/aizawa_fig5-768x562.png 768w\" sizes=\"(max-width: 982px) 100vw, 982px\" \/><\/figure>\n<\/div>\n<\/div>\n","protected":false},"excerpt":{"rendered":"<p>Multimedia, computer vision, image processing :Kiyoharu Aizawa Fundamentals of Recognition and Learning, Open  [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[3],"tags":[],"_links":{"self":[{"href":"https:\/\/nanotechnet.t.u-tokyo.ac.jp\/kairo\/index.php?rest_route=\/wp\/v2\/posts\/291"}],"collection":[{"href":"https:\/\/nanotechnet.t.u-tokyo.ac.jp\/kairo\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/nanotechnet.t.u-tokyo.ac.jp\/kairo\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/nanotechnet.t.u-tokyo.ac.jp\/kairo\/index.php?rest_route=\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/nanotechnet.t.u-tokyo.ac.jp\/kairo\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=291"}],"version-history":[{"count":1,"href":"https:\/\/nanotechnet.t.u-tokyo.ac.jp\/kairo\/index.php?rest_route=\/wp\/v2\/posts\/291\/revisions"}],"predecessor-version":[{"id":292,"href":"https:\/\/nanotechnet.t.u-tokyo.ac.jp\/kairo\/index.php?rest_route=\/wp\/v2\/posts\/291\/revisions\/292"}],"wp:attachment":[{"href":"https:\/\/nanotechnet.t.u-tokyo.ac.jp\/kairo\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=291"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/nanotechnet.t.u-tokyo.ac.jp\/kairo\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=291"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/nanotechnet.t.u-tokyo.ac.jp\/kairo\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=291"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}