<?xml version="1.0" encoding="UTF-8"?>
<?xml-stylesheet type="text/xsl" media="screen" href="/~d/styles/atom10full.xsl"?><?xml-stylesheet type="text/css" media="screen" href="http://feeds.feedburner.com/~d/styles/itemcontent.css"?><feed xmlns="http://www.w3.org/2005/Atom" xmlns:openSearch="http://a9.com/-/spec/opensearchrss/1.0/" xmlns:blogger="http://schemas.google.com/blogger/2008" xmlns:georss="http://www.georss.org/georss" xmlns:gd="http://schemas.google.com/g/2005" xmlns:thr="http://purl.org/syndication/thread/1.0" xmlns:feedburner="http://rssnamespace.org/feedburner/ext/1.0"><id>tag:blogger.com,1999:blog-21224994</id><updated>2017-06-15T10:19:11.339-07:00</updated><category term="Machine Learning" /><category term="Education" /><category term="University Relations" /><category term="Google Brain" /><category term="Publications" /><category term="Deep Learning" /><category term="TensorFlow" /><category term="Natural Language Processing" /><category term="Neural Networks" /><category term="open source" /><category term="conference" /><category term="MOOC" /><category term="Research Awards" /><category term="Computer Vision" /><category term="Natural Language Understanding" /><category term="Computer Science" /><category term="conferences" /><category term="Research" /><category term="Machine Intelligence" /><category term="datasets" /><category term="Awards" /><category term="YouTube" /><category term="Machine Perception" /><category term="Android" /><category term="K-12" /><category term="Speech" /><category term="Structured Data" /><category term="ACM" /><category term="Earth Engine" /><category term="Security and Privacy" /><category term="ph.d. fellowship" /><category term="HCI" /><category term="Voice Search" /><category term="grants" /><category term="Faculty Summit" /><category term="Health" /><category term="Machine Translation" /><category term="Quantum Computing" /><category term="Search" /><category term="Vision Research" /><category term="Collaboration" /><category term="Course Builder" /><category term="Google+" /><category term="NIPS" /><category term="User Experience" /><category term="Visualization" /><category term="WWW" /><category term="statistics" /><category term="CVPR" /><category term="Fusion Tables" /><category term="Google Books" /><category term="Google Cloud Platform" /><category term="Image Processing" /><category term="Moore's Law" /><category term="Ngram" /><category term="Translate" /><category term="market algorithms" /><category term="optimization" /><category term="renewable energy" /><category term="ACL" /><category term="Algorithms" /><category term="App Engine" /><category term="Computational Photography" /><category term="Diversity" /><category term="Gmail" /><category term="Google Genomics" /><category term="Google Play Apps" /><category term="Google Translate" /><category term="Information Retrieval" /><category term="Internet of Things" /><category term="Machine Hearing" /><category term="NLP" /><category term="Networks" /><category term="Semi-supervised Learning" /><category term="TTS" /><category term="UI" /><category term="accessibility" /><category term="crowd-sourcing" /><category term="economics" /><category term="internationalization" /><category term="publication" /><category term="search ads" /><category term="wikipedia" /><category term="API" /><category term="App Inventor" /><category term="Art" /><category term="Automatic Speech Recognition" /><category term="China" /><category term="Cloud Computing" /><category term="DeepMind" /><category term="EMEA" /><category term="Environment" /><category term="Exacycle" /><category term="Expander" /><category term="Google Drive" /><category term="Google Science Fair" /><category term="Graph" /><category term="Graph Mining" /><category term="Hardware" /><category term="ICLR" /><category term="ICML" /><category term="Inbox" /><category term="Labs" /><category term="MapReduce" /><category term="On-device Learning" /><category term="Optical Character Recognition" /><category term="Policy" /><category term="Robotics" /><category term="Social Networks" /><category term="Speech Recognition" /><category term="Supervised Learning" /><category term="Systems" /><category term="VLDB" /><category term="Video Analysis" /><category term="ads" /><category term="data science" /><category term="distributed systems" /><category term="schema.org" /><category term="trends" /><category term="AI" /><category term="Acoustic Modeling" /><category term="Adaptive Data Analysis" /><category term="Africa" /><category term="Android Wear" /><category term="April Fools" /><category term="Audio" /><category term="Australia" /><category term="Cantonese" /><category term="Chemistry" /><category term="Chrome" /><category term="Computational Imaging" /><category term="Conservation" /><category term="Data Center" /><category term="Data Discovery" /><category term="DeepDream" /><category term="EMNLP" /><category term="Electronic Commerce and Algorithms" /><category term="Encryption" /><category term="Entity Salience" /><category term="Europe" /><category term="Faculty Institute" /><category term="Flu Trends" /><category term="Gboard" /><category term="Google Docs" /><category term="Google Maps" /><category term="Google Photos" /><category term="Google Sheets" /><category term="Google Trips" /><category term="Google Voice Search" /><category term="Government" /><category term="High Dynamic Range Imaging" /><category term="ICSE" /><category term="IPython" /><category term="Image Annotation" /><category term="Image Classification" /><category term="Interspeech" /><category term="Journalism" /><category term="KDD" /><category term="Keyboard Input" /><category term="Klingon" /><category term="Korean" /><category term="Linear Optimization" /><category term="Low-Light Photography" /><category term="ML" /><category term="Magenta" /><category term="Market Research" /><category term="Mixed Reality" /><category term="Multimodal Learning" /><category term="NAACL" /><category term="Network Management" /><category term="Nexus" /><category term="PhD Fellowship" /><category term="PhotoScan" /><category term="PiLab" /><category term="Pixel" /><category term="Professional Development" /><category term="Proposals" /><category term="Public Data Explorer" /><category term="SIGCOMM" /><category term="SIGMOD" /><category term="Site Reliability Engineering" /><category term="Software" /><category term="Style Transfer" /><category term="TPU" /><category term="TV" /><category term="UNIX" /><category term="Virtual Reality" /><category term="Visiting Faculty" /><category term="Wiki" /><category term="adsense" /><category term="adwords" /><category term="correlate" /><category term="electronics" /><category term="entities" /><category term="gamification" /><category term="jsm" /><category term="jsm2011" /><category term="localization" /><category term="operating systems" /><category term="osdi" /><category term="osdi10" /><category term="patents" /><category term="resource optimization" /><category term="video" /><title type="text">Google Research Blog</title><subtitle type="html">The latest news on Google Research.</subtitle><link rel="alternate" type="text/html" href="http://research.googleblog.com/" /><link rel="next" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default?start-index=26&amp;max-results=25&amp;redirect=false" /><author><name>Google Blogs</name><email>noreply@blogger.com</email><gd:image rel="http://schemas.google.com/g/2005#thumbnail" width="32" height="32" src="//lh3.googleusercontent.com/-SMZmHeOVbFs/AAAAAAAAAAI/AAAAAAAAR7c/esUAZEvmr9M/s512-c/photo.jpg" /></author><generator version="7.00" uri="http://www.blogger.com">Blogger</generator><openSearch:totalResults>504</openSearch:totalResults><openSearch:startIndex>1</openSearch:startIndex><openSearch:itemsPerPage>25</openSearch:itemsPerPage><atom10:link xmlns:atom10="http://www.w3.org/2005/Atom" rel="self" type="application/atom+xml" href="http://feeds.feedburner.com/blogspot/gJZg" /><feedburner:info uri="blogspot/gjzg" /><atom10:link xmlns:atom10="http://www.w3.org/2005/Atom" rel="hub" href="http://pubsubhubbub.appspot.com/" /><entry><id>tag:blogger.com,1999:blog-21224994.post-776695066333371847</id><published>2017-06-15T10:00:00.000-07:00</published><updated>2017-06-15T10:19:11.348-07:00</updated><category scheme="http://www.blogger.com/atom/ns#" term="Computer Vision" /><category scheme="http://www.blogger.com/atom/ns#" term="Deep Learning" /><category scheme="http://www.blogger.com/atom/ns#" term="Machine Learning" /><category scheme="http://www.blogger.com/atom/ns#" term="Machine Perception" /><category scheme="http://www.blogger.com/atom/ns#" term="open source" /><category scheme="http://www.blogger.com/atom/ns#" term="TensorFlow" /><title type="text">Supercharge your Computer Vision models with the TensorFlow Object Detection API</title><content type="html">&lt;span class="byline-author"&gt;Posted by Jonathan Huang, Research Scientist and Vivek Rathod, Software Engineer&lt;/span&gt;&lt;br /&gt;&lt;br /&gt;&lt;i&gt;(Cross-posted on the &lt;a href="http://opensource.googleblog.com/2017/06/supercharge-your-computer-vision-models.html"&gt;Google Open Source Blog&lt;/a&gt;)&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;At Google, we develop flexible state-of-the-art machine learning (ML) systems for computer vision that not only can be used to improve our products and services, but also &lt;a href="https://research.googleblog.com/2016/08/improving-inception-and-image.html"&gt;spur progress in the research community&lt;/a&gt;. Creating accurate ML models capable of localizing and identifying multiple objects in a single image remains a core challenge in the field, and we invest a significant amount of time training and experimenting with these systems. &lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://2.bp.blogspot.com/-MO3T6Hybpkg/WUG-QjHrHbI/AAAAAAAAB2M/tQKa2ljTkRwYgtok3o_3Y6F5KCbC7a-qQCLcBGAs/s1600/image1.jpg" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" data-original-height="1067" data-original-width="1600" height="426" src="https://2.bp.blogspot.com/-MO3T6Hybpkg/WUG-QjHrHbI/AAAAAAAAB2M/tQKa2ljTkRwYgtok3o_3Y6F5KCbC7a-qQCLcBGAs/s640/image1.jpg" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Detected objects in a sample image (from the &lt;a href="http://mscoco.org/"&gt;COCO&lt;/a&gt; dataset) made by one of our models. Image credit: &lt;a href="https://www.flickr.com/photos/mike_miley/"&gt;Michael Miley&lt;/a&gt;, &lt;a href="https://www.flickr.com/photos/mike_miley/4678754542/in/photolist-88rQHL-88oBVp-88oC2B-88rS6J-88rSqm-88oBLv-88oBC4"&gt;original image&lt;/a&gt;.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;Last October, our in-house object detection system achieved new state-of-the-art results, and placed first in the &lt;a href="http://mscoco.org/dataset/#detections-leaderboard"&gt;COCO detection challenge&lt;/a&gt;. Since then, this system has generated results for a number of research publications&lt;sup&gt;1,2,3,4,5,6,7&lt;/sup&gt; and has been put to work in Google products such as &lt;a href="https://nest.com/nest-aware/"&gt;NestCam&lt;/a&gt;, the &lt;a href="https://www.blog.google/products/search/now-image-search-can-jump-start-your-search-style/"&gt;similar items and style ideas&lt;/a&gt; feature in Image Search and &lt;a href="https://research.googleblog.com/2017/05/updating-google-maps-with-deep-learning.html"&gt;street number and name detection&lt;/a&gt; in Street View.&lt;br /&gt;&lt;br /&gt;Today we are happy to make this system available to the broader research community via the &lt;a href="https://github.com/tensorflow/models/tree/master/object_detection"&gt;TensorFlow Object Detection API&lt;/a&gt;.  This codebase is an open-source framework built on top of &lt;a href="https://www.tensorflow.org/"&gt;TensorFlow&lt;/a&gt; that makes it easy to construct, train and deploy object detection models.  Our goals in designing this system was to support state-of-the-art models while allowing for rapid exploration and research.  Our first release contains the following:&lt;br /&gt;&lt;ul&gt;&lt;li&gt;A selection of trainable detection models, including:&lt;/li&gt;&lt;ul&gt;&lt;li&gt;&lt;a href="https://arxiv.org/abs/1512.02325"&gt;Single Shot Multibox Detector&lt;/a&gt; (SSD) with &lt;a href="http://research.googleblog.com/2017/06/mobilenets-open-source-models-for.html"&gt;MobileNets&lt;/a&gt;&lt;/li&gt;&lt;li&gt;SSD with &lt;a href="https://arxiv.org/abs/1512.00567"&gt;Inception V2&lt;/a&gt;&lt;/li&gt;&lt;li&gt;&lt;a href="https://arxiv.org/abs/1605.06409"&gt;Region-Based Fully Convolutional Networks&lt;/a&gt; (R-FCN) with &lt;a href="https://arxiv.org/abs/1512.03385"&gt;Resnet 101&lt;/a&gt;&lt;/li&gt;&lt;li&gt;&lt;a href="https://arxiv.org/abs/1506.01497"&gt;Faster RCNN&lt;/a&gt; with Resnet 101&lt;/li&gt;&lt;li&gt;&lt;a href="https://arxiv.org/abs/1506.01497"&gt;Faster RCNN&lt;/a&gt; with &lt;a href="https://arxiv.org/abs/1602.07261"&gt;Inception Resnet v2&lt;/a&gt;&lt;/li&gt;&lt;/ul&gt;&lt;li&gt;Frozen weights (trained on the &lt;a href="http://mscoco.org/"&gt;COCO dataset&lt;/a&gt;) for each of the above models to be used for out-of-the-box inference purposes.&lt;/li&gt;&lt;li&gt;A &lt;a href="https://github.com/tensorflow/models/tree/master/object_detection/object_detection_tutorial.ipynb"&gt;Jupyter notebook&lt;/a&gt; for performing out-of-the-box inference with one of our released models&lt;/li&gt;&lt;li&gt;Convenient local training scripts as well as distributed training and evaluation pipelines via Google Cloud&lt;/li&gt;&lt;/ul&gt;The SSD models that use MobileNet are lightweight, so that they can be comfortably run in real time on mobile devices. Our winning COCO submission in 2016 used an ensemble of the Faster RCNN models, which are are more computationally intensive but significantly more accurate.  For more details on the performance of these models, see our &lt;a href="https://arxiv.org/abs/1611.10012"&gt;CVPR 2017 paper&lt;/a&gt;.&lt;br /&gt;&lt;br /&gt;&lt;b&gt;Are you ready to get started?&lt;/b&gt;&lt;br /&gt;We’ve certainly found this code to be useful for our computer vision needs, and we hope that you will as well.  Contributions to the codebase are welcome and please stay tuned for our own further updates to the framework. To get started, download the code &lt;a href="https://github.com/tensorflow/models/tree/master/object_detection"&gt;here&lt;/a&gt; and try detecting objects in some of your own images using the &lt;a href="https://github.com/tensorflow/models/tree/master/object_detection/object_detection_tutorial.ipynb"&gt;Jupyter notebook&lt;/a&gt;, or &lt;a href="https://cloud.google.com/blog/big-data/2017/06/training-an-object-detector-using-cloud-machine-learning-engine"&gt;training your own pet detector on Cloud ML engine&lt;/a&gt;!   &lt;br /&gt;&lt;b&gt;&lt;br /&gt;&lt;/b&gt; &lt;b&gt;Acknowledgements&lt;/b&gt;&lt;br /&gt;The release of the Tensorflow Object Detection API and the pre-trained model zoo has been the result of widespread collaboration among Google researchers with feedback and testing from product groups. In particular we want to highlight the contributions of the following individuals:&lt;br /&gt;&lt;br /&gt;&lt;b&gt;Core Contributors:&lt;/b&gt; &lt;i&gt;Derek Chow, Chen Sun, Menglong Zhu, Matthew Tang, Anoop Korattikara, Alireza Fathi, Ian Fischer, Zbigniew Wojna, Yang Song, Sergio Guadarrama, Jasper Uijlings, Viacheslav Kovalevskyi, Kevin Murphy&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;b&gt;Also special thanks to:&lt;/b&gt; &lt;i&gt;Andrew Howard, Rahul Sukthankar, Vittorio Ferrari, Tom Duerig, Chuck Rosenberg, Hartwig Adam, Jing Jing Long, Victor Gomes, George Papandreou, Tyler Zhu&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;b&gt;References&lt;/b&gt;&lt;br /&gt;&lt;ol&gt;&lt;li&gt;&lt;a href="https://arxiv.org/abs/1611.10012"&gt;Speed/accuracy trade-offs for modern convolutional object detectors&lt;/a&gt;, &lt;i&gt;Huang et al., CVPR 2017 (paper describing this framework)&lt;/i&gt;&lt;/li&gt;&lt;li&gt;&lt;a href="https://arxiv.org/abs/1701.01779"&gt;Towards Accurate Multi-person Pose Estimation in the Wild&lt;/a&gt;, &lt;i&gt;Papandreou et al., CVPR 2017&lt;/i&gt;&lt;/li&gt;&lt;li&gt;&lt;a href="https://arxiv.org/abs/1702.00824"&gt;YouTube-BoundingBoxes: A Large High-Precision Human-Annotated Data Set for Object Detection in Video&lt;/a&gt;, &lt;i&gt;Real et al., CVPR 2017 (see also our &lt;a href="https://research.googleblog.com/2017/02/advancing-research-on-video.html"&gt;blog post&lt;/a&gt;)&lt;/i&gt;&lt;/li&gt;&lt;li&gt;&lt;a href="https://arxiv.org/abs/1612.06851"&gt;Beyond Skip Connections: Top-Down Modulation for Object Detection&lt;/a&gt;, &lt;i&gt;Shrivastava et al., arXiv preprint arXiv:1612.06851, 2016&lt;/i&gt;&lt;/li&gt;&lt;li&gt;&lt;a href="https://arxiv.org/abs/1612.02297"&gt;Spatially Adaptive Computation Time for Residual Networks&lt;/a&gt;, &lt;i&gt;Figurnov et al., CVPR 2017&lt;/i&gt;&lt;/li&gt;&lt;li&gt;&lt;a href="https://arxiv.org/abs/1705.08421"&gt;AVA: A Video Dataset of Spatio-temporally Localized Atomic Visual Actions&lt;/a&gt;, &lt;i&gt;Gu et al., arXiv preprint arXiv:1705.08421, 2017&lt;/i&gt;&lt;/li&gt;&lt;li&gt;&lt;a href="https://arxiv.org/abs/1704.04861"&gt;MobileNets: Efficient convolutional neural networks for mobile vision applications&lt;/a&gt;, &lt;i&gt;Howard et al., arXiv preprint arXiv:1704.04861, 2017&lt;/i&gt;&lt;/li&gt;&lt;/ol&gt;&lt;div class="feedflare"&gt;
&lt;a href="http://feeds.feedburner.com/~ff/blogspot/gJZg?a=jzLZSFviPok:OXC5NnMzlOQ:yIl2AUoC8zA"&gt;&lt;img src="http://feeds.feedburner.com/~ff/blogspot/gJZg?d=yIl2AUoC8zA" border="0"&gt;&lt;/img&gt;&lt;/a&gt;
&lt;/div&gt;&lt;img src="http://feeds.feedburner.com/~r/blogspot/gJZg/~4/jzLZSFviPok" height="1" width="1" alt=""/&gt;</content><link rel="replies" type="application/atom+xml" href="http://research.googleblog.com/feeds/776695066333371847/comments/default" title="Post Comments" /><link rel="replies" type="text/html" href="http://research.googleblog.com/2017/06/supercharge-your-computer-vision-models.html#comment-form" title="0 Comments" /><link rel="edit" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/776695066333371847" /><link rel="self" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/776695066333371847" /><link rel="alternate" type="text/html" href="http://feedproxy.google.com/~r/blogspot/gJZg/~3/jzLZSFviPok/supercharge-your-computer-vision-models.html" title="Supercharge your Computer Vision models with the TensorFlow Object Detection API" /><author><name>Research Blog</name><uri>https://plus.google.com/101673966767287570260</uri><email>noreply@blogger.com</email><gd:image rel="http://schemas.google.com/g/2005#thumbnail" width="32" height="32" src="//lh5.googleusercontent.com/-GXTJ2sdBxaM/AAAAAAAAAAI/AAAAAAAAA14/2q3ThT_9JIY/s512-c/photo.jpg" /></author><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://2.bp.blogspot.com/-MO3T6Hybpkg/WUG-QjHrHbI/AAAAAAAAB2M/tQKa2ljTkRwYgtok3o_3Y6F5KCbC7a-qQCLcBGAs/s72-c/image1.jpg" height="72" width="72" /><thr:total>0</thr:total><gd:extendedProperty name="commentSource" value="1" /><gd:extendedProperty name="commentModerationMode" value="FILTERED_POSTMOD" /><feedburner:origLink>http://research.googleblog.com/2017/06/supercharge-your-computer-vision-models.html</feedburner:origLink></entry><entry><id>tag:blogger.com,1999:blog-21224994.post-49284620403683891</id><published>2017-06-14T10:00:00.000-07:00</published><updated>2017-06-14T15:34:08.211-07:00</updated><category scheme="http://www.blogger.com/atom/ns#" term="Computer Vision" /><category scheme="http://www.blogger.com/atom/ns#" term="Machine Perception" /><category scheme="http://www.blogger.com/atom/ns#" term="Neural Networks" /><category scheme="http://www.blogger.com/atom/ns#" term="open source" /><category scheme="http://www.blogger.com/atom/ns#" term="TensorFlow" /><title type="text">MobileNets: Open-Source Models for Efficient On-Device Vision</title><content type="html">&lt;span class="byline-author"&gt;Posted by Andrew G. Howard, Senior Software Engineer and Menglong Zhu, Software Engineer&lt;/span&gt;&lt;br /&gt;&lt;br /&gt;&lt;i&gt;(Cross-posted on the &lt;a href="https://opensource.googleblog.com/2017/06/mobilenets-open-source-models-for.html"&gt;Google Open Source Blog&lt;/a&gt;)&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;Deep learning has fueled tremendous progress in the field of computer vision in recent years, with neural networks repeatedly pushing the &lt;a href="https://research.googleblog.com/2016/08/improving-inception-and-image.html"&gt;frontier of visual recognition technology&lt;/a&gt;. While many of those technologies such as object, landmark, logo and text recognition are provided for internet-connected devices through the &lt;a href="https://cloud.google.com/vision"&gt;Cloud Vision API&lt;/a&gt;, we believe that the ever-increasing computational power of mobile devices can enable the delivery of these technologies into the hands of our users, anytime, anywhere, regardless of internet connection. However, visual recognition for on device and embedded applications poses many challenges — models must run quickly with high accuracy in a resource-constrained environment making use of limited computation, power and space. &lt;br /&gt;&lt;br /&gt;Today we are pleased to announce the release of &lt;a href="https://github.com/tensorflow/models/blob/master/slim/nets/mobilenet_v1.md"&gt;MobileNets&lt;/a&gt;, a family of &lt;i&gt;mobile-first&lt;/i&gt; computer vision models for &lt;a href="https://www.tensorflow.org/"&gt;TensorFlow&lt;/a&gt;, designed to effectively maximize accuracy while being mindful of the restricted resources for an on-device or embedded application. MobileNets are small, low-latency, low-power models parameterized to meet the resource constraints of a variety of use cases. They can be built upon for classification, detection, embeddings and segmentation similar to how other popular large scale models, such as &lt;a href="https://arxiv.org/pdf/1602.07261.pdf"&gt;Inception&lt;/a&gt;, are used. &lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://3.bp.blogspot.com/-ujGePiv1gZ8/WUBjrgwrPmI/AAAAAAAAB14/zOw9URnrMnIbe7Vv8ftYT4PsnH7S-gJIQCLcBGAs/s1600/image1.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" data-original-height="395" data-original-width="1024" height="246" src="https://3.bp.blogspot.com/-ujGePiv1gZ8/WUBjrgwrPmI/AAAAAAAAB14/zOw9URnrMnIbe7Vv8ftYT4PsnH7S-gJIQCLcBGAs/s640/image1.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Example use cases include detection, fine-grain classification, attributes and geo-localization.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;This release contains the model definition for MobileNets in TensorFlow using &lt;a href="https://github.com/tensorflow/models/blob/master/inception/inception/slim/README.md"&gt;TF-Slim&lt;/a&gt;, as well as 16 pre-trained &lt;a href="http://www.image-net.org/challenges/LSVRC/"&gt;ImageNet&lt;/a&gt; classification checkpoints for use in mobile projects of all sizes. The models can be run efficiently on mobile devices with &lt;a href="https://www.tensorflow.org/mobile/"&gt;TensorFlow Mobile&lt;/a&gt;.&lt;br /&gt;&lt;div&gt;&lt;br /&gt;&lt;table border="1" cellpadding="1" cellspacing="0" style="width: 100%;"&gt;&lt;tbody&gt;&lt;tr&gt; &lt;td&gt;&lt;div style="background-color: lightblue; text-align: center;"&gt;&lt;b&gt;Model Checkpoint&lt;/b&gt;&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="background-color: lightblue; text-align: center;"&gt;&lt;b&gt;Million MACs&lt;/b&gt;&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="background-color: lightblue; text-align: center;"&gt;&lt;b&gt;Million Parameters&lt;/b&gt;&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="background-color: lightblue; text-align: center;"&gt;&lt;b&gt;Top-1 Accuracy&lt;/b&gt;&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="background-color: lightblue; text-align: center;"&gt;&lt;b&gt;Top-5 Accuracy&lt;/b&gt;&lt;/div&gt;&lt;/td&gt; &lt;/tr&gt;&lt;tr&gt; &lt;td&gt;&lt;div style="background-color: orange; text-align: center;"&gt;&lt;a href="http://download.tensorflow.org/models/mobilenet_v1_1.0_224_2017_06_14.tar.gz"&gt;MobileNet_v1_1.0_224&lt;/a&gt;&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;569&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;4.24&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;70.7&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;89.5&lt;/div&gt;&lt;/td&gt; &lt;/tr&gt;&lt;tr&gt; &lt;td&gt;&lt;div style="background-color: orange; text-align: center;"&gt;&lt;a href="http://download.tensorflow.org/models/mobilenet_v1_1.0_192_2017_06_14.tar.gz"&gt;MobileNet_v1_1.0_192&lt;/a&gt;&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;418&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;4.24&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;69.3&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;88.9&lt;/div&gt;&lt;/td&gt; &lt;/tr&gt;&lt;tr&gt; &lt;td&gt;&lt;div style="background-color: orange; text-align: center;"&gt;&lt;a href="http://download.tensorflow.org/models/mobilenet_v1_1.0_160_2017_06_14.tar.gz"&gt;MobileNet_v1_1.0_160&lt;/a&gt;&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;291&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;4.24&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;67.2&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;87.5&lt;/div&gt;&lt;/td&gt; &lt;/tr&gt;&lt;tr&gt; &lt;td&gt;&lt;div style="background-color: orange; text-align: center;"&gt;&lt;a href="http://download.tensorflow.org/models/mobilenet_v1_1.0_128_2017_06_14.tar.gz"&gt;MobileNet_v1_1.0_128&lt;/a&gt;&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;186&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;4.24&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;64.1&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;85.3&lt;/div&gt;&lt;/td&gt; &lt;/tr&gt;&lt;tr&gt; &lt;td&gt;&lt;div style="background-color: orange; text-align: center;"&gt;&lt;a href="http://http//download.tensorflow.org/models/mobilenet_v1_0.75_224_2017_06_14.tar.gz"&gt;MobileNet_v1_0.75_224&lt;/a&gt;&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;317&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;2.59&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;68.4&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;88.2&lt;/div&gt;&lt;/td&gt; &lt;/tr&gt;&lt;tr&gt; &lt;td&gt;&lt;div style="background-color: orange; text-align: center;"&gt;&lt;a href="http://download.tensorflow.org/models/mobilenet_v1_0.75_192_2017_06_14.tar.gz"&gt;MobileNet_v1_0.75_192&lt;/a&gt;&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;233&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;2.59&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;67.4&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;87.3&lt;/div&gt;&lt;/td&gt; &lt;/tr&gt;&lt;tr&gt; &lt;td&gt;&lt;div style="background-color: orange; text-align: center;"&gt;&lt;a href="http://download.tensorflow.org/models/mobilenet_v1_0.75_160_2017_06_14.tar.gz"&gt;MobileNet_v1_0.75_160&lt;/a&gt;&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;162&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;2.59&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;65.2&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;86.1&lt;/div&gt;&lt;/td&gt; &lt;/tr&gt;&lt;tr&gt; &lt;td&gt;&lt;div style="background-color: orange; text-align: center;"&gt;&lt;a href="http://download.tensorflow.org/models/mobilenet_v1_0.75_128_2017_06_14.tar.gz"&gt;MobileNet_v1_0.75_128&lt;/a&gt;&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;104&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;2.59&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;61.8&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;83.6&lt;/div&gt;&lt;/td&gt; &lt;/tr&gt;&lt;tr&gt; &lt;td&gt;&lt;div style="background-color: orange; text-align: center;"&gt;&lt;a href="http://download.tensorflow.org/models/mobilenet_v1_0.50_224_2017_06_14.tar.gz"&gt;MobileNet_v1_0.50_224&lt;/a&gt;&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;150&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;1.34&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;64.0&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;85.4&lt;/div&gt;&lt;/td&gt; &lt;/tr&gt;&lt;tr&gt; &lt;td&gt;&lt;div style="background-color: orange; text-align: center;"&gt;&lt;a href="http://download.tensorflow.org/models/mobilenet_v1_0.50_192_2017_06_14.tar.gz"&gt;MobileNet_v1_0.50_192&lt;/a&gt;&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;110&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;1.34&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;62.1&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;84.0&lt;/div&gt;&lt;/td&gt; &lt;/tr&gt;&lt;tr&gt; &lt;td&gt;&lt;div style="background-color: orange; text-align: center;"&gt;&lt;a href="http://download.tensorflow.org/models/mobilenet_v1_0.50_160_2017_06_14.tar.gz"&gt;MobileNet_v1_0.50_160&lt;/a&gt;&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;77&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;1.34&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;59.9&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;82.5&lt;/div&gt;&lt;/td&gt; &lt;/tr&gt;&lt;tr&gt; &lt;td&gt;&lt;div style="background-color: orange; text-align: center;"&gt;&lt;a href="http://download.tensorflow.org/models/mobilenet_v1_0.50_128_2017_06_14.tar.gz"&gt;MobileNet_v1_0.50_128&lt;/a&gt;&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;49&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;1.34&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;56.2&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;79.6&lt;/div&gt;&lt;/td&gt; &lt;/tr&gt;&lt;tr&gt; &lt;td&gt;&lt;div style="background-color: orange; text-align: center;"&gt;&lt;a href="http://download.tensorflow.org/models/mobilenet_v1_0.25_224_2017_06_14.tar.gz"&gt;MobileNet_v1_0.25_224&lt;/a&gt;&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;41&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;0.47&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;50.6&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;75.0&lt;/div&gt;&lt;/td&gt; &lt;/tr&gt;&lt;tr&gt; &lt;td&gt;&lt;div style="background-color: orange; text-align: center;"&gt;&lt;a href="http://download.tensorflow.org/models/mobilenet_v1_0.25_192_2017_06_14.tar.gz"&gt;MobileNet_v1_0.25_192&lt;/a&gt;&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;34&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;0.47&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;49.0&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;73.6&lt;/div&gt;&lt;/td&gt; &lt;/tr&gt;&lt;tr&gt; &lt;td&gt;&lt;div style="background-color: orange; text-align: center;"&gt;&lt;a href="http://download.tensorflow.org/models/mobilenet_v1_0.25_160_2017_06_14.tar.gz"&gt;MobileNet_v1_0.25_160&lt;/a&gt;&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;21&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;0.47&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;46.0&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;70.7&lt;/div&gt;&lt;/td&gt; &lt;/tr&gt;&lt;tr&gt; &lt;td&gt;&lt;div style="background-color: orange; text-align: center;"&gt;&lt;a href="http://download.tensorflow.org/models/mobilenet_v1_0.25_128_2017_06_14.tar.gz"&gt;MobileNet_v1_0.25_128&lt;/a&gt;&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;14&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;0.47&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;41.3&lt;/div&gt;&lt;/td&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;66.2&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;&lt;/div&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Choose the right MobileNet model to fit your latency and size budget. The size of the network in memory and on disk is proportional to the number of parameters. The latency and power usage of the network scales with the number of Multiply-Accumulates (MACs) which measures the number of fused Multiplication and Addition operations. Top-1 and Top-5 accuracies are measured on the &lt;a href="http://www.image-net.org/challenges/LSVRC/"&gt;ILSVRC dataset&lt;/a&gt;.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;We are excited to share MobileNets with the open-source community. Information for getting started can be found at the &lt;a href="https://github.com/tensorflow/models/blob/master/slim/README.md"&gt;TensorFlow-Slim Image Classification Library&lt;/a&gt;. To learn how to run models on-device please go to &lt;a href="https://www.tensorflow.org/mobile/"&gt;TensorFlow Mobile&lt;/a&gt;. You can read more about the technical details of MobileNets in our paper, &lt;a href="https://arxiv.org/abs/1704.04861"&gt;MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications&lt;/a&gt;.&lt;br /&gt;&lt;br /&gt;&lt;b&gt;Acknowledgements&lt;/b&gt;&lt;br /&gt;MobileNets were made possible with the hard work of many engineers and researchers throughout Google. Specifically we would like to thank:&lt;br /&gt;&lt;br /&gt;&lt;b&gt;Core Contributors:&lt;/b&gt; &lt;i&gt;Andrew G. Howard, Menglong Zhu, Bo Chen, Dmitry Kalenichenko, Weijun Wang, Tobias Weyand, Marco Andreetto, Hartwig Adam&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;b&gt;Special thanks to:&lt;/b&gt; &lt;i&gt;Benoit Jacob, Skirmantas Kligys, George Papandreou, Liang-Chieh Chen, Derek Chow, Sergio Guadarrama, Jonathan Huang, Andre Hentz, Pete Warden&lt;/i&gt;&lt;div class="feedflare"&gt;
&lt;a href="http://feeds.feedburner.com/~ff/blogspot/gJZg?a=U0VgnWN1kn4:9jLPOsIYGJU:yIl2AUoC8zA"&gt;&lt;img src="http://feeds.feedburner.com/~ff/blogspot/gJZg?d=yIl2AUoC8zA" border="0"&gt;&lt;/img&gt;&lt;/a&gt;
&lt;/div&gt;&lt;img src="http://feeds.feedburner.com/~r/blogspot/gJZg/~4/U0VgnWN1kn4" height="1" width="1" alt=""/&gt;</content><link rel="replies" type="application/atom+xml" href="http://research.googleblog.com/feeds/49284620403683891/comments/default" title="Post Comments" /><link rel="replies" type="text/html" href="http://research.googleblog.com/2017/06/mobilenets-open-source-models-for.html#comment-form" title="0 Comments" /><link rel="edit" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/49284620403683891" /><link rel="self" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/49284620403683891" /><link rel="alternate" type="text/html" href="http://feedproxy.google.com/~r/blogspot/gJZg/~3/U0VgnWN1kn4/mobilenets-open-source-models-for.html" title="MobileNets: Open-Source Models for Efficient On-Device Vision" /><author><name>Research Blog</name><uri>https://plus.google.com/101673966767287570260</uri><email>noreply@blogger.com</email><gd:image rel="http://schemas.google.com/g/2005#thumbnail" width="32" height="32" src="//lh5.googleusercontent.com/-GXTJ2sdBxaM/AAAAAAAAAAI/AAAAAAAAA14/2q3ThT_9JIY/s512-c/photo.jpg" /></author><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://3.bp.blogspot.com/-ujGePiv1gZ8/WUBjrgwrPmI/AAAAAAAAB14/zOw9URnrMnIbe7Vv8ftYT4PsnH7S-gJIQCLcBGAs/s72-c/image1.png" height="72" width="72" /><thr:total>0</thr:total><gd:extendedProperty name="commentSource" value="1" /><gd:extendedProperty name="commentModerationMode" value="FILTERED_POSTMOD" /><feedburner:origLink>http://research.googleblog.com/2017/06/mobilenets-open-source-models-for.html</feedburner:origLink></entry><entry><id>tag:blogger.com,1999:blog-21224994.post-6559586486192368933</id><published>2017-05-24T11:00:00.000-07:00</published><updated>2017-05-24T13:52:26.564-07:00</updated><category scheme="http://www.blogger.com/atom/ns#" term="Android" /><category scheme="http://www.blogger.com/atom/ns#" term="Gboard" /><category scheme="http://www.blogger.com/atom/ns#" term="Keyboard Input" /><category scheme="http://www.blogger.com/atom/ns#" term="Neural Networks" /><category scheme="http://www.blogger.com/atom/ns#" term="Speech Recognition" /><title type="text">The Machine Intelligence Behind Gboard</title><content type="html">&lt;span class="byline-author"&gt;Posted by Françoise Beaufays, Principal Scientist, Speech and Keyboard Team and Michael Riley, Principal Scientist, Speech and Languages Algorithms Team&lt;/span&gt;&lt;br /&gt;&lt;br /&gt;Most people spend a significant amount of time each day using mobile-device keyboards: composing emails, texting, engaging in social media, and more. Yet, mobile keyboards are still cumbersome to handle. The average user is roughly 35% slower typing on a mobile device than on a physical keyboard. To change that, we recently provided &lt;a href="https://blog.google/products/search/gboard-android-gets-new-languages-and-tools/"&gt;many exciting improvements&lt;/a&gt; to &lt;a href="https://play.google.com/store/apps/details?id=com.google.android.inputmethod.latin"&gt;Gboard for Android&lt;/a&gt;, working towards our vision of creating an intelligent mechanism that enables faster input while offering suggestions and correcting mistakes, in any language you choose. &lt;br /&gt;&lt;br /&gt;With the realization that the way a mobile keyboard translates touch inputs into text is similar to how a speech recognition system translates voice inputs into text, we leveraged our experience in Speech Recognition to pursue our vision. First, we created robust spatial models that map fuzzy sequences of raw touch points to keys on the keyboard, just like acoustic models map sequences of sound bites to phonetic units. Second, we built a powerful core decoding engine based on &lt;a href="https://en.wikipedia.org/wiki/Finite-state_transducer"&gt;finite state transducers&lt;/a&gt; (FST) to determine the likeliest word sequence given an input touch sequence. With its mathematical formalism and broad success in speech applications, we knew that an FST decoder would offer the flexibility needed to support a variety of complex keyboard input behaviors as well as language features. In this post, we will detail what went into the development of both of these systems.&lt;br /&gt;&lt;br /&gt;&lt;b&gt;Neural Spatial Models&lt;/b&gt;&lt;br /&gt;Mobile keyboard input is subject to errors that are generally attributed to “fat finger typing” (or tracing spatially similar words in glide typing, as illustrated below) along with cognitive and motor errors (manifesting in misspellings, character insertions, deletions or swaps, etc). An intelligent keyboard needs to be able to account for these errors and predict the intended words rapidly and accurately. As such, we built a spatial model for Gboard that addresses these errors at the character level, mapping the touch points on the screen to actual keys.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://4.bp.blogspot.com/-7htLmt4v5Yw/WSW8UsoXeKI/AAAAAAAAB1U/PsaG7sKYAtwUVaiTbU_qz-6xNwXzN3aOQCLcB/s1600/image2.gif" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" data-original-height="432" data-original-width="576" height="480" src="https://4.bp.blogspot.com/-7htLmt4v5Yw/WSW8UsoXeKI/AAAAAAAAB1U/PsaG7sKYAtwUVaiTbU_qz-6xNwXzN3aOQCLcB/s640/image2.gif" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Average glide trails for two spatially-similar words: “Vampire” and “Value”.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;Up to recently, Gboard used a Gaussian model to quantify the probability of tapping neighboring keys and a rule-based model to represent cognitive and motor errors. These models were simple and intuitive, but they didn’t allow us to directly optimize metrics that correlate with better typing quality. Drawing on our experience with &lt;a href="https://research.googleblog.com/2015/09/google-voice-search-faster-and-more.html"&gt;Voice Search acoustic models&lt;/a&gt; we replaced both the Gaussian and rule-based models with a single, highly efficient &lt;a href="https://en.wikipedia.org/wiki/Long_short-term_memory"&gt;long short-term memory&lt;/a&gt; (LSTM) model trained with a &lt;a href="ftp://ftp.idsia.ch/pub/juergen/icml2006.pdf"&gt;connectionist temporal classification&lt;/a&gt; (CTC) criterion.&lt;br /&gt;&lt;br /&gt;However, training this model turned out to be a lot more complicated than we had anticipated. While acoustic models are trained from human-transcribed audio data, one cannot easily transcribe millions of touch point sequences and glide traces. So the team exploited user-interaction signals, e.g. reverted auto-corrections and suggestion picks as negative and positive semi-supervised learning signals, to form rich training and test sets. &lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://2.bp.blogspot.com/-Q8bVYEwuTnw/WSW9PRtDRSI/AAAAAAAAB1c/vY4UfiCFETAAplzo1-QHW76PwkQUXvkXgCLcB/s1600/f2.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" data-original-height="728" data-original-width="1533" height="302" src="https://2.bp.blogspot.com/-Q8bVYEwuTnw/WSW9PRtDRSI/AAAAAAAAB1c/vY4UfiCFETAAplzo1-QHW76PwkQUXvkXgCLcB/s640/f2.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Raw data points corresponding to the word “could” (left), and normalized sampled trajectory with per-sample variances (right).&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;A plethora of techniques from the speech recognition literature was used to iterate on the NSM models to make them small and fast enough to be run on any device. The &lt;a href="https://www.tensorflow.org/"&gt;TensorFlow&lt;/a&gt; infrastructure was used to train hundreds of models, optimizing various signals surfaced by the keyboard: completions, suggestions, gliding, etc. After more than a year of work, the resulting models were about 6 times faster and 10 times smaller than the initial versions, they also showed about 15% reduction in bad autocorrects and 10% reduction in wrongly decoded gestures on offline datasets.&lt;br /&gt;&lt;br /&gt;&lt;b&gt;Finite-State Transducers&lt;/b&gt;&lt;br /&gt;While the NSM uses spatial information to help determine what was tapped or swiped, there are additional constraints — &lt;i&gt;lexical&lt;/i&gt; and &lt;i&gt;grammatical&lt;/i&gt; — that can be brought to bear. A lexicon tells us what words occur in a language and a probabilistic grammar tells us what words are likely to follow other words. To encode this information we use finite-state transducers. FSTs have long been a key component of Google’s speech recognition and synthesis systems. They provide a principled way to represent various probabilistic models (lexicons, grammars, normalizers, etc) used in natural language processing together with the mathematical framework needed to manipulate, optimize, combine and search the models&lt;a href="#1" name="top1"&gt;&lt;sup&gt;*&lt;/sup&gt;&lt;/a&gt;. &lt;br /&gt;&lt;br /&gt;In Gboard, a key-to-word transducer compactly represents the keyboard lexicon as shown in the figure below.  It encodes the mapping from key sequences to words, allowing for alternative key sequences and optional spaces.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://4.bp.blogspot.com/-J9yqtO7u-Nw/WSW9nXoRWiI/AAAAAAAAB1g/NLCjKa-0KSIVcQXtSTmuqdyO-xaNL0wHACLcB/s1600/image7.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" data-original-height="267" data-original-width="651" height="262" src="https://4.bp.blogspot.com/-J9yqtO7u-Nw/WSW9nXoRWiI/AAAAAAAAB1g/NLCjKa-0KSIVcQXtSTmuqdyO-xaNL0wHACLcB/s640/image7.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;This transducer encodes “I”, “I’ve”, “If” along paths from the start state (the bold circle 1) to final states (the double circle states 0 and 1). Each arc is labeled with an input key (before the “:”) and a corresponding output word (after the “:”) where ε encodes the empty symbol. The apostrophe in “I’ve” can be omitted. The user may skip the space bar sometimes. To account for that, the space key transition between words in the transducer is optional. The ε and space back arcs allow accepting more than one word.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;A probabilistic n-gram transducer is used to represent the language model for the keyboard. A state in the model represents an (up to) n-1 word context and an arc leaving that state is labeled with a successor word together with its probability of following that context (estimated from textual data). These, together with the spatial model that gives the likelihoods of sequences of key touches (discrete tap entries or continuous gestures in glide typing), are combined and explored with a&lt;a href="https://en.wikipedia.org/wiki/Beam_search"&gt; beam search&lt;/a&gt;.&lt;br /&gt;&lt;br /&gt;Generic FST principles, such as streaming, support for dynamic models, etc took us a long way towards building a new keyboard decoder, but several new functionalities also had to be added. When you speak, you don’t need the decoder to complete your words or guess what you will say next to save you a few syllables; but when you type, you appreciate the help of word completions and predictions. Also, we wanted the keyboard to provide seamless multilingual support, as shown below.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://1.bp.blogspot.com/-Oz-oMYfar8I/WSW90Jo866I/AAAAAAAAB1k/GO-9rpbpTcMhsfy3edD3lgcjXlLlTQjlwCLcB/s1600/image6.gif" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" data-original-height="800" data-original-width="600" height="640" src="https://1.bp.blogspot.com/-Oz-oMYfar8I/WSW90Jo866I/AAAAAAAAB1k/GO-9rpbpTcMhsfy3edD3lgcjXlLlTQjlwCLcB/s640/image6.gif" width="480" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Trilingual input typing in Gboard.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;It was a complex effort to get our new decoder off the ground, but the principled nature of FSTs has many benefits. For example, supporting transliterations for languages like Hindi is just a simple extension of the generic decoder.&lt;br /&gt;&lt;br /&gt;&lt;b&gt;Transliteration Models&lt;/b&gt;&lt;br /&gt;In many languages with complex scripts, romanization systems have been developed to map characters into the Latin alphabet, often according to their phonetic pronunciations. For example, the Pinyin “xièxiè” corresponds to the Chinese characters “谢谢” (“thank you”). A Pinyin keyboard allows users to conveniently type words on a QWERTY layout and have them automatically “translated” into the target script. Likewise, a transliterated Hindi keyboard allows users to type “daanth” for “दांत” (teeth). Whereas Pinyin is an agreed-upon romanization system, Hindi transliterations are more fuzzy; for example “daant” would be a valid alternative for “दांत”.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://2.bp.blogspot.com/-8CuSHDQtrd0/WSW9_xZDH3I/AAAAAAAAB1o/eDBYuZeA7okkwCBZ_8sZu0xBcQ8bbel_QCLcB/s1600/image5.gif" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" data-original-height="461" data-original-width="576" height="512" src="https://2.bp.blogspot.com/-8CuSHDQtrd0/WSW9_xZDH3I/AAAAAAAAB1o/eDBYuZeA7okkwCBZ_8sZu0xBcQ8bbel_QCLcB/s640/image5.gif" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Transliterated glide input for Hindi.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;Just as we have a transducer mapping from letter sequences to words (a lexicon) and a weighted language model automaton providing probabilities for word sequences, we built weighted transducer mappings between Latin key sequences and target script symbol sequences for 22 Indic languages. Some languages have multiple writing systems (Bodo for example can be written in the Bengali or Devanagari scripts) so between transliterated and native layouts, we built 57 new input methods in just a few months.&lt;br /&gt;&lt;br /&gt;The general nature of the FST decoder let us leverage all the work we had done to support completions, predictions, glide typing and many UI features with no extra effort, allowing us to offer a rich experience to our Indian users right from the start.&lt;br /&gt;&lt;br /&gt;&lt;b&gt;A More Intelligent Keyboard&lt;/b&gt;&lt;br /&gt;All in all, our recent work cut the decoding latency by 50%, reduced the fraction of words users have to manually correct by more than 10%, allowed us to launch transliteration support for the 22 official languages of India, and enabled many new features you may have noticed. &lt;br /&gt;&lt;br /&gt;While we hope that these recent changes improve your typing experience, we recognize that on-device typing is by no means solved. Gboard can still make suggestions that seem nonintuitive or of low utility and gestures can still be decoded to words a human would never pick. However, our shift towards powerful machine intelligence algorithms has opened new spaces that we’re actively exploring to make more useful tools and products for our users worldwide. &lt;br /&gt;&lt;br /&gt;&lt;b&gt;Acknowledgments&lt;/b&gt;&lt;br /&gt;This work was done by Cyril Allauzen, Ouais Alsharif, Lars Hellsten, Tom Ouyang, Brian Roark and David Rybach, with help from Speech Data Operation team. Special thanks go to Johan Schalkwyk and Corinna Cortes for their support.&lt;br /&gt;&lt;br /&gt;&lt;span class="Apple-style-span" style="font-size: small;"&gt;&lt;br /&gt;&lt;a name="1"&gt;&lt;b&gt;* &lt;/b&gt;&lt;/a&gt;The toolbox of relevant algorithms is available in the &lt;a href="http://www.openfst.org/"&gt;OpenFst&lt;/a&gt; open-source library.&lt;a href="#top1"&gt;&lt;sup&gt;↩&lt;/sup&gt;&lt;/a&gt;&lt;br /&gt;&lt;/span&gt;&lt;div class="feedflare"&gt;
&lt;a href="http://feeds.feedburner.com/~ff/blogspot/gJZg?a=GFnIEA0AMZI:xjlw7nEn2aY:yIl2AUoC8zA"&gt;&lt;img src="http://feeds.feedburner.com/~ff/blogspot/gJZg?d=yIl2AUoC8zA" border="0"&gt;&lt;/img&gt;&lt;/a&gt;
&lt;/div&gt;&lt;img src="http://feeds.feedburner.com/~r/blogspot/gJZg/~4/GFnIEA0AMZI" height="1" width="1" alt=""/&gt;</content><link rel="replies" type="application/atom+xml" href="http://research.googleblog.com/feeds/6559586486192368933/comments/default" title="Post Comments" /><link rel="replies" type="text/html" href="http://research.googleblog.com/2017/05/the-machine-intelligence-behind-gboard.html#comment-form" title="0 Comments" /><link rel="edit" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/6559586486192368933" /><link rel="self" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/6559586486192368933" /><link rel="alternate" type="text/html" href="http://feedproxy.google.com/~r/blogspot/gJZg/~3/GFnIEA0AMZI/the-machine-intelligence-behind-gboard.html" title="The Machine Intelligence Behind Gboard" /><author><name>Research Blog</name><uri>https://plus.google.com/101673966767287570260</uri><email>noreply@blogger.com</email><gd:image rel="http://schemas.google.com/g/2005#thumbnail" width="32" height="32" src="//lh5.googleusercontent.com/-GXTJ2sdBxaM/AAAAAAAAAAI/AAAAAAAAA14/2q3ThT_9JIY/s512-c/photo.jpg" /></author><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://4.bp.blogspot.com/-7htLmt4v5Yw/WSW8UsoXeKI/AAAAAAAAB1U/PsaG7sKYAtwUVaiTbU_qz-6xNwXzN3aOQCLcB/s72-c/image2.gif" height="72" width="72" /><thr:total>0</thr:total><gd:extendedProperty name="commentSource" value="1" /><gd:extendedProperty name="commentModerationMode" value="FILTERED_POSTMOD" /><feedburner:origLink>http://research.googleblog.com/2017/05/the-machine-intelligence-behind-gboard.html</feedburner:origLink></entry><entry><id>tag:blogger.com,1999:blog-21224994.post-3118081168848036629</id><published>2017-05-17T11:30:00.000-07:00</published><updated>2017-05-17T11:54:21.814-07:00</updated><category scheme="http://www.blogger.com/atom/ns#" term="Deep Learning" /><category scheme="http://www.blogger.com/atom/ns#" term="distributed systems" /><category scheme="http://www.blogger.com/atom/ns#" term="Machine Learning" /><category scheme="http://www.blogger.com/atom/ns#" term="Research" /><category scheme="http://www.blogger.com/atom/ns#" term="TensorFlow" /><category scheme="http://www.blogger.com/atom/ns#" term="TPU" /><title type="text">Introducing the TensorFlow Research Cloud</title><content type="html">&lt;span class="byline-author"&gt;Posted by Zak Stone, Product Manager for TensorFlow&lt;/span&gt;&lt;br /&gt;&lt;br /&gt;Researchers require enormous computational resources to train the machine learning (ML) models that have delivered recent breakthroughs in &lt;a href="https://research.googleblog.com/2017/03/assisting-pathologists-in-detecting.html"&gt;medical imaging&lt;/a&gt;, &lt;a href="https://research.googleblog.com/2016/09/a-neural-network-for-machine.html"&gt;neural machine translation&lt;/a&gt;, &lt;a href="https://deepmind.com/research/alphago/"&gt;game playing&lt;/a&gt;, and many other domains. We believe that significantly larger amounts of computation will make it possible for researchers to invent new types of ML models that will be even more accurate and useful. &lt;br /&gt;&lt;br /&gt;To accelerate the pace of open machine-learning research, we are introducing the &lt;a href="https://www.tensorflow.org/tfrc"&gt;TensorFlow Research Cloud&lt;/a&gt; (TFRC), a cluster of 1,000 &lt;a href="https://cloud.google.com/tpu/"&gt;Cloud TPUs&lt;/a&gt; that will be made available free of charge to support a broad range of computationally-intensive research projects that might not be possible otherwise.&lt;br /&gt;&lt;div class="separator" style="clear: both; text-align: center;"&gt;&lt;a href="https://2.bp.blogspot.com/-xojf3dn8Ngc/WRubNXxUZJI/AAAAAAAAB1A/0W7o1hR_n20QcWyXHXDI1OTo7vXBR8f7QCLcB/s1600/image2.png" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"&gt;&lt;img border="0" height="400" src="https://2.bp.blogspot.com/-xojf3dn8Ngc/WRubNXxUZJI/AAAAAAAAB1A/0W7o1hR_n20QcWyXHXDI1OTo7vXBR8f7QCLcB/s400/image2.png" width="400" /&gt;&lt;/a&gt;&lt;/div&gt;The TensorFlow Research Cloud offers researchers the following benefits:&lt;br /&gt;&lt;ul&gt;&lt;li&gt;Access to Google’s all-new &lt;a href="https://cloud.google.com/tpu/"&gt;Cloud TPUs&lt;/a&gt; that accelerate both training &lt;i&gt;and&lt;/i&gt; inference&lt;/li&gt;&lt;li&gt;Up to 180 teraflops of floating-point performance per Cloud TPU&lt;/li&gt;&lt;li&gt;64 GB of ultra-high-bandwidth memory per Cloud TPU&lt;/li&gt;&lt;li&gt;Familiar TensorFlow programming interfaces&lt;/li&gt;&lt;/ul&gt;You can &lt;a href="https://services.google.com/fb/forms/tpusignup/"&gt;sign up here&lt;/a&gt; to request to be notified when the TensorFlow Research Cloud application process opens, and you can optionally share more information about your computational needs. We plan to evaluate applications on a rolling basis in search of the most creative and ambitious proposals.&lt;br /&gt;&lt;br /&gt;The TensorFlow Research Cloud program is not limited to academia — we recognize that people with a wide range of affiliations, roles, and expertise are making major machine learning research contributions, and we especially encourage those with non-traditional backgrounds to apply. Access will be granted to selected individuals for limited amounts of compute time, and researchers are welcome to apply multiple times with multiple projects.&lt;br /&gt;&lt;div class="separator" style="clear: both; text-align: center;"&gt;&lt;a href="https://3.bp.blogspot.com/-JUR4rwPNfLc/WRubZSqnXTI/AAAAAAAAB1E/UDYOeQEJYwA362ko7ZptX0ROSvLsloqUgCLcB/s1600/image1.jpg" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"&gt;&lt;img border="0" height="425" src="https://3.bp.blogspot.com/-JUR4rwPNfLc/WRubZSqnXTI/AAAAAAAAB1E/UDYOeQEJYwA362ko7ZptX0ROSvLsloqUgCLcB/s640/image1.jpg" width="640" /&gt;&lt;/a&gt;&lt;/div&gt;Since the main goal of the TensorFlow Research Cloud is to benefit the open machine learning research community as a whole, successful applicants will be expected to do the following:&lt;br /&gt;&lt;ul&gt;&lt;li&gt;Share their TFRC-supported research with the world through peer-reviewed publications, open-source code, blog posts, or other open media&lt;/li&gt;&lt;li&gt;Share concrete, constructive feedback with Google to help us improve the TFRC program and the underlying Cloud TPU platform over time&lt;/li&gt;&lt;li&gt;Imagine a future in which ML acceleration is abundant and develop new kinds of machine learning models in anticipation of that future&lt;/li&gt;&lt;/ul&gt;For businesses interested in using Cloud TPUs for proprietary research and development, we will offer a parallel Cloud TPU Alpha program. You can &lt;a href="https://services.google.com/fb/forms/tpusignup/"&gt;sign up here&lt;/a&gt; to learn more about this program. We recommend participating in the Cloud TPU Alpha program if you are interested in any of the following:&lt;br /&gt;&lt;ul&gt;&lt;li&gt;Accelerating training of proprietary ML models; models that take weeks to train on other hardware can be trained in days or even hours on Cloud TPUs&lt;/li&gt;&lt;li&gt;Accelerating batch processing of industrial-scale datasets: images, videos, audio, unstructured text, structured data, etc.&lt;/li&gt;&lt;li&gt;Processing live requests in production using larger and more complex ML models than ever before&lt;/li&gt;&lt;/ul&gt;We hope the &lt;a href="https://www.tensorflow.org/tfrc"&gt;TensorFlow Research Cloud&lt;/a&gt; will allow as many researchers as possible to explore the frontier of machine learning research and extend it with new discoveries! We encourage you to &lt;a href="https://services.google.com/fb/forms/tpusignup"&gt;sign up today&lt;/a&gt; to be among the first to know as more information becomes available.&lt;div class="feedflare"&gt;
&lt;a href="http://feeds.feedburner.com/~ff/blogspot/gJZg?a=x4guySNUTEU:f-UEUXiO_jU:yIl2AUoC8zA"&gt;&lt;img src="http://feeds.feedburner.com/~ff/blogspot/gJZg?d=yIl2AUoC8zA" border="0"&gt;&lt;/img&gt;&lt;/a&gt;
&lt;/div&gt;&lt;img src="http://feeds.feedburner.com/~r/blogspot/gJZg/~4/x4guySNUTEU" height="1" width="1" alt=""/&gt;</content><link rel="replies" type="application/atom+xml" href="http://research.googleblog.com/feeds/3118081168848036629/comments/default" title="Post Comments" /><link rel="replies" type="text/html" href="http://research.googleblog.com/2017/05/introducing-tensorflow-research-cloud.html#comment-form" title="0 Comments" /><link rel="edit" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/3118081168848036629" /><link rel="self" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/3118081168848036629" /><link rel="alternate" type="text/html" href="http://feedproxy.google.com/~r/blogspot/gJZg/~3/x4guySNUTEU/introducing-tensorflow-research-cloud.html" title="Introducing the TensorFlow Research Cloud" /><author><name>Research Blog</name><uri>https://plus.google.com/101673966767287570260</uri><email>noreply@blogger.com</email><gd:image rel="http://schemas.google.com/g/2005#thumbnail" width="32" height="32" src="//lh5.googleusercontent.com/-GXTJ2sdBxaM/AAAAAAAAAAI/AAAAAAAAA14/2q3ThT_9JIY/s512-c/photo.jpg" /></author><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://2.bp.blogspot.com/-xojf3dn8Ngc/WRubNXxUZJI/AAAAAAAAB1A/0W7o1hR_n20QcWyXHXDI1OTo7vXBR8f7QCLcB/s72-c/image2.png" height="72" width="72" /><thr:total>0</thr:total><gd:extendedProperty name="commentSource" value="1" /><gd:extendedProperty name="commentModerationMode" value="FILTERED_POSTMOD" /><feedburner:origLink>http://research.googleblog.com/2017/05/introducing-tensorflow-research-cloud.html</feedburner:origLink></entry><entry><id>tag:blogger.com,1999:blog-21224994.post-6067111607818350524</id><published>2017-05-17T10:00:00.001-07:00</published><updated>2017-05-17T10:40:59.829-07:00</updated><category scheme="http://www.blogger.com/atom/ns#" term="Deep Learning" /><category scheme="http://www.blogger.com/atom/ns#" term="Google Brain" /><category scheme="http://www.blogger.com/atom/ns#" term="Machine Learning" /><title type="text">Using Machine Learning to Explore Neural Network Architecture</title><content type="html">&lt;span class="byline-author"&gt;Posted by Quoc Le &amp;amp; Barret Zoph, Research Scientists, Google Brain team&lt;/span&gt;&lt;br /&gt;&lt;br /&gt;At Google, we have successfully applied deep learning models to many applications, from &lt;a href="https://research.googleblog.com/2014/09/building-deeper-understanding-of-images.html"&gt;image recognition&lt;/a&gt; to &lt;a href="https://research.googleblog.com/2012/08/speech-recognition-and-deep-learning.html"&gt;speech recognition&lt;/a&gt; to &lt;a href="https://research.googleblog.com/2016/09/a-neural-network-for-machine.html"&gt;machine translation&lt;/a&gt;. Typically, our machine learning models are painstakingly designed by a team of engineers and scientists. This process of manually designing machine learning models is difficult because the search space of all possible models can be combinatorially large — a typical 10-layer network can have ~10&lt;sup&gt;10&lt;/sup&gt; candidate networks! For this reason, the process of designing networks often takes a significant amount of time and experimentation by those with significant machine learning expertise. &lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://3.bp.blogspot.com/-8Lsg0rnxl7k/WRtttN18MKI/AAAAAAAAB0o/KpHbFnYBmTYQ3dBjVLimPUkKphU_qLBfgCLcB/s1600/image2.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="188" src="https://3.bp.blogspot.com/-8Lsg0rnxl7k/WRtttN18MKI/AAAAAAAAB0o/KpHbFnYBmTYQ3dBjVLimPUkKphU_qLBfgCLcB/s640/image2.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Our &lt;a href="https://static.googleusercontent.com/media/research.google.com/en//pubs/archive/43022.pdf"&gt;GoogleNet&lt;/a&gt; architecture.  Design of this network required many years of careful experimentation and refinement from initial versions of convolutional architectures.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;To make this process of designing machine learning models much more accessible, we’ve been exploring ways to automate the design of machine learning models. Among many algorithms we’ve studied, &lt;a href="https://arxiv.org/abs/1703.01041"&gt;evolutionary algorithms&lt;/a&gt; [1] and &lt;a href="https://arxiv.org/abs/1611.01578"&gt;reinforcement learning algorithms&lt;/a&gt; [2] have shown great promise. But in this blog post, we’ll focus on our reinforcement learning approach and the early results we’ve gotten so far.&lt;br /&gt;&lt;br /&gt;In our approach (which we call "AutoML"), a controller neural net can propose a “child” model architecture, which can then be trained and evaluated for quality on a particular task. That feedback is then used to inform the controller how to improve its proposals for the next round. We repeat this process thousands of times — generating new architectures, testing them, and giving that feedback to the controller to learn from. Eventually the controller learns to assign high probability to areas of architecture space that achieve better accuracy on a held-out validation dataset, and low probability to areas of architecture space that score poorly. Here’s what the process looks like:&lt;br /&gt;&lt;div class="separator" style="clear: both; text-align: center;"&gt;&lt;a href="https://1.bp.blogspot.com/-0nzARW3QtkA/WRtuVsUJ02I/AAAAAAAAB0s/t6ncpAH6VfIzkr2tWW8CnE6U2Es2Bs1BgCLcB/s1600/image3.png" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"&gt;&lt;img border="0" height="290" src="https://1.bp.blogspot.com/-0nzARW3QtkA/WRtuVsUJ02I/AAAAAAAAB0s/t6ncpAH6VfIzkr2tWW8CnE6U2Es2Bs1BgCLcB/s640/image3.png" width="640" /&gt;&lt;/a&gt;&lt;/div&gt;We’ve applied this approach to two heavily benchmarked datasets in deep learning: image recognition with &lt;a href="https://www.cs.toronto.edu/~kriz/cifar.html"&gt;CIFAR-10&lt;/a&gt; and language modeling with &lt;a href="https://catalog.ldc.upenn.edu/ldc99t42"&gt;Penn Treebank&lt;/a&gt;. On both datasets, our approach can design models that achieve accuracies on par with state-of-art models designed by machine learning experts (including some on our own team!).&lt;br /&gt;&lt;br /&gt;So, what kind of neural nets does it produce? Let’s take one example: a recurrent architecture that’s trained to predict the next word on the Penn Treebank dataset. On the left here is a neural net designed by human experts. On the right is a recurrent architecture created by our method: &lt;br /&gt;&lt;div class="separator" style="clear: both; text-align: center;"&gt;&lt;a href="https://3.bp.blogspot.com/-i_T7k1EMd9w/WRtucgbRx1I/AAAAAAAAB0w/ZKxPOsmZCt4hnqCT0nhhWyHRFdD-xssUgCLcB/s1600/image1.png" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"&gt;&lt;img border="0" height="298" src="https://3.bp.blogspot.com/-i_T7k1EMd9w/WRtucgbRx1I/AAAAAAAAB0w/ZKxPOsmZCt4hnqCT0nhhWyHRFdD-xssUgCLcB/s640/image1.png" width="640" /&gt;&lt;/a&gt;&lt;/div&gt;&lt;br /&gt;The machine-chosen architecture does share some common features with the human design, such as using addition to combine input and previous hidden states. However, there are some notable new elements — for example, the machine-chosen architecture incorporates a multiplicative combination (the left-most blue node on the right diagram labeled “&lt;i&gt;elem_mult&lt;/i&gt;”). This type of combination is not common for recurrent networks, perhaps because researchers see no obvious benefit for having it. Interestingly, a simpler form of this approach was &lt;a href="https://arxiv.org/pdf/1606.06630.pdf"&gt;recently suggested&lt;/a&gt; by human designers, who also argued that this multiplicative combination can actually alleviate gradient vanishing/exploding issues, suggesting that the machine-chosen architecture was able to discover a useful new neural net architecture.&lt;br /&gt;&lt;br /&gt;This approach may also teach us something about why certain types of neural nets work so well. The architecture on the right here has many channels so that the gradient can flow backwards, which may help explain why &lt;a href="https://en.wikipedia.org/wiki/Long_short-term_memory"&gt;LSTM RNNs&lt;/a&gt; work better than standard &lt;a href="https://en.wikipedia.org/wiki/Recurrent_neural_network"&gt;RNNs&lt;/a&gt;. &lt;br /&gt;&lt;br /&gt;Going forward, we’ll work on careful analysis and testing of these machine-generated architectures to help refine our understanding of them. If we succeed, we think this can inspire new types of neural nets and make it possible for non-experts to create neural nets tailored to their particular needs, allowing machine learning to have a greater impact to everyone.&lt;br /&gt;&lt;br /&gt;&lt;b&gt;References&lt;/b&gt;&lt;br /&gt;&lt;b&gt;&lt;br /&gt;&lt;/b&gt; [1] &lt;a href="https://arxiv.org/abs/1703.01041"&gt;Large-Scale Evolution of Image Classifiers&lt;/a&gt;, &lt;i&gt;Esteban Real, Sherry Moore, Andrew Selle, Saurabh Saxena, Yutaka Leon Suematsu, Quoc Le, Alex Kurakin. International Conference on Machine Learning, 2017.&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;[2] &lt;a href="https://arxiv.org/abs/1611.01578"&gt;Neural Architecture Search with Reinforcement Learning&lt;/a&gt;, &lt;i&gt;Barret Zoph, Quoc V. Le. International Conference on Learning Representations, 2017.&lt;/i&gt;&lt;div class="feedflare"&gt;
&lt;a href="http://feeds.feedburner.com/~ff/blogspot/gJZg?a=MpgtmrorI_g:ePTLM9-U4CQ:yIl2AUoC8zA"&gt;&lt;img src="http://feeds.feedburner.com/~ff/blogspot/gJZg?d=yIl2AUoC8zA" border="0"&gt;&lt;/img&gt;&lt;/a&gt;
&lt;/div&gt;&lt;img src="http://feeds.feedburner.com/~r/blogspot/gJZg/~4/MpgtmrorI_g" height="1" width="1" alt=""/&gt;</content><link rel="replies" type="application/atom+xml" href="http://research.googleblog.com/feeds/6067111607818350524/comments/default" title="Post Comments" /><link rel="replies" type="text/html" href="http://research.googleblog.com/2017/05/using-machine-learning-to-explore.html#comment-form" title="0 Comments" /><link rel="edit" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/6067111607818350524" /><link rel="self" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/6067111607818350524" /><link rel="alternate" type="text/html" href="http://feedproxy.google.com/~r/blogspot/gJZg/~3/MpgtmrorI_g/using-machine-learning-to-explore.html" title="Using Machine Learning to Explore Neural Network Architecture" /><author><name>Research Blog</name><uri>https://plus.google.com/101673966767287570260</uri><email>noreply@blogger.com</email><gd:image rel="http://schemas.google.com/g/2005#thumbnail" width="32" height="32" src="//lh5.googleusercontent.com/-GXTJ2sdBxaM/AAAAAAAAAAI/AAAAAAAAA14/2q3ThT_9JIY/s512-c/photo.jpg" /></author><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://3.bp.blogspot.com/-8Lsg0rnxl7k/WRtttN18MKI/AAAAAAAAB0o/KpHbFnYBmTYQ3dBjVLimPUkKphU_qLBfgCLcB/s72-c/image2.png" height="72" width="72" /><thr:total>0</thr:total><gd:extendedProperty name="commentSource" value="1" /><gd:extendedProperty name="commentModerationMode" value="FILTERED_POSTMOD" /><feedburner:origLink>http://research.googleblog.com/2017/05/using-machine-learning-to-explore.html</feedburner:origLink></entry><entry><id>tag:blogger.com,1999:blog-21224994.post-3650479443015109456</id><published>2017-05-17T10:00:00.000-07:00</published><updated>2017-05-17T10:11:15.452-07:00</updated><category scheme="http://www.blogger.com/atom/ns#" term="Deep Learning" /><category scheme="http://www.blogger.com/atom/ns#" term="Gmail" /><category scheme="http://www.blogger.com/atom/ns#" term="Natural Language Processing" /><category scheme="http://www.blogger.com/atom/ns#" term="Natural Language Understanding" /><title type="text">Efficient Smart Reply, now for Gmail</title><content type="html">&lt;span class="byline-author"&gt;Posted by Brian Strope, Research Scientist, and Ray Kurzweil, Engineering Director, Google Research&lt;/span&gt;&lt;br /&gt;&lt;br /&gt;Last year we launched Smart Reply, a feature for &lt;a href="https://inbox.google.com/"&gt;Inbox by Gmail&lt;/a&gt; that uses &lt;a href="https://research.googleblog.com/2015/11/computer-respond-to-this-email.html"&gt;machine learning to suggest replies to email&lt;/a&gt;. Since the initial release, usage of Smart Reply has grown significantly, making up about 12% of replies in Inbox on mobile. Based on our examination of the use of Smart Reply in Inbox and our ideas about how humans learn and use language, we have created a new version of &lt;a href="https://blog.google/products/gmail/save-time-with-smart-reply-in-gmail"&gt;Smart Reply for Gmail&lt;/a&gt;. This version increases the percentage of usable suggestions and is more algorithmically efficient. &lt;br /&gt;&lt;br /&gt;&lt;b&gt;Novel thinking: hierarchy&lt;/b&gt;&lt;br /&gt;Inspired by how humans understand languages and concepts, we turned to hierarchical models of language, an approach that uses &lt;a href="https://en.wikipedia.org/wiki/How_to_Create_a_Mind"&gt;hierarchies of modules, each of which can learn, remember, and recognize a sequential pattern&lt;/a&gt;. &lt;br /&gt;&lt;br /&gt;The content of language is deeply hierarchical, reflected in the structure of language itself, going from letters to words to phrases to sentences to paragraphs to sections to chapters to books to authors to libraries, etc. Consider the message, "That interesting person at the cafe we like gave me a glance."  The hierarchical chunks in this sentence are highly variable. The subject of the sentence is "That interesting person at the cafe we like." The modifier "interesting" tells us something about the writer's past experiences with the person. We are told that the location of an incident involving both the writer and the person is "at the cafe." We are also told that "we," meaning the writer and the person being written to, like the cafe. Additionally, each word is itself part of a hierarchy, sometimes more than one. A cafe is a type of restaurant which is a type of store which is a type of establishment, and so on.  &lt;br /&gt;&lt;br /&gt;In proposing an appropriate response to this message we might consider the meaning of the word "glance," which is potentially ambiguous. Was it a positive gesture? In that case, we might respond, "Cool!" Or was it a negative gesture? If so, does the subject say anything about how the writer felt about the negative exchange? A lot of information about the world, and an ability to make reasoned judgments, are needed to make subtle distinctions.&lt;br /&gt;&lt;br /&gt;Given enough examples of language, a machine learning approach can discover many of these subtle distinctions. Moreover, a hierarchical approach to learning is well suited to the hierarchical nature of language. We have found that this approach works well for suggesting possible responses to emails. We use a hierarchy of modules, each of which considers features that correspond to sequences at different temporal scales, similar to how we understand speech and language. &lt;br /&gt;&lt;div class="separator" style="clear: both; text-align: center;"&gt;&lt;a href="https://4.bp.blogspot.com/-sOv8w9Ckt0Y/WRoYwN8tbEI/AAAAAAAAB0E/08ZM17H6UTogyteGE8e-ImRUwYJZvJLiwCLcB/s1600/image1.png" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"&gt;&lt;img border="0" height="351" src="https://4.bp.blogspot.com/-sOv8w9Ckt0Y/WRoYwN8tbEI/AAAAAAAAB0E/08ZM17H6UTogyteGE8e-ImRUwYJZvJLiwCLcB/s400/image1.png" width="400" /&gt;&lt;/a&gt;&lt;/div&gt;Each module processes inputs and provides transformed representations of those inputs on its outputs (which are, in turn, available for the next level). In the Smart Reply system, and the figure above, the repeated structure has two layers of hierarchy. The first makes each feature useful as a predictor of the final result, and the second combines these features. By definition, the second works at a more abstract representation and considers a wider timescale.&lt;br /&gt;&lt;br /&gt;By comparison, the initial release of Smart Reply encoded input emails word-by-word with a &lt;a href="https://en.wikipedia.org/wiki/Long_short-term_memory"&gt;long-short-term-memory&lt;/a&gt; (LSTM) recurrent neural network, and then decoded potential replies with yet another word-level LSTM. While this type of modeling is very effective in many contexts, even with Google infrastructure, it’s an approach that requires substantial computation resources. Instead of working word-by-word, we found an effective and highly efficient path by processing the problem more all-at-once, by comparing a simple hierarchy of vector representations of multiple features corresponding to longer time spans.&lt;br /&gt;&lt;br /&gt;&lt;b&gt;Semantics&lt;/b&gt;&lt;br /&gt;We have also considered whether the mathematical space of these vector representations is implicitly semantic. Do the hierarchical network representations reflect a coarse “understanding” of the actual meaning of the inputs and the responses in order to determine which go together, or do they reflect more consistent syntactical patterns? Given many real examples of which pairs go together and, perhaps more importantly which do not, we found that our networks are surprisingly effective and efficient at deriving representations that meet the training requirements. &lt;br /&gt;&lt;div class="separator" style="clear: both; text-align: center;"&gt;&lt;a href="https://1.bp.blogspot.com/-Lz7F4VVP8iM/WRoYVJ9YxfI/AAAAAAAAB0A/PpvcXHmVVBw71VKK5GUKXh8RUEkBKkNSwCLcB/s1600/Side-By-Side-v2.png" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"&gt;&lt;img border="0" height="600" src="https://1.bp.blogspot.com/-Lz7F4VVP8iM/WRoYVJ9YxfI/AAAAAAAAB0A/PpvcXHmVVBw71VKK5GUKXh8RUEkBKkNSwCLcB/s640/Side-By-Side-v2.png" width="640" /&gt;&lt;/a&gt;&lt;/div&gt;So far we see that the system can find responses that are on point, without an overlap of keywords or even synonyms of keywords.More directly, we’re delighted when the system suggests results that show understanding and are helpful.&lt;br /&gt;&lt;br /&gt;The key to this work is the confidence and trust people give us when they use the Smart Reply feature. As always, thank you for showing us the ways that work (and the ways that don’t!). With your help, we’ll do our best to keep learning.&lt;div class="feedflare"&gt;
&lt;a href="http://feeds.feedburner.com/~ff/blogspot/gJZg?a=iy2P2GkAo5o:0oc46H1vEhs:yIl2AUoC8zA"&gt;&lt;img src="http://feeds.feedburner.com/~ff/blogspot/gJZg?d=yIl2AUoC8zA" border="0"&gt;&lt;/img&gt;&lt;/a&gt;
&lt;/div&gt;&lt;img src="http://feeds.feedburner.com/~r/blogspot/gJZg/~4/iy2P2GkAo5o" height="1" width="1" alt=""/&gt;</content><link rel="replies" type="application/atom+xml" href="http://research.googleblog.com/feeds/3650479443015109456/comments/default" title="Post Comments" /><link rel="replies" type="text/html" href="http://research.googleblog.com/2017/05/efficient-smart-reply-now-for-gmail.html#comment-form" title="0 Comments" /><link rel="edit" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/3650479443015109456" /><link rel="self" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/3650479443015109456" /><link rel="alternate" type="text/html" href="http://feedproxy.google.com/~r/blogspot/gJZg/~3/iy2P2GkAo5o/efficient-smart-reply-now-for-gmail.html" title="Efficient Smart Reply, now for Gmail" /><author><name>Research Blog</name><uri>https://plus.google.com/101673966767287570260</uri><email>noreply@blogger.com</email><gd:image rel="http://schemas.google.com/g/2005#thumbnail" width="32" height="32" src="//lh5.googleusercontent.com/-GXTJ2sdBxaM/AAAAAAAAAAI/AAAAAAAAA14/2q3ThT_9JIY/s512-c/photo.jpg" /></author><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://4.bp.blogspot.com/-sOv8w9Ckt0Y/WRoYwN8tbEI/AAAAAAAAB0E/08ZM17H6UTogyteGE8e-ImRUwYJZvJLiwCLcB/s72-c/image1.png" height="72" width="72" /><thr:total>0</thr:total><gd:extendedProperty name="commentSource" value="1" /><gd:extendedProperty name="commentModerationMode" value="FILTERED_POSTMOD" /><feedburner:origLink>http://research.googleblog.com/2017/05/efficient-smart-reply-now-for-gmail.html</feedburner:origLink></entry><entry><id>tag:blogger.com,1999:blog-21224994.post-6281556223724387023</id><published>2017-05-16T11:00:00.000-07:00</published><updated>2017-05-16T11:00:24.864-07:00</updated><category scheme="http://www.blogger.com/atom/ns#" term="datasets" /><category scheme="http://www.blogger.com/atom/ns#" term="Machine Learning" /><category scheme="http://www.blogger.com/atom/ns#" term="Natural Language Processing" /><category scheme="http://www.blogger.com/atom/ns#" term="Publications" /><category scheme="http://www.blogger.com/atom/ns#" term="Research" /><category scheme="http://www.blogger.com/atom/ns#" term="Social Networks" /><title type="text">Coarse Discourse: A Dataset for Understanding Online Discussions</title><content type="html">&lt;span class="byline-author"&gt;Posted by Praveen Paritosh, Senior Research Scientist, Ka Wong, Senior Data Scientist&lt;/span&gt;&lt;br /&gt;&lt;br /&gt;Every day, participants of online communities form and share their opinions, experiences, advice and social support, most of which is expressed freely and without much constraint. These online discussions are often a key resource of information for many important topics, such as parenting, fitness, travel and more. However, these discussions also are intermixed with a clutter of disagreements, humor, flame wars and trolling, requiring readers to filter the content before getting the information they are looking for. And while the field of &lt;a href="https://en.wikipedia.org/wiki/Information_retrieval"&gt;Information Retrieval&lt;/a&gt; actively explores ways to allow users to more efficiently find, navigate and consume this content, there is a lack of shared datasets on forum discussions to aid in understanding these discussions a bit better.  &lt;br /&gt;&lt;br /&gt;To aid researchers in this space, we are releasing the &lt;a href="https://github.com/google-research-datasets/coarse-discourse"&gt;Coarse Discourse dataset&lt;/a&gt;, the largest dataset of annotated online discussions to date. The Coarse Discourse contains over half a million human annotations of publicly available online discussions on a random sample of over 9,000 threads from 130 communities from &lt;a href="http://www.reddit.com/"&gt;reddit.com&lt;/a&gt;. &lt;br /&gt;&lt;br /&gt;To create this dataset, we developed the Coarse Discourse taxonomy of forum comments by going through a small set of forum threads, reading every comment, and deciding what role the comments played in the discussion. We then repeated and revised this exercise with crowdsourced human editors to validate the reproducibility of the taxonomy's discourse types, which include: announcement, question, answer, agreement, disagreement, appreciation, negative reaction, elaboration, and humor.  From this data, over 100,000 comments were independently annotated by the crowdsourced editors for discourse type and relation. Along with the raw annotations from crowdsourced editors, we also provide the &lt;a href="https://github.com/google-research-datasets/coarse-discourse/blob/master/rating_guidelines.pdf"&gt;Coarse Discourse annotation task guidelines&lt;/a&gt; used by the editors to help with collecting data for other forums and refining the task further.  &lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://2.bp.blogspot.com/-Pc7dQZFAgO0/WRsz6MuGpCI/AAAAAAAAB0Y/pDheu9Y8mQ4hGGOp7cA5lIvdeBTBjFKJwCLcB/s1600/image1.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="480" src="https://2.bp.blogspot.com/-Pc7dQZFAgO0/WRsz6MuGpCI/AAAAAAAAB0Y/pDheu9Y8mQ4hGGOp7cA5lIvdeBTBjFKJwCLcB/s640/image1.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;An example thread annotated with discourse types and relations. &lt;a href="https://research.google.com/pubs/pub46055.html"&gt;Early findings&lt;/a&gt; suggest that question answering is a prominent use case in most communities, while some communities are more converationally focused, with back-and-forth interactions. &lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;For machine learning and natural language processing researchers trying to characterize the nature of online discussions, we hope that this dataset is a useful resource. Visit our &lt;a href="https://github.com/google-research-datasets/coarse-discourse"&gt;GitHub repository&lt;/a&gt; to download the data. For more details, check out our &lt;a href="http://www.icwsm.org/2017/index.php"&gt;ICWSM&lt;/a&gt; paper, “&lt;a href="https://research.google.com/pubs/pub46055.html"&gt;Characterizing Online Discussion Using Coarse Discourse Sequences&lt;/a&gt;.”&lt;br /&gt;&lt;br /&gt;&lt;b&gt;Acknowledgments&lt;/b&gt;&lt;br /&gt;This work was done by Amy Zhang during her internship at Google. We would also like to thank Bryan Culbertson, Olivia Rhinehart, Eric Altendorf, David Huynh, Nancy Chang, Chris Welty and our crowdsourced editors. &lt;div class="feedflare"&gt;
&lt;a href="http://feeds.feedburner.com/~ff/blogspot/gJZg?a=DxG6qAck2OM:wRJtGAvCBGo:yIl2AUoC8zA"&gt;&lt;img src="http://feeds.feedburner.com/~ff/blogspot/gJZg?d=yIl2AUoC8zA" border="0"&gt;&lt;/img&gt;&lt;/a&gt;
&lt;/div&gt;&lt;img src="http://feeds.feedburner.com/~r/blogspot/gJZg/~4/DxG6qAck2OM" height="1" width="1" alt=""/&gt;</content><link rel="replies" type="application/atom+xml" href="http://research.googleblog.com/feeds/6281556223724387023/comments/default" title="Post Comments" /><link rel="replies" type="text/html" href="http://research.googleblog.com/2017/05/coarse-discourse-dataset-for.html#comment-form" title="0 Comments" /><link rel="edit" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/6281556223724387023" /><link rel="self" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/6281556223724387023" /><link rel="alternate" type="text/html" href="http://feedproxy.google.com/~r/blogspot/gJZg/~3/DxG6qAck2OM/coarse-discourse-dataset-for.html" title="Coarse Discourse: A Dataset for Understanding Online Discussions" /><author><name>Research Blog</name><uri>https://plus.google.com/101673966767287570260</uri><email>noreply@blogger.com</email><gd:image rel="http://schemas.google.com/g/2005#thumbnail" width="32" height="32" src="//lh5.googleusercontent.com/-GXTJ2sdBxaM/AAAAAAAAAAI/AAAAAAAAA14/2q3ThT_9JIY/s512-c/photo.jpg" /></author><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://2.bp.blogspot.com/-Pc7dQZFAgO0/WRsz6MuGpCI/AAAAAAAAB0Y/pDheu9Y8mQ4hGGOp7cA5lIvdeBTBjFKJwCLcB/s72-c/image1.png" height="72" width="72" /><thr:total>0</thr:total><gd:extendedProperty name="commentSource" value="1" /><gd:extendedProperty name="commentModerationMode" value="FILTERED_POSTMOD" /><feedburner:origLink>http://research.googleblog.com/2017/05/coarse-discourse-dataset-for.html</feedburner:origLink></entry><entry><id>tag:blogger.com,1999:blog-21224994.post-827399624362227391</id><published>2017-05-11T06:00:00.000-07:00</published><updated>2017-05-11T09:28:35.532-07:00</updated><category scheme="http://www.blogger.com/atom/ns#" term="Art" /><category scheme="http://www.blogger.com/atom/ns#" term="Computer Vision" /><category scheme="http://www.blogger.com/atom/ns#" term="Supervised Learning" /><title type="text">Neural Network-Generated Illustrations in Allo</title><content type="html">&lt;span class="byline-author"&gt;Posted by Jennifer Daniel, Expressions Creative Director, Allo &lt;/span&gt;&lt;br /&gt;&lt;br /&gt;Taking, sharing, and viewing selfies has become a daily habit for many — the car selfie, the cute-outfit selfie, the travel selfie, the I-woke-up-like-this selfie. Apart from a social capacity, self-portraiture has long served as a means for self and identity exploration. For some, it’s about figuring out who they are. For others it’s about projecting how they want to be perceived. Sometimes it’s both.&lt;br /&gt;&lt;br /&gt;Photography in the form of a selfie is a very direct form of expression. It comes with a set of rules bounded by reality. Illustration, on the other hand, empowers people to define themselves - it’s warmer and less fraught than reality. &lt;br /&gt;&lt;div class="separator" style="clear: both; text-align: center;"&gt;&lt;a href="https://3.bp.blogspot.com/-v__QEAtAZM4/WROVRpJzWuI/AAAAAAAABzU/iPfpKEFbs4MG7JYEhalqQtkBfDsvUz_WACLcB/s1600/image3.gif" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"&gt;&lt;img border="0" height="220" src="https://3.bp.blogspot.com/-v__QEAtAZM4/WROVRpJzWuI/AAAAAAAABzU/iPfpKEFbs4MG7JYEhalqQtkBfDsvUz_WACLcB/s640/image3.gif" width="640" /&gt;&lt;/a&gt;&lt;/div&gt;Today, Google is introducing a feature in &lt;a href="https://play.google.com/store/apps/details?id=com.google.android.apps.fireball&amp;amp;hl=en"&gt;Allo&lt;/a&gt; that uses a combination of neural networks and the work of artists to turn your selfie into a personalized sticker pack. &lt;a href="https://support.google.com/allo/answer/7126107"&gt;Simply snap a selfie&lt;/a&gt;, and it’ll return an automatically generated illustrated version of you, on the fly, with customization options to help you personalize the stickers even further.&lt;br /&gt;&lt;div class="separator" style="clear: both; text-align: center;"&gt;&lt;a href="https://3.bp.blogspot.com/-AD4KlSCL-2E/WROV2QtjHSI/AAAAAAAABzc/G_f1w3s_RdgOXDx02RcdD1LvSBoOwBrPgCLcB/s1600/image2.gif" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"&gt;&lt;img border="0" height="640" src="https://3.bp.blogspot.com/-AD4KlSCL-2E/WROV2QtjHSI/AAAAAAAABzc/G_f1w3s_RdgOXDx02RcdD1LvSBoOwBrPgCLcB/s640/image2.gif" width="348" /&gt;&lt;/a&gt;&lt;/div&gt;&lt;b&gt;What makes you, &lt;i&gt;you&lt;/i&gt;?&lt;/b&gt;&lt;br /&gt;The traditional computer vision approach to mapping selfies to art would be to analyze the pixels of an image and algorithmically determine attribute values by looking at pixel values to measure color, shape, or texture. However, people today take selfies in all types of lighting conditions and poses. And while people can easily pick out and recognize qualitative features, like eye color, regardless of the lighting condition, this is a very complex task for computers. When people look at eye color, they don’t just interpret the pixel values of blue or green, but take into account the &lt;a href="https://www.buzzfeed.com/catesish/help-am-i-going-insane-its-definitely-blue?utm_term=.wbeyaDe9Jb#.tv77n1N0z8"&gt;surrounding visual context&lt;/a&gt;.&lt;br /&gt;&lt;br /&gt;In order to account for this, we explored how we could enable an algorithm to pick out qualitative features in a manner similar to the way people do, rather than the traditional approach of hand coding how to interpret every permutation of lighting condition, eye color, etc. While we could have trained a large convolutional neural network from scratch to attempt to accomplish this, we wondered if there was a more efficient way to get results, since we expected that learning to interpret a face into an illustration would be a very iterative process. &lt;br /&gt;&lt;br /&gt;That led us to run some experiments, similar to &lt;a href="https://en.wikipedia.org/wiki/DeepDream"&gt;DeepDream&lt;/a&gt;, on some of Google's existing more general-purpose computer vision neural networks. We discovered that a few neurons among the millions in these networks were good at focusing on things they weren’t explicitly trained to look at that seemed useful for creating personalized stickers. Additionally, by virtue of being large general-purpose neural networks they had already figured out how to abstract away things they didn’t need. All that was left to do was to provide a much smaller number of human labeled examples to teach the classifiers to isolate out the qualities that the neural network already knew about the image.&lt;br /&gt;&lt;br /&gt;To create an illustration of you that captures the qualities that would make it recognizable to your friends, we worked alongside an artistic team to create illustrations that represented a wide variety of features. Artists initially designed a set of hairstyles, for example, that they thought would be representative, and with the help of human raters we used these hairstyles to train the network to match the right illustration to the right selfie. We then asked human raters to judge the sticker output against the input image to see how well it did. In some instances, they determined that some styles were not well represented, so the artists created more that the neural network could learn to identify as well.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://3.bp.blogspot.com/-lISskYaU9Oo/WROWA5GrCyI/AAAAAAAABzg/Zm3RpvTcs_4xKNry1pGh3ConxADdmT1VQCLcB/s1600/image5.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="284" src="https://3.bp.blogspot.com/-lISskYaU9Oo/WROWA5GrCyI/AAAAAAAABzg/Zm3RpvTcs_4xKNry1pGh3ConxADdmT1VQCLcB/s640/image5.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Raters were asked to classify hairstyles that the icon on the left resembled closest. Then, once consensus was reached, resident artist &lt;a href="https://twitter.com/neo_rama"&gt;Lamar Abrams&lt;/a&gt; drew a representation of what they had in common.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;&lt;b&gt;Avoiding the uncanny valley&lt;/b&gt;&lt;br /&gt;In the study of aesthetics, a well-known problem is the &lt;a href="https://en.wikipedia.org/wiki/Uncanny_valley"&gt;uncanny valley&lt;/a&gt; - the hypothesis that human replicas which appear almost, but not exactly, like real human beings can feel repulsive. In machine learning, this could be compounded if were confronted by a computer’s perception of you, versus how you may think of yourself, which can be at odds.&lt;br /&gt;&lt;br /&gt;Rather than aim to replicate a person’s appearance exactly, pursuing a lower resolution model, like emojis and stickers, allows the team to explore expressive representation by returning an image that is less about reproducing reality and more about breaking the rules of representation. &lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://4.bp.blogspot.com/-Db9OdgOr4u8/WROXwQoBrdI/AAAAAAAABzs/mhNVO20sdYAekUGqTZrc73uxHJiImxWvgCLcB/s1600/image1.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="460" src="https://4.bp.blogspot.com/-Db9OdgOr4u8/WROXwQoBrdI/AAAAAAAABzs/mhNVO20sdYAekUGqTZrc73uxHJiImxWvgCLcB/s640/image1.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;The team worked with artist &lt;a href="https://twitter.com/neo_rama"&gt;Lamar Abrams&lt;/a&gt; to design the features that make up more than 563 quadrillion combinations.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;&lt;b&gt;Translating pixels to artistic illustrations&lt;/b&gt;&lt;br /&gt;Reconciling how the computer perceives you with how you perceive yourself and what you want to project is truly an artistic exercise. This makes a customization feature that includes different hairstyles, skin tones, and nose shapes, essential. After all, illustration by its very nature can be subjective. Aesthetics are defined by race, culture, and class which can lead to creating zones of exclusion without consciously trying. As such, we strove to create a space for a range of race, age, masculinity, femininity, and/or androgyny. Our teams continue to evaluate the research results to help prevent against incorporating biases while training the system.&lt;br /&gt;&lt;div class="separator" style="clear: both; text-align: center;"&gt;&lt;a href="https://4.bp.blogspot.com/-ChsVmG28OKk/WROYJV-ImhI/AAAAAAAABzw/PwkMgCNUxe0oK9CHz7Sfr1DGWkJRhBvNgCLcB/s1600/image4.gif" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"&gt;&lt;img border="0" height="640" src="https://4.bp.blogspot.com/-ChsVmG28OKk/WROYJV-ImhI/AAAAAAAABzw/PwkMgCNUxe0oK9CHz7Sfr1DGWkJRhBvNgCLcB/s640/image4.gif" width="350" /&gt;&lt;/a&gt;&lt;/div&gt;&lt;b&gt;Creating a broad palette for identity and sentiment&lt;/b&gt;&lt;br /&gt;There is no such thing as a ‘universal aesthetic’ or ‘a singular you’. The way people talk to their parents is different than how they talk to their friends which is different than how they talk to their colleagues. It’s not enough to make an avatar that is a literal representation of yourself when there are many versions of you. To address that, the Allo team is working with a range of artistic voices to help others extend their own voice. This first style that launched today speaks to your sarcastic side but the next pack might be more cute for those sincere moments. Then after that, maybe they’ll turn you into a dog. If emojis broadened the world of communication it’s not hard to imagine how this technology and language evolves. What will be most exciting is listening to what people say with it.&lt;br /&gt;&lt;br /&gt;This feature is starting to roll out in &lt;a href="https://play.google.com/store/apps/details?id=com.google.android.apps.fireball&amp;amp;hl=en"&gt;Allo today for Android&lt;/a&gt;, and will come soon to Allo on iOS. &lt;br /&gt;&lt;br /&gt;&lt;b&gt;Acknowledgements&lt;/b&gt;&lt;br /&gt;This work was made possible through a collaboration of the Allo Team and &lt;a href="https://research.google.com/pubs/MachinePerception.html"&gt;Machine Perception&lt;/a&gt; researchers at Google. We additionally thank Lamar Abrams, Koji Ashida, Forrester Cole, Jennifer Daniel, Shiraz Fuman, Dilip Krishnan, Inbar Mosseri, Aaron Sarna, Aaron Maschinot and Bhavik Singh.&lt;div class="feedflare"&gt;
&lt;a href="http://feeds.feedburner.com/~ff/blogspot/gJZg?a=XqIFBtFOr40:CUlGdmRUssk:yIl2AUoC8zA"&gt;&lt;img src="http://feeds.feedburner.com/~ff/blogspot/gJZg?d=yIl2AUoC8zA" border="0"&gt;&lt;/img&gt;&lt;/a&gt;
&lt;/div&gt;&lt;img src="http://feeds.feedburner.com/~r/blogspot/gJZg/~4/XqIFBtFOr40" height="1" width="1" alt=""/&gt;</content><link rel="replies" type="application/atom+xml" href="http://research.googleblog.com/feeds/827399624362227391/comments/default" title="Post Comments" /><link rel="replies" type="text/html" href="http://research.googleblog.com/2017/05/neural-network-generated-illustrations.html#comment-form" title="0 Comments" /><link rel="edit" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/827399624362227391" /><link rel="self" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/827399624362227391" /><link rel="alternate" type="text/html" href="http://feedproxy.google.com/~r/blogspot/gJZg/~3/XqIFBtFOr40/neural-network-generated-illustrations.html" title="Neural Network-Generated Illustrations in Allo" /><author><name>Research Blog</name><uri>https://plus.google.com/101673966767287570260</uri><email>noreply@blogger.com</email><gd:image rel="http://schemas.google.com/g/2005#thumbnail" width="32" height="32" src="//lh5.googleusercontent.com/-GXTJ2sdBxaM/AAAAAAAAAAI/AAAAAAAAA14/2q3ThT_9JIY/s512-c/photo.jpg" /></author><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://3.bp.blogspot.com/-v__QEAtAZM4/WROVRpJzWuI/AAAAAAAABzU/iPfpKEFbs4MG7JYEhalqQtkBfDsvUz_WACLcB/s72-c/image3.gif" height="72" width="72" /><thr:total>0</thr:total><gd:extendedProperty name="commentSource" value="1" /><gd:extendedProperty name="commentModerationMode" value="FILTERED_POSTMOD" /><feedburner:origLink>http://research.googleblog.com/2017/05/neural-network-generated-illustrations.html</feedburner:origLink></entry><entry><id>tag:blogger.com,1999:blog-21224994.post-6551648794507159411</id><published>2017-05-03T11:00:00.000-07:00</published><updated>2017-05-03T17:50:12.813-07:00</updated><category scheme="http://www.blogger.com/atom/ns#" term="Computer Vision" /><category scheme="http://www.blogger.com/atom/ns#" term="Deep Learning" /><category scheme="http://www.blogger.com/atom/ns#" term="Google Brain" /><category scheme="http://www.blogger.com/atom/ns#" term="Optical Character Recognition" /><category scheme="http://www.blogger.com/atom/ns#" term="TensorFlow" /><title type="text">Updating Google Maps with Deep Learning and Street View</title><content type="html">&lt;span class="byline-author"&gt;Posted by Julian Ibarz, Staff Software Engineer, Google Brain Team and Sujoy Banerjee, Product Manager, Ground Truth Team&lt;/span&gt;&lt;br /&gt;&lt;br /&gt;Every day, Google Maps provides useful directions, real-time traffic information and information on businesses to millions of people. In order to provide the best experience for our users, this information has to constantly mirror an ever-changing world. While Street View cars collect millions of images daily, it is impossible to manually analyze more than 80 billion high resolution images collected to date in order to find new, or updated, information for Google Maps. One of the goals of the Google’s Ground Truth team is to enable the automatic extraction of information from our geo-located imagery to improve Google Maps.&lt;br /&gt;&lt;br /&gt;In “&lt;a href="https://arxiv.org/abs/1704.03549"&gt;Attention-based Extraction of Structured Information from Street View Imagery&lt;/a&gt;”, we describe our approach to accurately read street names out of very challenging Street View images in many countries, automatically, using a deep neural network. Our algorithm achieves 84.2% accuracy on the challenging &lt;a href="https://github.com/tensorflow/models/blob/master/street/README.md"&gt;French Street Name Signs&lt;/a&gt; (FSNS) dataset, significantly outperforming the previous state-of-the-art systems. Importantly, our system is easily extensible to extract other types of information out of Street View images as well, and now helps us automatically extract business names from store fronts. We are excited to announce that this model is now &lt;a href="https://github.com/tensorflow/models/tree/master/attention_ocr"&gt;publicly available&lt;/a&gt;!&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://3.bp.blogspot.com/-MVn24olbp3o/WQoVkm8OEYI/AAAAAAAABy0/hdhF4Ad79KIKA6V0Hv0tZ2w3h8_Yi4v6gCLcB/s1600/fig1.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="188" src="https://3.bp.blogspot.com/-MVn24olbp3o/WQoVkm8OEYI/AAAAAAAABy0/hdhF4Ad79KIKA6V0Hv0tZ2w3h8_Yi4v6gCLcB/s640/fig1.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Example of street name from the FSNS dataset correctly transcribed by our system. Up to four views of the same sign are provided.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;&lt;div class="separator" style="clear: both; text-align: center;"&gt;&lt;/div&gt;Text recognition in a natural environment is a challenging computer vision and machine learning problem. While traditional &lt;a href="https://en.wikipedia.org/wiki/Optical_character_recognition"&gt;Optical Character Recognition &lt;/a&gt;(OCR) systems mainly focus on extracting text from scanned documents, text acquired from natural scenes is more challenging due to visual artifacts, such as distortion, occlusions, directional blur, cluttered background or different viewpoints. Our efforts to solve this research challenge first began in 2008, when we used &lt;a href="https://research.google.com/pubs/pub35481.html"&gt;neural networks to blur faces and license plates&lt;/a&gt; in Street View images to protect the privacy of our users. From this initial research, we realized that with enough labeled data, we could additionally use machine learning not only to protect the privacy of our users, but also to automatically improve Google Maps with relevant up-to-date information.&lt;br /&gt;&lt;br /&gt;In 2014, Google’s Ground Truth team published a state-of-the-art method for &lt;a href="https://security.googleblog.com/2014/04/street-view-and-recaptcha-technology.html"&gt;reading street numbers&lt;/a&gt; on the &lt;a href="http://ufldl.stanford.edu/housenumbers/"&gt;Street View House Numbers&lt;/a&gt; (SVHN) dataset, implemented by then summer intern (now Googler) &lt;a href="https://research.google.com/pubs/105214.html"&gt;Ian Goodfellow&lt;/a&gt;. This work was not only of academic interest but was critical in making Google Maps more accurate. Today, over one-third of addresses globally have had their location improved thanks to this system. In some countries, such as Brazil, this algorithm has improved more than 90% of the addresses in Google Maps today, greatly improving the usability of our maps.&lt;br /&gt;&lt;br /&gt;The next logical step was to extend these techniques to street names. To solve this problem, we created and &lt;a href="https://github.com/tensorflow/models/blob/master/street/README.md"&gt;released French Street Name Signs&lt;/a&gt; (FSNS), a large training dataset of more than 1 million street names. The FSNS dataset was a multi-year effort designed to allow anyone to improve their OCR models on a challenging and real use case. FSNS dataset is much larger and more challenging than SVHN in that accurate recognition of street signs may require combining information from many different images.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://1.bp.blogspot.com/-ZYdBFgbis3M/WQoVxc8aPRI/AAAAAAAABy4/QeCwCnH0l_ULx3Prb50FdgxmfVzyS49LwCLcB/s1600/Fig2.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="338" src="https://1.bp.blogspot.com/-ZYdBFgbis3M/WQoVxc8aPRI/AAAAAAAABy4/QeCwCnH0l_ULx3Prb50FdgxmfVzyS49LwCLcB/s640/Fig2.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;These are examples of challenging signs that are properly transcribed by our system by selecting or combining understanding across images. The second example is extremely challenging by itself, but the model learned a language model prior that enables it to remove ambiguity and correctly read the street name. Note that in the FSNS dataset, random noise is used in the case where less than four independent views are available of the same physical sign.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;With this training set, Google intern Zbigniew Wojna spent the summer of 2016 developing a deep learning model architecture to automatically label new Street View imagery. One of the interesting strengths of our new model is that it can normalize the text to be consistent with our naming conventions, as well as ignore extraneous text, directly from the data itself.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://2.bp.blogspot.com/-L4XeR28uk_E/WQoV8wigUnI/AAAAAAAABy8/UWMSthqTsTM3NWvOgBGn83K4L53v_-fyACLcB/s1600/fig3.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="186" src="https://2.bp.blogspot.com/-L4XeR28uk_E/WQoV8wigUnI/AAAAAAAABy8/UWMSthqTsTM3NWvOgBGn83K4L53v_-fyACLcB/s640/fig3.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Example of text normalization learned from data in Brazil. Here it changes “AV.” into “Avenida” and “Pres.” into “Presidente” which is what we desire.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://4.bp.blogspot.com/-KQ3UJyKHaSE/WQoWO_yZo8I/AAAAAAAABzA/MRgZPw5xZN0j5YfWFlvW1aP49eJ9bzQiACLcB/s1600/fig4.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="148" src="https://4.bp.blogspot.com/-KQ3UJyKHaSE/WQoWO_yZo8I/AAAAAAAABzA/MRgZPw5xZN0j5YfWFlvW1aP49eJ9bzQiACLcB/s640/fig4.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;In this example, the model is not confused from the fact that there is two street names, properly normalizes “Av” into “Avenue” as well as correctly ignores the number “1600”.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;While this model is accurate, it did show a sequence error rate of 15.8%. However, after analyzing failure cases, we found that 48% of them were due to ground truth errors, highlighting the fact that this model is on par with the label quality (a full analysis our error rate can be found in &lt;a href="https://arxiv.org/abs/1704.03549"&gt;our paper&lt;/a&gt;).&lt;br /&gt;&lt;br /&gt;This new system, combined with the one extracting street numbers, allows us to create new addresses directly from imagery, where we previously didn’t know the name of the street, or the location of the addresses. Now, whenever a Street View car drives on a newly built road, our system can analyze the tens of thousands of images that would be captured, extract the street names and numbers, and properly create and locate the new addresses, automatically, on Google Maps.&lt;br /&gt;&lt;br /&gt;But automatically creating addresses for Google Maps is not enough -- additionally we want to be able to provide navigation to businesses by name. In 2015, we published “&lt;a href="https://arxiv.org/abs/1512.05430"&gt;Large Scale Business Discovery from Street View Imagery&lt;/a&gt;”, which proposed an approach to accurately detect business store-front signs in Street View images. However, once a store front is detected, one still needs to accurately extract its name for it to be useful -- the model must figure out which text is the business name, and which text is not relevant. We call this extracting “structured text” information out of imagery. It is not just text, it is text with semantic meaning attached to it. &lt;br /&gt;&lt;br /&gt;Using different training data, the same model architecture that we used to read street names can also be used to accurately extract business names out of business facades. In this particular case, we are able to only extract the business name which enables us to verify if we already know about this business in Google Maps, allowing us to have more accurate and up-to-date business listings.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://3.bp.blogspot.com/-qfh-IuJQlxU/WQoWr_KWPtI/AAAAAAAABzE/wHjU2imEKHUsmhUTFbxy2r1pHmz6pmb_QCLcB/s1600/image4.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="636" src="https://3.bp.blogspot.com/-qfh-IuJQlxU/WQoWr_KWPtI/AAAAAAAABzE/wHjU2imEKHUsmhUTFbxy2r1pHmz6pmb_QCLcB/s640/image4.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;The system is correctly able to predict the business name ‘Zelina Pneus’, despite not receiving any data about the true location of the name in the image. Model is not confused by the tire brands that the sign indicates are available at the store.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;Applying these large models across our more than 80 billion Street View images requires a lot of computing power. This is why the Ground Truth team was the first user of Google's TPUs, which were &lt;a href="https://cloudplatform.googleblog.com/2017/04/quantifying-the-performance-of-the-TPU-our-first-machine-learning-chip.html"&gt;publicly announced earlier this year&lt;/a&gt;, to drastically reduce the computational cost of the inferences of our pipeline.&lt;br /&gt;&lt;br /&gt;People rely on the accuracy of Google Maps in order to assist them. While keeping Google Maps up-to-date with the ever-changing landscape of cities, roads and businesses presents a technical challenge that is far from solved, it is the goal of the Ground Truth team to drive cutting-edge innovation in machine learning to create a better experience for over one billion Google Maps users.&lt;br /&gt;&lt;br /&gt;&lt;div class="feedflare"&gt;
&lt;a href="http://feeds.feedburner.com/~ff/blogspot/gJZg?a=UanQlzl060Q:9KRLezog8VA:yIl2AUoC8zA"&gt;&lt;img src="http://feeds.feedburner.com/~ff/blogspot/gJZg?d=yIl2AUoC8zA" border="0"&gt;&lt;/img&gt;&lt;/a&gt;
&lt;/div&gt;&lt;img src="http://feeds.feedburner.com/~r/blogspot/gJZg/~4/UanQlzl060Q" height="1" width="1" alt=""/&gt;</content><link rel="replies" type="application/atom+xml" href="http://research.googleblog.com/feeds/6551648794507159411/comments/default" title="Post Comments" /><link rel="replies" type="text/html" href="http://research.googleblog.com/2017/05/updating-google-maps-with-deep-learning.html#comment-form" title="0 Comments" /><link rel="edit" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/6551648794507159411" /><link rel="self" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/6551648794507159411" /><link rel="alternate" type="text/html" href="http://feedproxy.google.com/~r/blogspot/gJZg/~3/UanQlzl060Q/updating-google-maps-with-deep-learning.html" title="Updating Google Maps with Deep Learning and Street View" /><author><name>Research Blog</name><uri>https://plus.google.com/101673966767287570260</uri><email>noreply@blogger.com</email><gd:image rel="http://schemas.google.com/g/2005#thumbnail" width="32" height="32" src="//lh5.googleusercontent.com/-GXTJ2sdBxaM/AAAAAAAAAAI/AAAAAAAAA14/2q3ThT_9JIY/s512-c/photo.jpg" /></author><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://3.bp.blogspot.com/-MVn24olbp3o/WQoVkm8OEYI/AAAAAAAABy0/hdhF4Ad79KIKA6V0Hv0tZ2w3h8_Yi4v6gCLcB/s72-c/fig1.png" height="72" width="72" /><thr:total>0</thr:total><gd:extendedProperty name="commentSource" value="1" /><gd:extendedProperty name="commentModerationMode" value="FILTERED_POSTMOD" /><feedburner:origLink>http://research.googleblog.com/2017/05/updating-google-maps-with-deep-learning.html</feedburner:origLink></entry><entry><id>tag:blogger.com,1999:blog-21224994.post-98421893308618512</id><published>2017-04-25T09:00:00.000-07:00</published><updated>2017-04-25T09:00:00.153-07:00</updated><category scheme="http://www.blogger.com/atom/ns#" term="Computational Photography" /><category scheme="http://www.blogger.com/atom/ns#" term="Image Processing" /><category scheme="http://www.blogger.com/atom/ns#" term="Low-Light Photography" /><category scheme="http://www.blogger.com/atom/ns#" term="Nexus" /><category scheme="http://www.blogger.com/atom/ns#" term="Pixel" /><title type="text">Experimental Nighttime Photography with Nexus and Pixel</title><content type="html">&lt;span class="byline-author"&gt;Posted by Florian Kainz, Software Engineer, Google Daydream&lt;/span&gt;&lt;br /&gt;&lt;br /&gt;On a full moon night last year I carried a professional DSLR camera, a heavy lens and a tripod up to a hilltop in the &lt;a href="https://en.wikipedia.org/wiki/Marin_Headlands"&gt;Marin Headlands&lt;/a&gt; just north of San Francisco to take a picture of the Golden Gate Bridge and the lights of the city behind it. &lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://1.bp.blogspot.com/-JJORxMmYFuw/WPkESrozTuI/AAAAAAAAByU/pkjoul07knIfKWyvM-xmlaCzvODG28FJQCLcB/s1600/image5.jpg" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" src="https://1.bp.blogspot.com/-JJORxMmYFuw/WPkESrozTuI/AAAAAAAAByU/pkjoul07knIfKWyvM-xmlaCzvODG28FJQCLcB/s1600/image5.jpg" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;A view of the Golden Gate Bridge from the Marin Headlands, taken with a DSLR camera (Canon 1DX, Zeiss Otus 28mm f/1.4 ZE). Click &lt;a href="https://drive.google.com/open?id=0B0i7HeK_sTi8dE53b1ZYVm0tM1E"&gt;here&lt;/a&gt; for the full resolution image.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;I thought the photo of the moonlit landscape came out well so I showed it to my (then) teammates in &lt;a href="https://blog.x.company/meet-gcam-the-x-graduate-that-gave-us-a-whole-new-point-of-view-3ee86657d6c9"&gt;Gcam&lt;/a&gt;, a Google Research team that focuses on &lt;a href="https://en.wikipedia.org/wiki/Computational_photography"&gt;computational photography&lt;/a&gt; - developing algorithms that assist in taking pictures, usually with smartphones and similar small cameras. Seeing my nighttime photo, one of the Gcam team members challenged me to re-take it, but with a phone camera instead of a DSLR. Even though cameras on cellphones have come a long way, I wasn’t sure whether it would be possible to come close to the DSLR shot.&lt;br /&gt;&lt;br /&gt;Probably the most successful Gcam project to date is &lt;a href="https://research.googleblog.com/2014/10/hdr-low-light-and-high-dynamic-range.html"&gt;the image processing pipeline that enables the HDR+ mode&lt;/a&gt; in the camera app on Nexus and Pixel phones. HDR+ allows you to take photos at low-light levels by rapidly shooting a burst of up to ten short exposures and averaging them them into a single image, reducing blur due to camera shake while collecting enough total light to yield &lt;a href="https://plus.google.com/photos/105100397789364595413/albums/6071336333050111233"&gt;surprisingly good pictures&lt;/a&gt;. Of course there are limits to what HDR+ can do. Once it gets dark enough the camera just cannot gather enough light and challenging shots like nighttime landscapes are still beyond reach. &lt;br /&gt;&lt;br /&gt;&lt;b&gt;The Challenges&lt;/b&gt;&lt;br /&gt;To learn what was possible with a cellphone camera in extremely low-light conditions, I looked to the experimental &lt;a href="https://www.youtube.com/watch?v=S7lbnMd56Ys"&gt;SeeInTheDark&lt;/a&gt; app, written by Marc Levoy and presented at the &lt;a href="http://extremeimaging.csail.mit.edu/"&gt;ICCV 2015 Extreme Imaging Workshop&lt;/a&gt;, which can produce pictures with even less light than HDR+. It does this by accumulating more exposures, and merging them under the assumption that the scene is static and any differences between successive exposures must be due to camera motion or sensor noise. The app reduces noise further by dropping image resolution to about 1 MPixel. With SeeInTheDark it is just possible to take pictures, albeit fairly grainy ones, by the light of the full moon.&lt;br /&gt;&lt;br /&gt;However, in order to keep motion blur due to camera shake and moving objects in the scene at acceptable levels, both HDR+ and SeeInTheDark must keep the exposure times for individual frames below roughly one tenth of a second. Since the user can’t hold the camera perfectly still for extended periods, it doesn’t make sense to attempt to merge a large number of frames into a single picture. Therefore, HDR+ merges at most ten frames, while SeeInTheDark progressively discounts older frames as new ones are captured. This limits how much light the camera can gather and thus affects the quality of the final pictures at very low light levels.&lt;br /&gt;&lt;br /&gt;Of course, if we want to take high-quality pictures of low-light scenes (such as a landscape illuminated only by the moon), increasing the exposure time to more than one second and mounting the phone on a tripod or placing it on some other solid support makes the task a lot easier. Google’s Nexus 6P and Pixel phones support exposure times of 4 and 2 seconds respectively. As long as the scene is static, we should be able to record and merge dozens of frames to produce a single final image, even if shooting those frames takes several minutes.&lt;br /&gt;&lt;br /&gt;Even with the use of a tripod, a sharp picture requires the camera’s lens to be focused on the subject, and this can be tricky in scenes with very low light levels. The two autofocus mechanisms employed by cellphone cameras — &lt;a href="https://en.wikipedia.org/wiki/Autofocus#Contrast_detection"&gt;contrast detection&lt;/a&gt; and &lt;a href="https://en.wikipedia.org/wiki/Autofocus#Phase_detection"&gt;phase detection&lt;/a&gt; — fail when it’s dark enough that the camera's image sensor returns mostly noise. Fortunately, the interesting parts of outdoor scenes tend to be far enough away that simply setting the focus distance to infinity produces sharp images.&lt;br /&gt;&lt;br /&gt;&lt;b&gt;Experiments &amp;amp; Results&lt;/b&gt;&lt;br /&gt;Taking all this into account, I wrote a simple Android camera app with manual control over exposure time, ISO and focus distance. When the shutter button is pressed the app waits a few seconds and then records up to 64 frames with the selected settings. The app saves the raw frames captured from the sensor as &lt;a href="https://en.wikipedia.org/wiki/Digital_Negative"&gt;DNG&lt;/a&gt; files, which can later be downloaded onto a PC for processing.&lt;br /&gt;&lt;br /&gt;To test my app, I visited the &lt;a href="https://www.nps.gov/pore/learn/historyculture/people_maritime_lighthouse.htm"&gt;Point Reyes lighthouse&lt;/a&gt; on the California coast some thirty miles northwest of San Francisco on a full moon night. I pointed a Nexus 6P phone at the building and shot a burst of 32 four-second frames at ISO 1600. After covering the camera lens with opaque adhesive tape I shot an additional 32 black frames. Back at the office I loaded the raw files into Photoshop. The individual frames were very grainy, as one would expect given the tiny sensor in a cellphone camera, but computing the mean of all 32 frames cleaned up most of the grain, and subtracting the mean of the 32 black frames removed faint grid-like patterns caused by local variations in the sensor's black level. The resulting image, shown below, looks surprisingly good.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://1.bp.blogspot.com/-cPw2kFKkOoU/WPkERgagOjI/AAAAAAAAByk/jiKvOPtbbP839LEBffR2ISpUDpud7Lp_wCEw/s1600/image10.jpg" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" src="https://1.bp.blogspot.com/-cPw2kFKkOoU/WPkERgagOjI/AAAAAAAAByk/jiKvOPtbbP839LEBffR2ISpUDpud7Lp_wCEw/s1600/image10.jpg" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Point Reyes lighthouse at night, photographed with Google Nexus 6P (full resolution image &lt;a href="https://drive.google.com/open?id=0B0i7HeK_sTi8LUh0eHdJVmhpcUk"&gt;here&lt;/a&gt;).&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;The lantern in the lighthouse is overexposed, but the rest of the scene is sharp, not too grainy, and has pleasing, natural looking colors. For comparison, a hand-held HDR+ shot of the same scene looks like this:&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://2.bp.blogspot.com/-VnXHLDk9Q2o/WPkESPnc-aI/AAAAAAAAByk/u5gXv-GKU5YqI2hR8rRBYysbuePiObf_wCEw/s1600/image3.jpg" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" src="https://2.bp.blogspot.com/-VnXHLDk9Q2o/WPkESPnc-aI/AAAAAAAAByk/u5gXv-GKU5YqI2hR8rRBYysbuePiObf_wCEw/s1600/image3.jpg" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Point Reyes Lighthouse at night, hand-held HDR+ shot (full resolution image &lt;a href="https://drive.google.com/open?id=0B0i7HeK_sTi8UWotTlY2OFdVNzQ"&gt;here&lt;/a&gt;). The inset rectangle has been brightened in Photoshop to roughly match the previous picture.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;Satisfied with these results, I wanted to see if I could capture a nighttime landscape as well as the stars in the clear sky above it, all in one picture. When I took the photo of the lighthouse a thin layer of clouds conspired with the bright moonlight to make the stars nearly invisible, but on a clear night a two or four second exposure can easily capture the brighter stars. The stars are not stationary, though; &lt;a href="https://en.wikipedia.org/wiki/Pole_star"&gt;they appear to rotate around the celestial poles&lt;/a&gt;, completing a full turn every 24 hours. The motion is slow enough to be invisible in exposures of only a few seconds, but over the minutes it takes to record a few dozen frames the stars move enough to turn into streaks when the frames are merged. Here is an example:&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://3.bp.blogspot.com/-gYWccAfZOpQ/WPkETK0qp-I/AAAAAAAAByk/0qMjxR-Mvzwwz9ReqHzcPFNA6uIbJFJ2wCEw/s1600/image9.jpg" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" src="https://3.bp.blogspot.com/-gYWccAfZOpQ/WPkETK0qp-I/AAAAAAAAByk/0qMjxR-Mvzwwz9ReqHzcPFNA6uIbJFJ2wCEw/s1600/image9.jpg" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;The North Star above Mount Burdell, single 2-second exposure. (full resolution image &lt;a href="https://drive.google.com/open?id=0B0i7HeK_sTi8em5tajZjSFB0b1E"&gt;here&lt;/a&gt;).&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://4.bp.blogspot.com/-vRDfsctVbcA/WPkEShKZMLI/AAAAAAAAByk/wB7q748vrCY6yKfNSukUGrbL0lK4aES2gCEw/s1600/image6.jpg" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" src="https://4.bp.blogspot.com/-vRDfsctVbcA/WPkEShKZMLI/AAAAAAAAByk/wB7q748vrCY6yKfNSukUGrbL0lK4aES2gCEw/s1600/image6.jpg" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Mean of 32 2-second exposures (full resolution image &lt;a href="https://drive.google.com/open?id=0B0i7HeK_sTi8aXZwUC10b09QZ00"&gt;here&lt;/a&gt;).&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;Seeing streaks instead of pinpoint stars in the sky can be avoided by shifting and rotating the original frames such that the stars align. Merging the aligned frames produces an image with a clean-looking sky, and many faint stars that were hidden by noise in the individual frames become visible. Of course, the ground is now motion-blurred as if the camera had followed the rotation of the sky.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://2.bp.blogspot.com/-WPcQB1VtBA0/WPkESzcZEFI/AAAAAAAAByk/OCUFMe-MkWsaXKCU2hOQU2Lcnn7ZAUFlwCEw/s1600/image7.jpg" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" src="https://2.bp.blogspot.com/-WPcQB1VtBA0/WPkESzcZEFI/AAAAAAAAByk/OCUFMe-MkWsaXKCU2hOQU2Lcnn7ZAUFlwCEw/s1600/image7.jpg" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Mean of 32 2-second exposures, stars aligned (full resolution image &lt;a href="https://drive.google.com/open?id=0B0i7HeK_sTi8V0RGZUZYOE9zdzg"&gt;here&lt;/a&gt;).&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;We now have two images; one where the ground is sharp, and one where the sky is sharp, and we can combine them into a single picture that is sharp everywhere. In Photoshop the easiest way to do that is with a hand-painted layer mask. After adjusting brightness and colors to taste, slight cropping, and removing an ugly "No Trespassing" sign we get a presentable picture:&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://3.bp.blogspot.com/-w5LA9zAqZQk/WPkESFAvKgI/AAAAAAAAByk/2UsxK3l5k1UJQ1wf0Hkt4Zi0jdtp6of7gCEw/s1600/image2.jpg" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" src="https://3.bp.blogspot.com/-w5LA9zAqZQk/WPkESFAvKgI/AAAAAAAAByk/2UsxK3l5k1UJQ1wf0Hkt4Zi0jdtp6of7gCEw/s1600/image2.jpg" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;The North Star above Mount Burdell, shot with Google Pixel, final image (full resolution image &lt;a href="https://drive.google.com/open?id=0B0i7HeK_sTi8eWtMZ00tblhvdGM"&gt;here&lt;/a&gt;).&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;&lt;b&gt;Using Even Less Light&lt;/b&gt;&lt;br /&gt;The pictures I've shown so far were shot on nights with a full moon, when it was bright enough that one could easily walk outside without a lantern or a flashlight. I wanted to find out if it was possible to take cellphone photos in even less light. Using a Pixel phone, I tried a scene illuminated by a three-quarter moon low in the sky, and another one with no moon at all. Anticipating more noise in the individual exposures, I shot 64-frame bursts. The processed final images still look fine:&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://1.bp.blogspot.com/-R0SPBSFHgsM/WPkERhEsrZI/AAAAAAAAByk/cWnGPkdSKB4sp51SeZ9YPySU6U2AMtxPwCEw/s1600/image1.jpg" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" src="https://1.bp.blogspot.com/-R0SPBSFHgsM/WPkERhEsrZI/AAAAAAAAByk/cWnGPkdSKB4sp51SeZ9YPySU6U2AMtxPwCEw/s1600/image1.jpg" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Wrecked fishing boat in Inverness and the Big Dipper, 64 2-second exposures, shot with Google Pixel (full resolution image &lt;a href="https://drive.google.com/open?id=0B0i7HeK_sTi8TlZhTzdwazVhNG8"&gt;here&lt;/a&gt;).&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://3.bp.blogspot.com/-g2oeHIFx908/WPkESY6QzsI/AAAAAAAAByk/6OgqSRwoG2IhYJLzrTOV5azLo0aZ1FmtQCEw/s1600/image4.jpg" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" src="https://3.bp.blogspot.com/-g2oeHIFx908/WPkESY6QzsI/AAAAAAAAByk/6OgqSRwoG2IhYJLzrTOV5azLo0aZ1FmtQCEw/s1600/image4.jpg" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Stars above Pierce Point Ranch, 64 2-second exposures, shot with Google Pixel (full resolution image &lt;a href="https://drive.google.com/open?id=0B0i7HeK_sTi8aVl2WEs0T3doNms"&gt;here&lt;/a&gt;).&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;In the second image the distant lights of the cities around the San Francisco Bay caused the sky near the horizon to glow, but without moonlight the night was still dark enough to make the Milky Way visible. The picture looks noticeably grainier than my earlier moonlight shots, but it's not too bad.&lt;br /&gt;&lt;br /&gt;&lt;b&gt;Pushing the Limits&lt;/b&gt;&lt;br /&gt;How far can we go? Can we take a cellphone photo with only starlight - no moon, no artificial light sources nearby, and no background glow from a distant city?&lt;br /&gt;&lt;br /&gt;To test this I drove to a point on the California coast a little north of the mouth of the &lt;a href="https://en.wikipedia.org/wiki/Russian_River_(California)"&gt;Russian River&lt;/a&gt;, where nights can get really dark, and pointed my Pixel phone at the summer sky above the ocean. Combining 64 two-second exposures taken at ISO 12800, and 64 corresponding black black frames did produce a recognizable image of the Milky Way. The constellations &lt;a href="https://en.wikipedia.org/wiki/Scorpius"&gt;Scorpius&lt;/a&gt; and &lt;a href="https://en.wikipedia.org/wiki/Sagittarius_(constellation)"&gt;Sagittarius&lt;/a&gt; are clearly visible, and squinting hard enough one can just barely make out the horizon and one or two rocks in the ocean, but overall, this is not a picture you'd want to print out and frame. Still, this may be the lowest-light cellphone photo ever taken.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://2.bp.blogspot.com/-2SAtZ4kFLJY/WPkESBFiKTI/AAAAAAAAByk/fswFdIgEhqs8r5qtooaIt2fD8c2cXlApgCEw/s1600/image12.jpg" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" src="https://2.bp.blogspot.com/-2SAtZ4kFLJY/WPkESBFiKTI/AAAAAAAAByk/fswFdIgEhqs8r5qtooaIt2fD8c2cXlApgCEw/s1600/image12.jpg" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Only starlight, shot with Google Pixel (full resolution image &lt;a href="https://drive.google.com/open?id=0B0i7HeK_sTi8UkdjQ0hjcGdoT1k"&gt;here&lt;/a&gt;).&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;Here we are approaching the limits of what the Pixel camera can do. The camera cannot handle exposure times longer than two seconds. If this restriction was removed we could expose individual frames for eight to ten seconds, and the stars still would not show noticeable motion blur. With longer exposures we could lower the ISO setting, which would significantly reduce noise in the individual frames, and we would get a correspondingly cleaner and more detailed final picture.&lt;br /&gt;&lt;br /&gt;Getting back to the original challenge - using a cellphone to reproduce a night-time DSLR shot of the Golden Gate - I did that. Here is what I got:&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://4.bp.blogspot.com/-yo9hRd3xLLg/WPkERdbOs7I/AAAAAAAAByk/UDE_MAh3bXszzir_Grxs672mpiKY82DlACEw/s1600/image11.jpg" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" src="https://4.bp.blogspot.com/-yo9hRd3xLLg/WPkERdbOs7I/AAAAAAAAByk/UDE_MAh3bXszzir_Grxs672mpiKY82DlACEw/s1600/image11.jpg" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Golden Gate Bridge at night, shot with Google Nexus 6P (full resolution image &lt;a href="https://drive.google.com/open?id=0B0i7HeK_sTi8OVdjTFZvWmVuTVU"&gt;here&lt;/a&gt;).&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://4.bp.blogspot.com/-d7AUDMCZGvY/WPkES2Ava6I/AAAAAAAAByk/_NHAY42k5KAHoueJscA8DZKjdFd2mAAmgCEw/s1600/image8.jpg" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" src="https://4.bp.blogspot.com/-d7AUDMCZGvY/WPkES2Ava6I/AAAAAAAAByk/_NHAY42k5KAHoueJscA8DZKjdFd2mAAmgCEw/s1600/image8.jpg" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;The Moon above San Francisco, shot with Google Nexus 6P (full resolution image &lt;a href="https://drive.google.com/open?id=0B0i7HeK_sTi8bnZJdHBTQ0tmb1U"&gt;here&lt;/a&gt;).&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;At 9 to 10 MPixels the resolution of these pictures is not as high as what a DSLR camera might produce, but otherwise image quality is surprisingly good: the photos are sharp all the way into the corners, there is not much visible noise, the captured dynamic range is sufficient to avoid saturating all but the brightest highlights, and the colors are pleasing.&lt;br /&gt;&lt;br /&gt;Trying to find out if phone cameras might be suitable for outdoor nighttime photography was a fun experiment, and clearly the result is yes, they are. However, arriving at the final images required a lot of careful post-processing on a desktop computer, and the procedure is too cumbersome for all but the most dedicated cellphone photographers. However, with the right software a phone should be able to process the images internally, and if steps such as painting layer masks by hand can be eliminated, it might be possible to do point-and-shoot photography in very low light conditions. Almost - the cellphone would still have to rest on the ground or be mounted on a tripod.&lt;br /&gt;&lt;br /&gt;&lt;a href="https://goo.gl/photos/cMm7mb7ASr84HkTy5"&gt;Here’s a Google Photos album&lt;/a&gt; with more examples of photos that were created with the technique described above. &lt;div class="feedflare"&gt;
&lt;a href="http://feeds.feedburner.com/~ff/blogspot/gJZg?a=-VTaimOXtas:XbTFZyzDD5M:yIl2AUoC8zA"&gt;&lt;img src="http://feeds.feedburner.com/~ff/blogspot/gJZg?d=yIl2AUoC8zA" border="0"&gt;&lt;/img&gt;&lt;/a&gt;
&lt;/div&gt;&lt;img src="http://feeds.feedburner.com/~r/blogspot/gJZg/~4/-VTaimOXtas" height="1" width="1" alt=""/&gt;</content><link rel="replies" type="application/atom+xml" href="http://research.googleblog.com/feeds/98421893308618512/comments/default" title="Post Comments" /><link rel="replies" type="text/html" href="http://research.googleblog.com/2017/04/experimental-nighttime-photography-with.html#comment-form" title="0 Comments" /><link rel="edit" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/98421893308618512" /><link rel="self" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/98421893308618512" /><link rel="alternate" type="text/html" href="http://feedproxy.google.com/~r/blogspot/gJZg/~3/-VTaimOXtas/experimental-nighttime-photography-with.html" title="Experimental Nighttime Photography with Nexus and Pixel" /><author><name>Research Blog</name><uri>https://plus.google.com/101673966767287570260</uri><email>noreply@blogger.com</email><gd:image rel="http://schemas.google.com/g/2005#thumbnail" width="32" height="32" src="//lh5.googleusercontent.com/-GXTJ2sdBxaM/AAAAAAAAAAI/AAAAAAAAA14/2q3ThT_9JIY/s512-c/photo.jpg" /></author><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://1.bp.blogspot.com/-JJORxMmYFuw/WPkESrozTuI/AAAAAAAAByU/pkjoul07knIfKWyvM-xmlaCzvODG28FJQCLcB/s72-c/image5.jpg" height="72" width="72" /><thr:total>0</thr:total><gd:extendedProperty name="commentSource" value="1" /><gd:extendedProperty name="commentModerationMode" value="FILTERED_POSTMOD" /><feedburner:origLink>http://research.googleblog.com/2017/04/experimental-nighttime-photography-with.html</feedburner:origLink></entry><entry><id>tag:blogger.com,1999:blog-21224994.post-8614643642628792169</id><published>2017-04-24T00:00:00.000-07:00</published><updated>2017-04-26T08:16:09.242-07:00</updated><category scheme="http://www.blogger.com/atom/ns#" term="conference" /><category scheme="http://www.blogger.com/atom/ns#" term="conferences" /><category scheme="http://www.blogger.com/atom/ns#" term="Deep Learning" /><category scheme="http://www.blogger.com/atom/ns#" term="ICLR" /><category scheme="http://www.blogger.com/atom/ns#" term="Machine Learning" /><category scheme="http://www.blogger.com/atom/ns#" term="Publications" /><category scheme="http://www.blogger.com/atom/ns#" term="Research" /><title type="text">Research at Google and ICLR 2017</title><content type="html">&lt;span class="byline-author"&gt;Posted by Ian Goodfellow, Staff Research Scientist, Google Brain Team&lt;/span&gt; &lt;br /&gt;&lt;br /&gt;This week, Toulon, France hosts the &lt;a href="http://www.iclr.cc/doku.php?id=ICLR2017:main&amp;amp;redirect=1"&gt;5th International Conference on Learning Representations&lt;/a&gt; (ICLR 2017), a conference focused on how one can learn meaningful and useful representations of data for &lt;a href="https://en.wikipedia.org/wiki/Machine_learning"&gt;Machine Learning&lt;/a&gt;. ICLR includes conference and workshop tracks, with invited talks along with oral and poster presentations of some of the latest research on deep learning, metric learning, kernel learning, compositional models, non-linear structured prediction, and issues regarding non-convex optimization.&lt;br /&gt;&lt;br /&gt;At the forefront of innovation in cutting-edge technology in &lt;a href="https://en.wikipedia.org/wiki/Artificial_neural_network"&gt;Neural Networks&lt;/a&gt; and &lt;a href="https://en.wikipedia.org/wiki/Deep_learning"&gt;Deep Learning&lt;/a&gt;, Google focuses on both theory and application, developing learning approaches to understand and generalize. As Platinum Sponsor of ICLR 2017, Google will have a strong presence with over 50 researchers attending (many from the &lt;a href="https://research.google.com/teams/brain/"&gt;Google Brain team&lt;/a&gt; and &lt;a href="https://research.google.com/teams/gre/"&gt;Google Research Europe&lt;/a&gt;), contributing to and learning from the broader academic research community by presenting papers and posters, in addition to participating on organizing committees and in workshops.&lt;br /&gt;&lt;br /&gt;If you are attending ICLR 2017, we hope you'll stop by our booth and chat with our researchers about the projects and opportunities at Google that go into solving interesting problems for billions of people. You can also learn more about our research being presented at ICLR 2017 in the list below (Googlers highlighted in &lt;span style="color: #3d85c6;"&gt;blue&lt;/span&gt;).&lt;br /&gt;&lt;br /&gt;&lt;b&gt;&lt;u&gt;Area Chairs include:&lt;/u&gt;&lt;/b&gt;&lt;br /&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;George Dahl&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Slav Petrov&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Vikas Sindhwani&lt;/span&gt;&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;b&gt;&lt;u&gt;Program Chairs include:&lt;/u&gt;&lt;/b&gt;&lt;br /&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;Hugo Larochelle&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Tara Sainath&lt;/span&gt;&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;b&gt;&lt;u&gt;Contributed Talks&lt;/u&gt;&lt;/b&gt;&lt;br /&gt;&lt;a href="https://openreview.net/forum?id=Sy8gdB9xx&amp;amp;noteId=Sy8gdB9xx"&gt;Understanding Deep Learning Requires Rethinking Generalization&lt;/a&gt; &lt;strong&gt;&lt;i&gt;(Best Paper Award)&lt;/i&gt;&lt;/strong&gt;&lt;br /&gt;&lt;i&gt;Chiyuan Zhang*, &lt;span style="color: #3d85c6;"&gt;Samy Bengio&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Moritz Hardt&lt;/span&gt;, Benjamin Recht*, Oriol Vinyals&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://openreview.net/forum?id=HkwoSDPgg&amp;amp;noteId=HkwoSDPgg"&gt;Semi-Supervised Knowledge Transfer for Deep Learning from Private Training Data&lt;/a&gt;&amp;nbsp;&lt;strong&gt;&lt;i&gt;(Best Paper Award)&lt;/i&gt;&lt;/strong&gt;&lt;br /&gt;&lt;i&gt;Nicolas Papernot*, &lt;span style="color: #3d85c6;"&gt;Martín Abadi&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Úlfar Erlingsson&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Ian Goodfellow&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Kunal&lt;br /&gt;Talwar&lt;/span&gt;&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://openreview.net/forum?id=SJ3rcZcxl&amp;amp;noteId=SJ3rcZcxl"&gt;Q-Prop: Sample-Efficient Policy Gradient with An Off-Policy Critic&lt;/a&gt;&lt;br /&gt;&lt;i&gt;Shixiang (Shane) Gu*, Timothy Lillicrap, Zoubin Ghahramani, Richard E.&lt;br /&gt;Turner, &lt;span style="color: #3d85c6;"&gt;Sergey Levine&lt;/span&gt;&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://openreview.net/forum?id=r1Ue8Hcxg&amp;amp;noteId=r1Ue8Hcxg"&gt;Neural Architecture Search with Reinforcement Learning&lt;/a&gt;&lt;br /&gt;&lt;i&gt;&lt;span style="color: #45818e;"&gt;Barret Zoph&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #45818e;"&gt;&amp;nbsp;Quoc Le&lt;/span&gt;&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;b&gt;&lt;u&gt;Posters&lt;/u&gt;&lt;/b&gt;&lt;br /&gt;&lt;a href="https://openreview.net/forum?id=BJm4T4Kgx&amp;amp;noteId=BJm4T4Kgx"&gt;Adversarial Machine Learning at Scale&lt;/a&gt;&lt;br /&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;Alexey Kurakin&lt;/span&gt;, &lt;span style="color: #3d85c6;"&gt;Ian J. Goodfellow&lt;/span&gt;&lt;sup&gt;†&lt;/sup&gt;, &lt;span style="color: #3d85c6;"&gt;Samy Bengio&lt;/span&gt;&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://openreview.net/forum?id=BydARw9ex&amp;amp;noteId=BydARw9ex"&gt;Capacity and Trainability in Recurrent Neural Networks&lt;/a&gt;&lt;br /&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;Jasmine Collins&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Jascha Sohl-Dickstein&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;David Sussillo&lt;/span&gt;&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://openreview.net/forum?id=ryT4pvqll&amp;amp;noteId=ryT4pvqll"&gt;Improving Policy Gradient by Exploring Under-Appreciated Rewards&lt;/a&gt;&lt;br /&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;Ofir Nachum&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Mohammad Norouzi&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Dale Schuurmans&lt;/span&gt;&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://openreview.net/forum?id=B1ckMDqlg&amp;amp;noteId=B1ckMDqlg"&gt;Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer&lt;/a&gt;&lt;br /&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;Noam Shazeer&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Azalia Mirhoseini&lt;/span&gt;, Krzysztof Maziarz,&lt;span style="color: #3d85c6;"&gt; Andy Davis&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Quoc Le&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&amp;nbsp;&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;Geoffrey Hinton&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Jeff Dean&lt;/span&gt;&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://openreview.net/forum?id=BydrOIcle&amp;amp;noteId=BydrOIcle"&gt;Unrolled Generative Adversarial Networks&lt;/a&gt;&lt;br /&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;Luke Metz&lt;/span&gt;, Ben Poole*, David Pfau, &lt;span style="color: #3d85c6;"&gt;Jascha Sohl-Dickstein&lt;/span&gt;&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://openreview.net/forum?id=rkE3y85ee&amp;amp;noteId=rkE3y85ee"&gt;Categorical Reparameterization with Gumbel-Softmax&lt;/a&gt;&lt;br /&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;Eric Jang&lt;/span&gt;, Shixiang (Shane) Gu*, Ben Poole*&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://openreview.net/forum?id=rkEFLFqee&amp;amp;noteId=rkEFLFqee"&gt;Decomposing Motion and Content for Natural Video Sequence Prediction&lt;/a&gt;&lt;br /&gt;&lt;i&gt;Ruben Villegas, Jimei Yang, Seunghoon Hong, Xunyu Lin, &lt;span style="color: #3d85c6;"&gt;Honglak Lee&lt;/span&gt;&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://openreview.net/forum?id=HkpbnH9lx&amp;amp;noteId=HkpbnH9lx"&gt;Density Estimation Using Real NVP&lt;/a&gt;&lt;br /&gt;&lt;i&gt;Laurent Dinh*, &lt;span style="color: #3d85c6;"&gt;Jascha Sohl-Dickstein&lt;/span&gt;, &lt;span style="color: #3d85c6;"&gt;Samy Bengio&lt;/span&gt;&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://openreview.net/forum?id=SyQq185lg&amp;amp;noteId=SyQq185lg"&gt;Latent Sequence Decompositions&lt;/a&gt;&lt;br /&gt;&lt;i&gt;William Chan*, Yu Zhang*, &lt;span style="color: #3d85c6;"&gt;Quoc Le&lt;/span&gt;, Navdeep Jaitly*&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://openreview.net/forum?id=ry2YOrcge&amp;amp;noteId=ry2YOrcge"&gt;Learning a Natural Language Interface with Neural Programmer&lt;/a&gt;&lt;br /&gt;&lt;i&gt;Arvind Neelakantan*, &lt;span style="color: #3d85c6;"&gt;Quoc V. Le&lt;/span&gt;, &lt;span style="color: #3d85c6;"&gt;Martín Abadi&lt;/span&gt;, Andrew McCallum*, Dario&lt;/i&gt;&lt;br /&gt;&lt;i&gt; Amodei*&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://openreview.net/forum?id=H1W1UN9gg&amp;amp;noteId=H1W1UN9gg"&gt;Deep Information Propagation&lt;/a&gt;&lt;br /&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;Samuel Schoenholz&lt;/span&gt;, &lt;span style="color: #3d85c6;"&gt;Justin Gilmer&lt;/span&gt;, Surya Ganguli, &lt;span style="color: #3d85c6;"&gt;Jascha Sohl-Dickstein&lt;/span&gt;&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://openreview.net/forum?id=ryxB0Rtxx&amp;amp;noteId=ryxB0Rtxx"&gt;Identity Matters in Deep Learning&lt;/a&gt;&lt;br /&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;Moritz Hardt&lt;/span&gt;, Tengyu Ma&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://openreview.net/forum?id=BJO-BuT1g&amp;amp;noteId=BJO-BuT1g"&gt;A Learned Representation For Artistic Style&lt;/a&gt;&lt;br /&gt;&lt;i&gt;Vincent Dumoulin*, &lt;span style="color: #3d85c6;"&gt;Jonathon Shlens&lt;/span&gt;, &lt;span style="color: #3d85c6;"&gt;Manjunath Kudlur&lt;/span&gt;&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://openreview.net/forum?id=r1X3g2_xl&amp;amp;noteId=r1X3g2_xl"&gt;Adversarial Training Methods for Semi-Supervised Text Classification&lt;/a&gt;&lt;br /&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;Takeru Miyato&lt;/span&gt;, &lt;span style="color: #3d85c6;"&gt;Andrew M. Dai&lt;/span&gt;, &lt;span style="color: #3d85c6;"&gt;Ian Goodfellow&lt;/span&gt;&lt;sup&gt;†&lt;/sup&gt;&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://openreview.net/forum?id=rkpACe1lx&amp;amp;noteId=rkpACe1lx"&gt;HyperNetworks&lt;/a&gt;&lt;br /&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;David Ha&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Andrew Dai&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Quoc V. Le&lt;/span&gt;&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://openreview.net/forum?id=SJTQLdqlg&amp;amp;noteId=SJTQLdqlg"&gt;Learning to Remember Rare Events&lt;/a&gt;&lt;br /&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;Lukasz Kaiser&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Ofir Nachum, &lt;/span&gt;Aurko Roy*&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Samy Bengio&lt;/span&gt;&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://openreview.net/forum?id=ryrGawqex&amp;amp;noteId=ryrGawqex"&gt;Deep Learning with Dynamic Computation Graphs&lt;/a&gt;&lt;br /&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;Moshe Looks&lt;/span&gt;, &lt;span style="color: #3d85c6;"&gt;Marcello Herreshof&lt;/span&gt;, &lt;span style="color: #3d85c6;"&gt;DeLesley Hutchins&lt;/span&gt;, &lt;span style="color: #3d85c6;"&gt;Peter Norvig&lt;/span&gt;&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://openreview.net/forum?id=ryuxYmvel&amp;amp;noteId=ryuxYmvel"&gt;HolStep: A Machine Learning Dataset for Higher-order Logic Theorem Proving&lt;/a&gt;&lt;br /&gt;&lt;i&gt;Cezary Kaliszyk, &lt;span style="color: #3d85c6;"&gt;François Chollet&lt;/span&gt;, &lt;span style="color: #3d85c6;"&gt;Christian Szegedy&lt;/span&gt;&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;b&gt;&lt;u&gt;Workshop Track Abstracts&lt;/u&gt;&lt;/b&gt;&lt;br /&gt;&lt;a href="https://openreview.net/forum?id=BJyBKyHKg&amp;amp;noteId=BJyBKyHKg"&gt;Particle Value Functions&lt;/a&gt;&lt;br /&gt;&lt;i&gt;Chris J. Maddison, &lt;span style="color: #3d85c6;"&gt;Dieterich Lawson&lt;/span&gt;, &lt;span style="color: #3d85c6;"&gt;George Tucker&lt;/span&gt;, Nicolas Heess, Arnaud Doucet, Andriy Mnih, Yee Whye Teh&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://openreview.net/forum?id=rJY3vK9eg&amp;amp;noteId=rJY3vK9eg"&gt;Neural Combinatorial Optimization with Reinforcement Learning&lt;/a&gt;&lt;br /&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;Irwan Bello&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Hieu Pham&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Quoc V. Le&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Mohammad Norouzi&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Samy Bengio&lt;/span&gt;&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://openreview.net/forum?id=r1br_2Kge&amp;amp;noteId=r1br_2Kge"&gt;Short and Deep: Sketching and Neural Networks&lt;/a&gt;&lt;br /&gt;&lt;span style="color: #3d85c6;"&gt;Amit Daniely&lt;/span&gt;&lt;i&gt;,&lt;/i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Nevena Lazic&lt;/span&gt;&lt;i&gt;,&lt;/i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Yoram Singer&lt;/span&gt;&lt;i&gt;,&lt;/i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Kunal Talwar&lt;/span&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://openreview.net/pdf?id=HkXKUTVFl"&gt;Explaining the Learning Dynamics of Direct Feedback Alignment&lt;/a&gt;&lt;br /&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;Justin Gilmer&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Colin Raffel&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Samuel S. Schoenholz&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Maithra Raghu&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Jascha Sohl-Dickstein&lt;/span&gt;&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://arxiv.org/abs/1702.06914"&gt;Training a Subsampling Mechanism in Expectation&lt;/a&gt;&lt;br /&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;Colin Raffel&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Dieterich Lawson&lt;/span&gt;&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://openreview.net/forum?id=BJ8fyHceg&amp;amp;noteId=BJ8fyHceg"&gt;Tuning Recurrent Neural Networks with Reinforcement Learning&lt;/a&gt;&lt;br /&gt;&lt;i&gt;Natasha Jaques&lt;/i&gt;*&lt;i&gt;, Shixiang (Shane) Gu*, Richard E. Turner, &lt;span style="color: #3d85c6;"&gt;Douglas Eck&lt;/span&gt;&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://openreview.net/pdf?id=ryBDyehOl"&gt;REBAR: Low-Variance, Unbiased Gradient Estimates for Discrete Latent Variable Models&lt;/a&gt;&lt;br /&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;George Tucker&lt;/span&gt;, Andriy Mnih, Chris J. Maddison, &lt;span style="color: #3d85c6;"&gt;Jascha Sohl-Dickstein&lt;/span&gt;&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://openreview.net/forum?id=S1OufnIlx&amp;amp;noteId=S1OufnIlx"&gt;Adversarial Examples in the Physical World&lt;/a&gt;&lt;br /&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;Alexey Kurakin&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Ian Goodfellow&lt;/span&gt;&lt;sup&gt;†&lt;/sup&gt;, &lt;span style="color: #3d85c6;"&gt;Samy Bengio&lt;/span&gt;&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://openreview.net/forum?id=HkCjNI5ex&amp;amp;noteId=HkCjNI5ex"&gt;Regularizing Neural Networks by Penalizing Confident Output Distributions&lt;/a&gt;&lt;br /&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;Gabriel Pereyra&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;George Tucker&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Jan Chorowski&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Lukasz Kaiser&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Geoffrey Hinton&lt;/span&gt;&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://openreview.net/forum?id=Bkul3t9ee&amp;amp;noteId=Bkul3t9ee"&gt;Unsupervised Perceptual Rewards for Imitation Learning&lt;/a&gt;&lt;br /&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;Pierre Sermanet&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Kelvin Xu&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Sergey Levine&lt;/span&gt;&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://arxiv.org/abs/1702.07780"&gt;Changing Model Behavior at Test-time Using Reinforcement Learning&lt;/a&gt;&lt;br /&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;Augustus Odena&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Dieterich Lawson&lt;/span&gt;&lt;/i&gt;&lt;i&gt;,&lt;/i&gt;&lt;i&gt;&lt;span style="color: #3d85c6;"&gt;&amp;nbsp;Christopher Olah&lt;/span&gt;&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;* Work performed while at Google&lt;br /&gt;† Work performed while at OpenAI&lt;/i&gt;&lt;div class="feedflare"&gt;
&lt;a href="http://feeds.feedburner.com/~ff/blogspot/gJZg?a=-S4Nyf5tIyY:uiw4EgQqloQ:yIl2AUoC8zA"&gt;&lt;img src="http://feeds.feedburner.com/~ff/blogspot/gJZg?d=yIl2AUoC8zA" border="0"&gt;&lt;/img&gt;&lt;/a&gt;
&lt;/div&gt;&lt;img src="http://feeds.feedburner.com/~r/blogspot/gJZg/~4/-S4Nyf5tIyY" height="1" width="1" alt=""/&gt;</content><link rel="replies" type="application/atom+xml" href="http://research.googleblog.com/feeds/8614643642628792169/comments/default" title="Post Comments" /><link rel="replies" type="text/html" href="http://research.googleblog.com/2017/04/research-at-google-and-iclr-2017.html#comment-form" title="0 Comments" /><link rel="edit" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/8614643642628792169" /><link rel="self" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/8614643642628792169" /><link rel="alternate" type="text/html" href="http://feedproxy.google.com/~r/blogspot/gJZg/~3/-S4Nyf5tIyY/research-at-google-and-iclr-2017.html" title="Research at Google and ICLR 2017" /><author><name>Research Blog</name><uri>https://plus.google.com/101673966767287570260</uri><email>noreply@blogger.com</email><gd:image rel="http://schemas.google.com/g/2005#thumbnail" width="32" height="32" src="//lh5.googleusercontent.com/-GXTJ2sdBxaM/AAAAAAAAAAI/AAAAAAAAA14/2q3ThT_9JIY/s512-c/photo.jpg" /></author><thr:total>0</thr:total><gd:extendedProperty name="commentSource" value="1" /><gd:extendedProperty name="commentModerationMode" value="FILTERED_POSTMOD" /><feedburner:origLink>http://research.googleblog.com/2017/04/research-at-google-and-iclr-2017.html</feedburner:origLink></entry><entry><id>tag:blogger.com,1999:blog-21224994.post-4543747413734193468</id><published>2017-04-20T10:00:00.000-07:00</published><updated>2017-04-20T10:42:50.233-07:00</updated><category scheme="http://www.blogger.com/atom/ns#" term="Computational Photography" /><category scheme="http://www.blogger.com/atom/ns#" term="Computer Vision" /><category scheme="http://www.blogger.com/atom/ns#" term="Google Photos" /><category scheme="http://www.blogger.com/atom/ns#" term="PhotoScan" /><title type="text">PhotoScan: Taking Glare-Free Pictures of Pictures</title><content type="html">&lt;span class="byline-author"&gt;Posted by Ce Liu, Michael Rubinstein, Mike Krainin and Bill Freeman, Research Scientists&lt;/span&gt;&lt;br /&gt;&lt;br /&gt;Yesterday, we released an &lt;a href="https://blog.google/products/photos/scan-your-printed-photos-just-one-tap/"&gt;update to PhotoScan&lt;/a&gt;, an app for &lt;a href="https://itunes.apple.com/app/apple-store/id1165525994?pt=9008&amp;amp;ct=ogb&amp;amp;mt=8"&gt;iOS&lt;/a&gt; and &lt;a href="https://play.google.com/store/apps/details?id=com.google.android.apps.photos.scanner&amp;amp;referrer=utm_source%3Dogb"&gt;Android&lt;/a&gt; that allows you to digitize photo prints with just a smartphone.  One of the key features of PhotoScan is the ability to remove glare from prints, which are often glossy and reflective, as are the plastic album pages or glass-covered picture frames that host them. To create this feature, we developed a unique blend of computer vision and image processing techniques that can carefully align and combine several slightly different pictures of a print to separate the glare from the image underneath.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://1.bp.blogspot.com/-NRMc4h_25y0/WPfVg7YLn7I/AAAAAAAABvU/IQE9IVz6yeAQbrg50Qryn0-i97MePc-1wCLcB/s1600/Fig.1.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="438" src="https://1.bp.blogspot.com/-NRMc4h_25y0/WPfVg7YLn7I/AAAAAAAABvU/IQE9IVz6yeAQbrg50Qryn0-i97MePc-1wCLcB/s640/Fig.1.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;&lt;b&gt;Left:&lt;/b&gt; A regular digital picture of a physical print.  &lt;b&gt;Right:&lt;/b&gt; Glare-free digital output from PhotoScan&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;When taking a single picture of a photo, determining which regions of the picture are the actual photo and which regions are glare is challenging to do automatically. Moreover, the glare may often saturate regions in the picture, rendering it impossible to see or recover the parts of the photo underneath it. But if we take several pictures of the photo while moving the camera, the position of the glare tends to change, covering different regions of the photo. In most cases we found that every pixel of the photo is likely not to be covered by glare in at least one of the pictures. While no single view may be glare-free, we can combine multiple pictures of the printed photo taken at different angles to remove the glare. The challenge is that the images need to be aligned very accurately in order to combine them properly, and this processing needs to run very quickly on the phone to provide a near instant experience. &lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://2.bp.blogspot.com/-6GF9Zk-KZwM/WPfWJkAqRTI/AAAAAAAABvY/7sx1tcHJvOU0QVjBA2ju3UNKKwogEjX_QCLcB/s1600/image24.gif" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="396" src="https://2.bp.blogspot.com/-6GF9Zk-KZwM/WPfWJkAqRTI/AAAAAAAABvY/7sx1tcHJvOU0QVjBA2ju3UNKKwogEjX_QCLcB/s640/image24.gif" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;&lt;b&gt;Left:&lt;/b&gt; The captured, input images (5 in total). &lt;b&gt;Right:&lt;/b&gt; If we stabilize the images on the photo, we can see just the glare moving, covering different parts of the photo. Notice no single image is glare-free.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;Our technique is inspired by our earlier work published at &lt;a href="http://s2015.siggraph.org/"&gt;SIGGRAPH 2015&lt;/a&gt;, which we dubbed “&lt;a href="https://sites.google.com/site/obstructionfreephotography/"&gt;obstruction-free photography&lt;/a&gt;”. It uses similar principles to remove various types of obstructions from the field of view. However, the algorithm we originally proposed was based on a generative model where the motion and appearance of both the main scene and the obstruction layer are estimated. While that model is quite powerful and can remove a variety of obstructions, it is too computationally expensive to be run on smartphones.  We therefore developed a simpler model that treats glare as an outlier, and only attempts to register the underlying, glare-free photo. While this model is simpler, the task is still quite challenging as the registration needs to be highly accurate and robust.&lt;br /&gt;&lt;br /&gt;&lt;b&gt;How it Works&lt;/b&gt;&lt;br /&gt;We start from a series of pictures of the print taken by the user while moving the camera. The first picture - the “reference frame” - defines the desired output viewpoint. The user is then instructed to take four additional frames. In each additional frame, we detect &lt;a href="https://en.wikipedia.org/wiki/Feature_detection_(computer_vision)"&gt;sparse feature points&lt;/a&gt;  (we compute &lt;a href="http://www.willowgarage.com/sites/default/files/orb_final.pdf"&gt;ORB features&lt;/a&gt; on &lt;a href="https://en.wikipedia.org/wiki/Corner_detection"&gt;Harris corners&lt;/a&gt;) and use them to establish &lt;a href="https://en.wikipedia.org/wiki/Homography_(computer_vision)"&gt;homographies&lt;/a&gt; mapping each frame to the reference frame.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://3.bp.blogspot.com/-Y5HVN6R4RXI/WPfYQ6LUq6I/AAAAAAAABv0/5Ms1GdyJLMw-1oCULAo6NT-oLclsqrVCgCLcB/s1600/image1.gif" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="256" src="https://3.bp.blogspot.com/-Y5HVN6R4RXI/WPfYQ6LUq6I/AAAAAAAABv0/5Ms1GdyJLMw-1oCULAo6NT-oLclsqrVCgCLcB/s640/image1.gif" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Left: Detected feature matches between the reference frame and each other frame (left), and the warped frames according to the estimated homographies (right).&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;While the technique may sound straightforward, there is a catch - homographies are only able to align flat images. But printed photos are often not entirely flat (as is the case with the example shown above). Therefore, we use &lt;a href="https://en.wikipedia.org/wiki/Optical_flow"&gt;optical flow&lt;/a&gt; — a fundamental, computer vision representation for motion, which establishes pixel-wise mapping between two images — to correct the non-planarities. We start from the homography-aligned frames, and compute “flow fields” to warp the images and further refine the registration. In the example below, notice how the corners of the photo on the left slightly “move” after registering the frames using only homographies. The right hand side shows how the photo is better aligned after refining the registration using optical flow.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://1.bp.blogspot.com/-diqjle2sxtE/WPfZAD_D7bI/AAAAAAAABv8/OjXhzF7iwE094ko_kZOp1qL_6LuzniOjQCLcB/s1600/image13.gif" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="396" src="https://1.bp.blogspot.com/-diqjle2sxtE/WPfZAD_D7bI/AAAAAAAABv8/OjXhzF7iwE094ko_kZOp1qL_6LuzniOjQCLcB/s640/image13.gif" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Comparison between the warped frames using homographies (left) and after the additional warp refinement using optical flow (right).&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;The difference in the registration is subtle, but has a big impact on the end result. Notice how small misalignments manifest themselves as duplicated image structures in the result, and how these artifacts are alleviated with the additional flow refinement.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://4.bp.blogspot.com/-Bio4tf8MoPM/WPfZ_IZUZ4I/AAAAAAAABwE/w1QCvmCCooM_OMYyW6HARJYNQfkR6FAKACLcB/s1600/image36.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="228" src="https://4.bp.blogspot.com/-Bio4tf8MoPM/WPfZ_IZUZ4I/AAAAAAAABwE/w1QCvmCCooM_OMYyW6HARJYNQfkR6FAKACLcB/s640/image36.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Comparison between the glare removal result with (right) and without (left) optical flow refinement. In the result using homographies only (left), notice artifacts around the eye, nose and teeth of the person, and duplicated stems and flower petals on the fabric.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;Here too, the challenge was to make optical flow, a naturally slow algorithm, work very quickly on the phone. Instead of computing optical flow at each pixel as done traditionally (the number of flow vectors computed is equal to the number of input pixels), we represent a flow field by a smaller number of control points, and express the motion at each pixel in the image as a function of the motion at the control points. Specifically, we divide each image into tiled, non-overlapping cells to form a coarse grid, and represent the flow of a pixel in a cell as the &lt;a href="https://en.wikipedia.org/wiki/Bilinear_interpolation"&gt;bilinear combination&lt;/a&gt; of the flow at the four corners of the cell that contains it.&lt;br /&gt;&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://1.bp.blogspot.com/-2BGs2hD11Gc/WPfaf1EFLfI/AAAAAAAABwM/4oowfe_O0zkAxTInqRasClj0rLUA3NP1wCLcB/s1600/image15.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="252" src="https://1.bp.blogspot.com/-2BGs2hD11Gc/WPfaf1EFLfI/AAAAAAAABwM/4oowfe_O0zkAxTInqRasClj0rLUA3NP1wCLcB/s640/image15.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;The grid setup for grid optical flow. A point p is represented as the &lt;a href="https://en.wikipedia.org/wiki/Bilinear_interpolation"&gt;bilinear interpolation&lt;/a&gt; of the four corner points of the cell that encapsulates it.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://1.bp.blogspot.com/-1ByxAzXYXxQ/WPfaX27TXPI/AAAAAAAABwI/di3WahqYU44ku1LTr4qgDGf1D9TqLNPywCEw/s1600/image16.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="440" src="https://1.bp.blogspot.com/-1ByxAzXYXxQ/WPfaX27TXPI/AAAAAAAABwI/di3WahqYU44ku1LTr4qgDGf1D9TqLNPywCEw/s640/image16.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;&lt;b&gt;Left:&lt;/b&gt; Illustration of the computed flow field on one of the frames. &lt;b&gt;Right:&lt;/b&gt; The flow color coding: orientation and magnitude represented by hue and saturation, respectively.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;This results in a much smaller problem to solve, since the number of flow vectors to compute now equals the number of grid points, which is typically much smaller than the number of pixels. This process is similar in nature to the spline-based image registration described in &lt;a href="https://pdfs.semanticscholar.org/8f57/f2e7cd7ee8042a5600b4d5b5638a10f06fe2.pdf"&gt;Szeliski and Coughlan (1997)&lt;/a&gt;. With this algorithm, we were able to reduce the optical flow computation time by a factor of ~40 on a Pixel phone!&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://4.bp.blogspot.com/-Q5bnNur88ME/WPfbie00jGI/AAAAAAAABwU/46YVBXgKuGoW5dtsu2tyFp6Gwt8MS78WQCLcB/s1600/image30.gif" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="492" src="https://4.bp.blogspot.com/-Q5bnNur88ME/WPfbie00jGI/AAAAAAAABwU/46YVBXgKuGoW5dtsu2tyFp6Gwt8MS78WQCLcB/s640/image30.gif" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Flipping between the homography-registered frame and the flow-refined warped frame (using the above flow field), superimposed on the (clean) reference frame, shows how the computed flow field “snaps” image parts to their corresponding parts in the reference frame, improving the registration.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;Finally, in order to compose the glare-free output, for any given location in the registered frames, we examine the pixel values, and use a soft minimum algorithm to obtain the darkest observed value. More specifically, we compute the expectation of the minimum brightness over the registered frames, assigning less weight to pixels close to the (warped) image boundaries. We use this method rather than computing the minimum directly across the frames due to the fact that corresponding pixels at each frame may have slightly different brightness. Therefore, per-pixel minimum can produce visible seams due to sudden intensity changes at boundaries between overlaid images.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://1.bp.blogspot.com/-Cqdgdu3SvFY/WPfb0q4nFFI/AAAAAAAABwY/3GYnLUGGDTwUKOTRCaeZLPKp_VTE8dyAQCLcB/s1600/image20.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="640" src="https://1.bp.blogspot.com/-Cqdgdu3SvFY/WPfb0q4nFFI/AAAAAAAABwY/3GYnLUGGDTwUKOTRCaeZLPKp_VTE8dyAQCLcB/s640/image20.png" width="628" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Regular minimum (left) versus soft minimum (right) over the registered frames.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;The algorithm can support a variety of scanning conditions — matte and gloss prints, photos inside or outside albums, magazine covers.&lt;br /&gt;&lt;br /&gt;&lt;div&gt;&lt;table border="0" cellpadding="0" cellspacing="0" style="width: 100%;"&gt;&lt;tbody&gt;&lt;tr&gt; &lt;td style="text-align: center;"&gt;&lt;span id="docs-internal-guid-01f4804a-885a-6bd2-c1c9-d80fb7fb89e0"&gt;&lt;b&gt;&lt;span style="font-family: &amp;quot;arial&amp;quot;; font-size: 11pt; font-style: italic; vertical-align: baseline; white-space: pre-wrap;"&gt;Input&lt;/span&gt;&lt;span style="font-family: &amp;quot;arial&amp;quot;; font-size: 11pt; vertical-align: baseline; white-space: pre-wrap;"&gt;&lt;span class="Apple-tab-span" style="white-space: pre;"&gt; &lt;/span&gt;&lt;/span&gt;&lt;span style="font-family: &amp;quot;arial&amp;quot;; font-size: 11pt; vertical-align: baseline; white-space: pre-wrap;"&gt;&lt;span class="Apple-tab-span" style="white-space: pre;"&gt; &lt;/span&gt;&lt;/span&gt;&lt;span style="font-family: &amp;quot;arial&amp;quot;; font-size: 11pt; vertical-align: baseline; white-space: pre-wrap;"&gt;&lt;span class="Apple-tab-span" style="white-space: pre;"&gt; &lt;/span&gt;&lt;/span&gt;&lt;span style="font-family: &amp;quot;arial&amp;quot;; font-size: 11pt; vertical-align: baseline; white-space: pre-wrap;"&gt;&lt;span class="Apple-tab-span" style="white-space: pre;"&gt; &lt;/span&gt;&lt;/span&gt;&lt;span style="font-family: &amp;quot;arial&amp;quot;; font-size: 11pt; vertical-align: baseline; white-space: pre-wrap;"&gt;&lt;span class="Apple-tab-span" style="white-space: pre;"&gt;                             &lt;/span&gt;&lt;/span&gt;&lt;span style="font-family: &amp;quot;arial&amp;quot;; font-size: 11pt; vertical-align: baseline; white-space: pre-wrap;"&gt; &amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&lt;/span&gt;&lt;span style="font-family: &amp;quot;arial&amp;quot;; font-size: 11pt; font-style: italic; vertical-align: baseline; white-space: pre-wrap;"&gt;Registered&lt;/span&gt;&lt;span style="font-family: &amp;quot;arial&amp;quot;; font-size: 11pt; vertical-align: baseline; white-space: pre-wrap;"&gt;&lt;span class="Apple-tab-span" style="white-space: pre;"&gt; &lt;/span&gt;&lt;/span&gt;&lt;span style="font-family: &amp;quot;arial&amp;quot;; font-size: 11pt; vertical-align: baseline; white-space: pre-wrap;"&gt;&lt;span class="Apple-tab-span" style="white-space: pre;"&gt; &lt;/span&gt;&lt;/span&gt;&lt;span style="font-family: &amp;quot;arial&amp;quot;; font-size: 11pt; vertical-align: baseline; white-space: pre-wrap;"&gt;&lt;span class="Apple-tab-span" style="white-space: pre;"&gt; &lt;/span&gt;&lt;/span&gt;&lt;span style="font-family: &amp;quot;arial&amp;quot;; font-size: 11pt; vertical-align: baseline; white-space: pre-wrap;"&gt;&lt;span class="Apple-tab-span" style="white-space: pre;"&gt;                 &lt;/span&gt;&lt;/span&gt;&lt;span style="font-family: &amp;quot;arial&amp;quot;; font-size: 11pt; vertical-align: baseline; white-space: pre-wrap;"&gt; &amp;nbsp;            &lt;/span&gt;&lt;span style="font-family: &amp;quot;arial&amp;quot;; font-size: 11pt; vertical-align: baseline; white-space: pre-wrap;"&gt;&lt;span class="Apple-tab-span" style="white-space: pre;"&gt; &lt;/span&gt;&lt;/span&gt;&lt;span style="font-family: &amp;quot;arial&amp;quot;; font-size: 11pt; vertical-align: baseline; white-space: pre-wrap;"&gt; &amp;nbsp;&amp;nbsp;&lt;/span&gt;&lt;span style="font-family: &amp;quot;arial&amp;quot;; font-size: 11pt; font-style: italic; vertical-align: baseline; white-space: pre-wrap;"&gt;Glare-free&lt;/span&gt;&lt;/b&gt;&lt;/span&gt;&lt;/td&gt; &lt;/tr&gt;&lt;tr&gt; &lt;td&gt;&lt;div style="text-align: center;"&gt;&lt;a href="https://2.bp.blogspot.com/-fKAEe3-A8Us/WPjx_IqGqNI/AAAAAAAABxY/FWqYMA_2kiUf_TYa2bX1oVr5tVrjjfpbgCLcB/s1600/album.gif" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"&gt;&lt;img border="0" height="156" src="https://2.bp.blogspot.com/-fKAEe3-A8Us/WPjx_IqGqNI/AAAAAAAABxY/FWqYMA_2kiUf_TYa2bX1oVr5tVrjjfpbgCLcB/s640/album.gif" width="640" /&gt;&lt;/a&gt;&lt;/div&gt;&lt;/td&gt; &lt;/tr&gt;&lt;tr&gt; &lt;td&gt;&lt;a href="https://1.bp.blogspot.com/-bMXYqQndqAc/WPjyCV-GZtI/AAAAAAAABxc/yMcwOYxbMCUieygGl6m3UaD-tVEC_GGRQCLcB/s1600/156.gif" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"&gt;&lt;img border="0" height="272" src="https://1.bp.blogspot.com/-bMXYqQndqAc/WPjyCV-GZtI/AAAAAAAABxc/yMcwOYxbMCUieygGl6m3UaD-tVEC_GGRQCLcB/s640/156.gif" width="640" /&gt;&lt;/a&gt;&lt;/td&gt; &lt;/tr&gt;&lt;tr&gt; &lt;td&gt;&lt;a href="https://1.bp.blogspot.com/-YqCLhUEc8gw/WPjyFdQMSNI/AAAAAAAABxg/pke73LmNmzgJs-XaPYwzgpXelQCoDW98gCLcB/s1600/photo2.gif" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"&gt;&lt;img border="0" height="142" src="https://1.bp.blogspot.com/-YqCLhUEc8gw/WPjyFdQMSNI/AAAAAAAABxg/pke73LmNmzgJs-XaPYwzgpXelQCoDW98gCLcB/s640/photo2.gif" width="640" /&gt;&lt;/a&gt;&lt;/td&gt; &lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;&lt;/div&gt;To get the final result, the Photos team has developed a method that automatically detects and crops the photo area, and rectifies it to a frontal view. Because of perspective distortion, the scanned rectangular photo usually appears to be a quadrangle on the image. The method analyzes image signals, like color and edges, to figure out the exact boundary of the original photo on the scanned image, then applies a geometric transformation to rectify the quadrangle area back to its original rectangular shape yielding high-quality, glare-free digital version of the photo.&lt;br /&gt;&lt;div class="separator" style="clear: both; text-align: center;"&gt;&lt;a href="https://1.bp.blogspot.com/-BBtJ_YvtR0o/WPfcipl1qiI/AAAAAAAABwg/wlYhAYSkY2cKIbV349H9_ZCJjtowdcvZgCLcB/s1600/image17.png" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"&gt;&lt;img border="0" height="640" src="https://1.bp.blogspot.com/-BBtJ_YvtR0o/WPfcipl1qiI/AAAAAAAABwg/wlYhAYSkY2cKIbV349H9_ZCJjtowdcvZgCLcB/s640/image17.png" width="426" /&gt;&lt;/a&gt;&lt;/div&gt;So overall, quite a lot going on under the hood, and all done almost instantaneously on your phone! To give PhotoScan a try, download the app on &lt;a href="https://play.google.com/store/apps/details?id=com.google.android.apps.photos.scanner&amp;amp;referrer=utm_source%3Dogb"&gt;Android&lt;/a&gt; or &lt;a href="https://itunes.apple.com/app/apple-store/id1165525994?pt=9008&amp;amp;ct=ogb&amp;amp;mt=8"&gt;iOS&lt;/a&gt;.&lt;div class="feedflare"&gt;
&lt;a href="http://feeds.feedburner.com/~ff/blogspot/gJZg?a=RBJhlm7R83w:vOvOvyCdQL8:yIl2AUoC8zA"&gt;&lt;img src="http://feeds.feedburner.com/~ff/blogspot/gJZg?d=yIl2AUoC8zA" border="0"&gt;&lt;/img&gt;&lt;/a&gt;
&lt;/div&gt;&lt;img src="http://feeds.feedburner.com/~r/blogspot/gJZg/~4/RBJhlm7R83w" height="1" width="1" alt=""/&gt;</content><link rel="replies" type="application/atom+xml" href="http://research.googleblog.com/feeds/4543747413734193468/comments/default" title="Post Comments" /><link rel="replies" type="text/html" href="http://research.googleblog.com/2017/04/photoscan-taking-glare-free-pictures-of.html#comment-form" title="0 Comments" /><link rel="edit" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/4543747413734193468" /><link rel="self" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/4543747413734193468" /><link rel="alternate" type="text/html" href="http://feedproxy.google.com/~r/blogspot/gJZg/~3/RBJhlm7R83w/photoscan-taking-glare-free-pictures-of.html" title="PhotoScan: Taking Glare-Free Pictures of Pictures" /><author><name>Research Blog</name><uri>https://plus.google.com/101673966767287570260</uri><email>noreply@blogger.com</email><gd:image rel="http://schemas.google.com/g/2005#thumbnail" width="32" height="32" src="//lh5.googleusercontent.com/-GXTJ2sdBxaM/AAAAAAAAAAI/AAAAAAAAA14/2q3ThT_9JIY/s512-c/photo.jpg" /></author><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://1.bp.blogspot.com/-NRMc4h_25y0/WPfVg7YLn7I/AAAAAAAABvU/IQE9IVz6yeAQbrg50Qryn0-i97MePc-1wCLcB/s72-c/Fig.1.png" height="72" width="72" /><thr:total>0</thr:total><gd:extendedProperty name="commentSource" value="1" /><gd:extendedProperty name="commentModerationMode" value="FILTERED_POSTMOD" /><feedburner:origLink>http://research.googleblog.com/2017/04/photoscan-taking-glare-free-pictures-of.html</feedburner:origLink></entry><entry><id>tag:blogger.com,1999:blog-21224994.post-8410495058131747007</id><published>2017-04-13T10:00:00.000-07:00</published><updated>2017-04-13T16:37:04.126-07:00</updated><category scheme="http://www.blogger.com/atom/ns#" term="Google Brain" /><category scheme="http://www.blogger.com/atom/ns#" term="Machine Learning" /><category scheme="http://www.blogger.com/atom/ns#" term="Magenta" /><title type="text">Teaching Machines to Draw</title><content type="html">&lt;span class="byline-author"&gt;Posted by David Ha, Google Brain Resident&lt;/span&gt;&lt;br /&gt;&lt;br /&gt;Abstract visual communication is a key part of how people convey ideas to one another. From a young age, children develop the ability to depict objects, and arguably even emotions, with only a few pen strokes. These simple drawings may not resemble reality as captured by a photograph, but they do tell us something about how people represent and reconstruct images of the world around them.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://1.bp.blogspot.com/-X0o-v6SQXtI/WO6R8X22wlI/AAAAAAAABts/Qbvu5QoVwZIfKB-zEV7Po_Juh9AqEb28gCLcB/s1600/twitter_cover.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" src="https://1.bp.blogspot.com/-X0o-v6SQXtI/WO6R8X22wlI/AAAAAAAABts/Qbvu5QoVwZIfKB-zEV7Po_Juh9AqEb28gCLcB/s1600/twitter_cover.png" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Vector drawings produced by sketch-rnn.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;In our recent paper, “&lt;a href="https://arxiv.org/abs/1704.03477"&gt;A Neural Representation of Sketch Drawings&lt;/a&gt;”, we present a generative &lt;a href="https://en.wikipedia.org/wiki/Recurrent_neural_network"&gt;recurrent neural network&lt;/a&gt; capable of producing sketches of common objects, with the goal of training a machine to draw and generalize abstract concepts in a manner similar to humans. We train our model on a dataset of hand-drawn sketches, each represented as a sequence of motor actions controlling a pen: which direction to move, when to lift the pen up, and when to stop drawing. In doing so, we created a model that potentially has many applications, from assisting the creative process of an artist, to helping teach students how to draw.&lt;br /&gt;&lt;br /&gt;While there is a already a large body of existing work on generative modelling of images using neural networks, most of the work focuses on modelling &lt;a href="https://en.wikipedia.org/wiki/Raster_graphics"&gt;raster images&lt;/a&gt; represented as a 2D grid of pixels. While these models are currently able to generate realistic images, due to the high dimensionality of a 2D grid of pixels, a key challenge for them is to generate images with coherent structure. For example, these models sometimes produce amusing images of cats with three or more eyes, or dogs with multiple heads.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://2.bp.blogspot.com/-6s_kqOJ291M/WO6TjERWCdI/AAAAAAAABtw/AJqbZKwFDN0Grrji1DVypTy2G0b8O2XmgCLcB/s1600/image02.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="384" src="https://2.bp.blogspot.com/-6s_kqOJ291M/WO6TjERWCdI/AAAAAAAABtw/AJqbZKwFDN0Grrji1DVypTy2G0b8O2XmgCLcB/s640/image02.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Examples of animals generated with the wrong number of body parts, produced using previous GAN models trained on 128x128 ImageNet dataset. The image above is Figure 29 of&lt;br /&gt;&lt;a href="https://arxiv.org/abs/1701.00160"&gt;Generative Adversarial Networks&lt;/a&gt;, Ian Goodfellow, NIPS 2016 Tutorial.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;In this work, we investigate a lower-dimensional vector-based representation inspired by how people draw. Our model, sketch-rnn, is based on the &lt;a href="https://research.google.com/pubs/pub43155.html"&gt;sequence-to-sequence&lt;/a&gt; (seq2seq) autoencoder framework. It incorporates &lt;a href="https://research.googleblog.com/2014/12/advances-in-variational-inference.html"&gt;variational inference&lt;/a&gt; and utilizes &lt;a href="https://research.google.com/pubs/pub45823.html"&gt;hypernetworks&lt;/a&gt; as recurrent neural network cells. The goal of a seq2seq autoencoder is to train a network to encode an input sequence into a vector of floating point numbers, called a &lt;i&gt;latent&lt;/i&gt; vector, and from this latent vector reconstruct an output sequence using a decoder that replicates the input sequence as closely as possible.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://1.bp.blogspot.com/-iuML0km0cv0/WO6U0ukCaYI/AAAAAAAABt0/wsBg174LgtQslJgG2Q6jKeOTFP1EH3o3ACLcB/s1600/sketch_rnn.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" src="https://1.bp.blogspot.com/-iuML0km0cv0/WO6U0ukCaYI/AAAAAAAABt0/wsBg174LgtQslJgG2Q6jKeOTFP1EH3o3ACLcB/s1600/sketch_rnn.png" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Schematic of sketch-rnn.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;In our model, we deliberately add noise to the latent vector. In our paper, we show that by inducing noise into the communication channel between the encoder and the decoder, the model is no longer be able to reproduce the input sketch exactly, but instead must learn to capture the essence of the sketch as a noisy latent vector. Our decoder takes this latent vector and produces a sequence of motor actions used to construct a new sketch. In the figure below, we feed several actual sketches of cats into the encoder to produce reconstructed sketches using the decoder.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://4.bp.blogspot.com/-oAPlGiBet6s/WO6VOPRPzkI/AAAAAAAABt4/vXkP4Yy8-0guIDU6AhKfP0VdnAJBx5ssgCLcB/s1600/image11.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="224" src="https://4.bp.blogspot.com/-oAPlGiBet6s/WO6VOPRPzkI/AAAAAAAABt4/vXkP4Yy8-0guIDU6AhKfP0VdnAJBx5ssgCLcB/s640/image11.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Reconstructions from a model trained on cat sketches.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;&lt;br /&gt;It is important to emphasize that the reconstructed cat sketches are &lt;i&gt;not&lt;/i&gt; copies of the input sketches, but are instead &lt;i&gt;new&lt;/i&gt; sketches of cats with similar characteristics as the inputs. To demonstrate that the model is not simply copying from the input sequence, and that it actually learned something about the way people draw cats, we can try to feed in non-standard sketches into the encoder:&lt;br /&gt;&lt;div class="separator" style="clear: both; text-align: center;"&gt;&lt;a href="https://1.bp.blogspot.com/-0T75JKlZl7s/WO6Vp1ZIrZI/AAAAAAAABt8/0GUe3SVcEIMntEtSJ2oNTRb4tIvYTGkHQCLcB/s1600/image13.png" imageanchor="1"&gt;&lt;img border="0" height="361" src="https://1.bp.blogspot.com/-0T75JKlZl7s/WO6Vp1ZIrZI/AAAAAAAABt8/0GUe3SVcEIMntEtSJ2oNTRb4tIvYTGkHQCLcB/s640/image13.png" width="640" /&gt;&lt;/a&gt;&lt;/div&gt;When we feed in a sketch of a three-eyed cat, the model generates a similar looking cat that has two eyes instead, suggesting that our model has learned that cats usually only have two eyes. To show that our model is not simply choosing the closest normal-looking cat from a large collection of memorized cat-sketches, we can try to input something totally different, like a sketch of a toothbrush. We see that the network generates a cat-like figure with long whiskers that mimics the features and orientation of the toothbrush. This suggests that the network has learned to encode an input sketch into a set of abstract cat-concepts embedded into the latent vector, and is also able to reconstruct an entirely new sketch based on this latent vector.&lt;br /&gt;&lt;br /&gt;Not convinced? We repeat the experiment again on a model trained on pig sketches and arrive at similar conclusions. When presented with an eight-legged pig, the model generates a similar pig with only four legs. If we feed a truck into the pig-drawing model, we get a pig that looks a bit like the truck.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://3.bp.blogspot.com/-5LztXFr0lE8/WO6V6ym5VOI/AAAAAAAABuA/b5XkHbZtSc4gkE1r_SgDOtKcMjZcJe_BQCLcB/s1600/image10.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="202" src="https://3.bp.blogspot.com/-5LztXFr0lE8/WO6V6ym5VOI/AAAAAAAABuA/b5XkHbZtSc4gkE1r_SgDOtKcMjZcJe_BQCLcB/s640/image10.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Reconstructions from a model trained on pig sketches.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;To investigate how these latent vectors encode conceptual animal features, in the figure below, we first obtain two latent vectors encoded from two very different pigs, in this case a pig head (in the green box) and a full pig (in the orange box).  We want to get a sense of how our model learned to represent pigs, and one way to do this is to interpolate between the two different latent vectors, and visualize each generated sketch from each interpolated latent vector.  In the figure below, we visualize how the sketch of the pig head slowly morphs into the sketch of the full pig, and in the process show how the model organizes the concepts of pig sketches.  We see that the latent vector controls the relatively position and size of the nose relative to the head, and also the existence of the body and legs in the sketch.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://3.bp.blogspot.com/-tTBaBZUBnD0/WO6WlYrMwTI/AAAAAAAABuE/CTutYj42bW8lBDm3U2-SGg3HKtMrC84BQCLcB/s1600/image04.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" src="https://3.bp.blogspot.com/-tTBaBZUBnD0/WO6WlYrMwTI/AAAAAAAABuE/CTutYj42bW8lBDm3U2-SGg3HKtMrC84BQCLcB/s1600/image04.png" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Latent space interpolations generated from a model trained on pig sketches.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;We would also like to know if our model can learn representations of multiple animals, and if so, what would they look like? In the figure below, we generate sketches from interpolating latent vectors between a cat head and a full pig. We see how the representation slowly transitions from a cat head, to a cat with a tail, to a cat with a fat body, and finally into a full pig. Like a child learning to draw animals, our model learns to construct animals by attaching a head, feet, and a tail to its body. We see that the model is also able to draw cat heads that are distinct from pig heads.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://1.bp.blogspot.com/-gou3INIvWG4/WO6XBiwE20I/AAAAAAAABuI/QJWHeUGjBNwG5zRHknw2Rcq1hzY868LGgCLcB/s1600/image12.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="156" src="https://1.bp.blogspot.com/-gou3INIvWG4/WO6XBiwE20I/AAAAAAAABuI/QJWHeUGjBNwG5zRHknw2Rcq1hzY868LGgCLcB/s640/image12.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Latent Space Interpolations from a model trained on sketches of both cats and pigs.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;These interpolation examples suggest that the latent vectors indeed encode conceptual features of a sketch. But can we use these features to augment other sketches without such features - for example, adding a body to a cat's head?&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://4.bp.blogspot.com/-yK7t-68XTzA/WO6XNTIb66I/AAAAAAAABuM/PqB64rz_YgMYM3EkP2BEJojnCLHsAMXqgCLcB/s1600/image05.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="190" src="https://4.bp.blogspot.com/-yK7t-68XTzA/WO6XNTIb66I/AAAAAAAABuM/PqB64rz_YgMYM3EkP2BEJojnCLHsAMXqgCLcB/s640/image05.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Learned relationships between abstract concepts, explored using latent vector arithmetic.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;Indeed, we find that sketch drawing analogies are possible for our model trained on both cat and pig sketches. For example, we can subtract the latent vector of an encoded pig head from the latent vector of a full pig, to arrive at a vector that represents the concept of a body.  Adding this difference to the latent vector of a cat head results in a full cat (i.e. cat head + body = full cat). These drawing analogies allow us to explore how the model organizes its latent space to represent different concepts in the manifold of generated sketches.&lt;br /&gt;&lt;br /&gt;&lt;b&gt;Creative Applications&lt;/b&gt;&lt;br /&gt;In addition to the research component of this work, we are also super excited about potential creative applications of sketch-rnn. For instance, even in the simplest use case, pattern designers can apply sketch-rnn to generate a large number of similar, but unique designs for textile or wallpaper prints.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://4.bp.blogspot.com/-JKzxGG1pusY/WO6YNlzvfFI/AAAAAAAABuQ/_diVcKJWHHEecMeixOHCSnG8v4rHDt3cwCLcB/s1600/multi_cat_examples.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" src="https://4.bp.blogspot.com/-JKzxGG1pusY/WO6YNlzvfFI/AAAAAAAABuQ/_diVcKJWHHEecMeixOHCSnG8v4rHDt3cwCLcB/s1600/multi_cat_examples.png" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Similar, but unique cats, generated from a single input sketch (green and yellow boxes).&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;As we saw earlier, a model trained to draw pigs can be made to draw pig-like trucks if given an input sketch of a truck. We can extend this result to applications that might help creative designers come up with abstract designs that can resonate more with their target audience.&lt;br /&gt;&lt;br /&gt;For instance, in the figure below, we feed sketches of four different chairs into our cat-drawing model to produce four chair-like cats. We can go further and incorporate the interpolation methodology described earlier to explore the latent space of chair-like cats, and produce a large grid of generated designs to select from. &lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://2.bp.blogspot.com/-XSXUQnJGAiU/WO6ZBUTS4GI/AAAAAAAABuU/IXrSlolMzpoSZPrHX797htnP65uvRHwCQCLcB/s1600/image03.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="430" src="https://2.bp.blogspot.com/-XSXUQnJGAiU/WO6ZBUTS4GI/AAAAAAAABuU/IXrSlolMzpoSZPrHX797htnP65uvRHwCQCLcB/s640/image03.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Exploring the latent space of generated chair-cats.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;Exploring the latent space between different objects can potentially enable creative designers to find interesting intersections and relationships between different drawings.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://2.bp.blogspot.com/-F1C-Zcwb_xI/WO6ZPv3A7sI/AAAAAAAABuY/LiLxxRW-KlY2T0aFDhJq2kUgmjyJQoQAgCLcB/s1600/multiple_interpolations.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" src="https://2.bp.blogspot.com/-F1C-Zcwb_xI/WO6ZPv3A7sI/AAAAAAAABuY/LiLxxRW-KlY2T0aFDhJq2kUgmjyJQoQAgCLcB/s1600/multiple_interpolations.png" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Exploring the latent space of generated sketches of everyday objects.&lt;br /&gt;Latent space interpolation from left to right, and then top to bottom.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;We can also use the decoder module of sketch-rnn as a standalone model and train it to predict different possible endings of incomplete sketches. This technique can lead to applications where the model assists the creative process of an artist by suggesting alternative ways to finish an incomplete sketch. In the figure below, we draw different incomplete sketches (in red), and have the model come up with different possible ways to complete the drawings.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://3.bp.blogspot.com/-Tpppe68EuSw/WO6ZgdENDVI/AAAAAAAABuc/3VHMomdZ1lITXqM65s4JS1xvqoPyyaHzACLcB/s1600/full_predictions.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" src="https://3.bp.blogspot.com/-Tpppe68EuSw/WO6ZgdENDVI/AAAAAAAABuc/3VHMomdZ1lITXqM65s4JS1xvqoPyyaHzACLcB/s1600/full_predictions.png" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;The model can start with incomplete sketches (the red partial sketches to the left of the vertical line) and automatically generate different completions.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;We can take this concept even further, and have different models complete the same incomplete sketch. In the figures below, we see how to make the same circle and square figures become a part of various ants, flamingos, helicopters, owls, couches and even paint brushes. By using a diverse set of models trained to draw various objects, designers can explore creative ways to communicate meaningful visual messages to their audience.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://1.bp.blogspot.com/-5l1FYHe6O20/WO6ZwaMIf2I/AAAAAAAABug/EEcoxzlb9mgCMNJaUhHPwFTuFChrZhm3ACLcB/s1600/multi_prediction.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" src="https://1.bp.blogspot.com/-5l1FYHe6O20/WO6ZwaMIf2I/AAAAAAAABug/EEcoxzlb9mgCMNJaUhHPwFTuFChrZhm3ACLcB/s1600/multi_prediction.png" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Predicting the endings of the same circle and square figures (center) using various sketch-rnn models trained to draw different objects.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;We are very excited about the future possibilities of generative vector image modelling. These models will enable many exciting new creative applications in a variety of different directions. They can also serve as a tool to help us improve our understanding of our own creative thought processes. Learn more about sketch-rnn by reading our paper, “&lt;a href="https://arxiv.org/abs/1704.03477"&gt;A Neural Representation of Sketch Drawings&lt;/a&gt;”.&lt;br /&gt;&lt;br /&gt;&lt;b&gt;Acknowledgements&lt;/b&gt;&lt;br /&gt;We thank Ian Johnson, Jonas Jongejan, Martin Wattenberg, Mike Schuster, Ben Poole, Kyle Kastner, Junyoung Chung, Kyle McDonald for their help with this project. This work was done as part of the &lt;a href="http://g.co/brainresidency"&gt;Google Brain Residency&lt;/a&gt; program.&lt;div class="feedflare"&gt;
&lt;a href="http://feeds.feedburner.com/~ff/blogspot/gJZg?a=vwpFovXgCf0:_b-8nFRF-bo:yIl2AUoC8zA"&gt;&lt;img src="http://feeds.feedburner.com/~ff/blogspot/gJZg?d=yIl2AUoC8zA" border="0"&gt;&lt;/img&gt;&lt;/a&gt;
&lt;/div&gt;&lt;img src="http://feeds.feedburner.com/~r/blogspot/gJZg/~4/vwpFovXgCf0" height="1" width="1" alt=""/&gt;</content><link rel="replies" type="application/atom+xml" href="http://research.googleblog.com/feeds/8410495058131747007/comments/default" title="Post Comments" /><link rel="replies" type="text/html" href="http://research.googleblog.com/2017/04/teaching-machines-to-draw.html#comment-form" title="0 Comments" /><link rel="edit" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/8410495058131747007" /><link rel="self" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/8410495058131747007" /><link rel="alternate" type="text/html" href="http://feedproxy.google.com/~r/blogspot/gJZg/~3/vwpFovXgCf0/teaching-machines-to-draw.html" title="Teaching Machines to Draw" /><author><name>Research Blog</name><uri>https://plus.google.com/101673966767287570260</uri><email>noreply@blogger.com</email><gd:image rel="http://schemas.google.com/g/2005#thumbnail" width="32" height="32" src="//lh5.googleusercontent.com/-GXTJ2sdBxaM/AAAAAAAAAAI/AAAAAAAAA14/2q3ThT_9JIY/s512-c/photo.jpg" /></author><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://1.bp.blogspot.com/-X0o-v6SQXtI/WO6R8X22wlI/AAAAAAAABts/Qbvu5QoVwZIfKB-zEV7Po_Juh9AqEb28gCLcB/s72-c/twitter_cover.png" height="72" width="72" /><thr:total>0</thr:total><gd:extendedProperty name="commentSource" value="1" /><gd:extendedProperty name="commentModerationMode" value="FILTERED_POSTMOD" /><feedburner:origLink>http://research.googleblog.com/2017/04/teaching-machines-to-draw.html</feedburner:origLink></entry><entry><id>tag:blogger.com,1999:blog-21224994.post-8615799087445291746</id><published>2017-04-11T11:00:00.000-07:00</published><updated>2017-04-11T13:12:30.769-07:00</updated><category scheme="http://www.blogger.com/atom/ns#" term="Deep Learning" /><category scheme="http://www.blogger.com/atom/ns#" term="Google Brain" /><category scheme="http://www.blogger.com/atom/ns#" term="Machine Translation" /><category scheme="http://www.blogger.com/atom/ns#" term="open source" /><category scheme="http://www.blogger.com/atom/ns#" term="TensorFlow" /><title type="text">Introducing tf-seq2seq: An Open Source Sequence-to-Sequence Framework in TensorFlow</title><content type="html">&lt;span class="byline-author"&gt;Posted by Anna Goldie and Denny Britz, Research Software Engineer and Google Brain Resident, Google Brain Team&lt;/span&gt;&lt;br /&gt;&lt;br /&gt;&lt;i&gt;(Crossposted on the &lt;a href="http://opensource.googleblog.com/2017/04/tf-seq2seq-sequence-to-sequence-framework-in-tensorflow.html"&gt;Google Open Source Blog&lt;/a&gt;)&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;Last year, we announced &lt;a href="https://research.googleblog.com/2016/09/a-neural-network-for-machine.html"&gt;Google Neural Machine Translation&lt;/a&gt; (GNMT), a sequence-to-sequence (“&lt;a href="https://papers.nips.cc/paper/5346-sequence-to-sequence-learning-with-neural-networks.pdf"&gt;seq2seq&lt;/a&gt;”) model which is now used in Google Translate production systems. While GNMT achieved huge improvements in translation quality, its impact was limited by the fact that the framework for training these models was unavailable to external researchers.&lt;br /&gt;&lt;br /&gt;Today, we are excited to introduce &lt;a href="https://google.github.io/seq2seq/"&gt;tf-seq2seq&lt;/a&gt;, an open source seq2seq framework in TensorFlow that makes it easy to experiment with seq2seq models and achieve state-of-the-art results. To that end, we made the tf-seq2seq codebase clean and modular, maintaining full test coverage and documenting all of its functionality.&lt;br /&gt;&lt;br /&gt;Our framework supports various configurations of the standard seq2seq model, such as depth of the encoder/decoder, attention mechanism, RNN cell type, or beam size. This versatility allowed us to discover optimal hyperparameters and outperform other frameworks, as described in our paper, “&lt;a href="https://arxiv.org/pdf/1703.03906.pdf"&gt;Massive Exploration of Neural Machine Translation Architectures&lt;/a&gt;.”&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://4.bp.blogspot.com/-6DALk3-hPtA/WO04i5GgXLI/AAAAAAAABtc/2t9mYz4nQDg9jLoHdTkywDUfxIOFJfC_gCLcB/s1600/Seq2SeqDiagram.gif" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="382" src="https://4.bp.blogspot.com/-6DALk3-hPtA/WO04i5GgXLI/AAAAAAAABtc/2t9mYz4nQDg9jLoHdTkywDUfxIOFJfC_gCLcB/s640/Seq2SeqDiagram.gif" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;A seq2seq model translating from Mandarin to English. At each time step, the encoder takes in one Chinese character and its own previous state (black arrow), and produces an output vector (blue arrow). The decoder then generates an English translation word-by-word, at each time step taking in the last word, the previous state, and a weighted combination of all the outputs of the encoder (aka attention [3], depicted in blue) and then producing the next English word. Please note that in our implementation we use wordpieces [4] to handle rare words.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;In addition to machine translation, tf-seq2seq can also be applied to any other sequence-to-sequence task (i.e. learning to produce an output sequence given an input sequence), including machine summarization, image captioning, speech recognition, and conversational modeling. We carefully designed our framework to maintain this level of generality and provide tutorials, preprocessed data, and other utilities for &lt;a href="https://google.github.io/seq2seq/nmt/"&gt;machine translation&lt;/a&gt;.&lt;br /&gt;&lt;br /&gt;We hope that you will use tf-seq2seq to accelerate (or kick off) your own deep learning research. We also welcome your contributions to our &lt;a href="https://github.com/google/seq2seq"&gt;GitHub repository&lt;/a&gt;, where we have a variety of open issues that we would love to have your help with!&lt;br /&gt;&lt;br /&gt;&lt;b&gt;Acknowledgments:&lt;/b&gt;&lt;br /&gt;We’d like to thank Eugene Brevdo, Melody Guan, Lukasz Kaiser, Quoc V. Le, Thang Luong, and Chris Olah for all their help. For a deeper dive into how seq2seq models work, please see the resources below. &lt;br /&gt;&lt;br /&gt;&lt;b&gt;References:&lt;/b&gt;&lt;br /&gt;[1] &lt;a href="https://arxiv.org/pdf/1703.03906.pdf"&gt;Massive Exploration of Neural Machine Translation Architectures&lt;/a&gt;, &lt;i&gt;Denny Britz, Anna Goldie, Minh-Thang Luong, Quoc Le&lt;/i&gt;&lt;br /&gt;[2] &lt;a href="https://papers.nips.cc/paper/5346-sequence-to-sequence-learning-with-neural-networks.pdf"&gt;Sequence to Sequence Learning with Neural Networks&lt;/a&gt;, &lt;i&gt;Ilya Sutskever, Oriol Vinyals, Quoc V. Le. NIPS, 2014&lt;/i&gt;&lt;br /&gt;[3] &lt;a href="https://arxiv.org/abs/1409.0473"&gt;Neural Machine Translation by Jointly Learning to Align and Translate&lt;/a&gt;, &lt;i&gt;Dzmitry Bahdanau, Kyunghyun Cho, Yoshua Bengio. ICLR, 2015&lt;/i&gt;&lt;br /&gt;[4] &lt;a href="https://arxiv.org/abs/1609.08144"&gt;Google’s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation&lt;/a&gt;, &lt;i&gt;Yonghui Wu, Mike Schuster, Zhifeng Chen, Quoc V. Le, Mohammad Norouzi, Wolfgang Macherey, Maxim Krikun, Yuan Cao, Qin Gao, Klaus Macherey, Jeff Klingner, Apurva Shah, Melvin Johnson, Xiaobing Liu, Łukasz Kaiser, Stephan Gouws, Yoshikiyo Kato, Taku Kudo, Hideto Kazawa, Keith Stevens, George Kurian, Nishant Patil, Wei Wang, Cliff Young, Jason Smith, Jason Riesa, Alex Rudnick, Oriol Vinyals, Greg Corrado, Macduff Hughes, Jeffrey Dean. Technical Report, 2016&lt;/i&gt;&lt;br /&gt;[5] &lt;a href="http://distill.pub/2016/augmented-rnns/"&gt;Attention and Augmented Recurrent Neural Networks&lt;/a&gt;, &lt;i&gt;Chris Olah, Shan Carter. Distill, 2016&lt;/i&gt;&lt;br /&gt;[6] &lt;a href="https://arxiv.org/abs/1703.01619"&gt;Neural Machine Translation and Sequence-to-sequence Models: A Tutorial&lt;/a&gt;, &lt;i&gt;Graham Neubig&lt;/i&gt;&lt;br /&gt;[7] &lt;a href="https://www.tensorflow.org/tutorials/seq2seq"&gt;Sequence-to-Sequence Models&lt;/a&gt;, &lt;i&gt;TensorFlow.org&lt;/i&gt;&lt;div class="feedflare"&gt;
&lt;a href="http://feeds.feedburner.com/~ff/blogspot/gJZg?a=LZ4x4YzhLQ8:YmKPsCyKlHQ:yIl2AUoC8zA"&gt;&lt;img src="http://feeds.feedburner.com/~ff/blogspot/gJZg?d=yIl2AUoC8zA" border="0"&gt;&lt;/img&gt;&lt;/a&gt;
&lt;/div&gt;&lt;img src="http://feeds.feedburner.com/~r/blogspot/gJZg/~4/LZ4x4YzhLQ8" height="1" width="1" alt=""/&gt;</content><link rel="replies" type="application/atom+xml" href="http://research.googleblog.com/feeds/8615799087445291746/comments/default" title="Post Comments" /><link rel="replies" type="text/html" href="http://research.googleblog.com/2017/04/introducing-tf-seq2seq-open-source.html#comment-form" title="0 Comments" /><link rel="edit" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/8615799087445291746" /><link rel="self" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/8615799087445291746" /><link rel="alternate" type="text/html" href="http://feedproxy.google.com/~r/blogspot/gJZg/~3/LZ4x4YzhLQ8/introducing-tf-seq2seq-open-source.html" title="Introducing tf-seq2seq: An Open Source Sequence-to-Sequence Framework in TensorFlow" /><author><name>Research Blog</name><uri>https://plus.google.com/101673966767287570260</uri><email>noreply@blogger.com</email><gd:image rel="http://schemas.google.com/g/2005#thumbnail" width="32" height="32" src="//lh5.googleusercontent.com/-GXTJ2sdBxaM/AAAAAAAAAAI/AAAAAAAAA14/2q3ThT_9JIY/s512-c/photo.jpg" /></author><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://4.bp.blogspot.com/-6DALk3-hPtA/WO04i5GgXLI/AAAAAAAABtc/2t9mYz4nQDg9jLoHdTkywDUfxIOFJfC_gCLcB/s72-c/Seq2SeqDiagram.gif" height="72" width="72" /><thr:total>0</thr:total><gd:extendedProperty name="commentSource" value="1" /><gd:extendedProperty name="commentModerationMode" value="FILTERED_POSTMOD" /><feedburner:origLink>http://research.googleblog.com/2017/04/introducing-tf-seq2seq-open-source.html</feedburner:origLink></entry><entry><id>tag:blogger.com,1999:blog-21224994.post-6501613459508973924</id><published>2017-04-10T10:00:00.000-07:00</published><updated>2017-04-10T10:00:23.006-07:00</updated><category scheme="http://www.blogger.com/atom/ns#" term="ph.d. fellowship" /><category scheme="http://www.blogger.com/atom/ns#" term="University Relations" /><title type="text">Announcing the 2017 Google PhD Fellows for North America, Europe and the Middle East</title><content type="html">&lt;span class="byline-author"&gt;Posted by Michael Rennaker, Program Manager&lt;/span&gt;&lt;br /&gt;&lt;br /&gt;Google created the &lt;a href="http://research.google.com/research-outreach.html#/research-outreach/graduate-fellowships"&gt;PhD Fellowship program&lt;/a&gt; in 2009 to recognize and support outstanding graduate students doing exceptional research in Computer Science and related disciplines. Now in its eighth year, our fellowship program has supported hundreds of future faculty, industry researchers, innovators and entrepreneurs.&lt;br /&gt;&lt;br /&gt;Reflecting our continuing commitment to supporting and building relationships with the academic community, we are excited to announce the 33 recipients from North America, Europe and the Middle East. We offer our sincere congratulations to Google’s 2017 Class of Google PhD Fellows.&lt;br /&gt;&lt;br /&gt;&lt;b&gt;&lt;u&gt;Algorithms, Optimizations and Markets&lt;/u&gt;&lt;/b&gt;&lt;br /&gt;Chiu Wai Sam Wong, &lt;i&gt;University of California, Berkeley&lt;/i&gt;&lt;br /&gt;Eric Balkanski, &lt;i&gt;Harvard University&lt;/i&gt;&lt;br /&gt;Haifeng Xu, &lt;i&gt;University of Southern California&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;b&gt;&lt;u&gt;Human-Computer Interaction&lt;/u&gt;&lt;/b&gt;&lt;br /&gt;Motahhare Eslami, &lt;i&gt;University of Illinois, Urbana-Champaign&lt;/i&gt;&lt;br /&gt;Sarah D'Angelo, &lt;i&gt;Northwestern University&lt;/i&gt;&lt;br /&gt;Sarah Mcroberts, &lt;i&gt;University of Minnesota - Twin Cities&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;b&gt;&lt;u&gt;Machine Learning&lt;/u&gt;&lt;/b&gt;&lt;br /&gt;Aude Genevay, &lt;i&gt;Fondation Sciences Mathématiques de Paris&lt;/i&gt;&lt;br /&gt;Dustin Tran, &lt;i&gt;Columbia University&lt;/i&gt;&lt;br /&gt;Jamie Hayes, &lt;i&gt;University College London&lt;/i&gt;&lt;br /&gt;Martin Arjovsky, &lt;i&gt;New York University&lt;/i&gt;&lt;br /&gt;Taco Cohen, &lt;i&gt;University of Amsterdam&lt;/i&gt;&lt;br /&gt;Yuhuai Wu, &lt;i&gt;University of Toronto&lt;/i&gt;&lt;br /&gt;Yunye Gong, &lt;i&gt;Cornell University&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;b&gt;&lt;u&gt;Machine Perception, Speech Technology and Computer Vision&lt;/u&gt;&lt;/b&gt;&lt;br /&gt;Franziska Müller, &lt;i&gt;Saarland University - Saarbrücken GSCS and MPI Institute for Informatics&lt;/i&gt;&lt;br /&gt;George Trigeorgis, &lt;i&gt;Imperial College London&lt;/i&gt;&lt;br /&gt;Iro Armeni, &lt;i&gt;Stanford University&lt;/i&gt;&lt;br /&gt;Saining Xie, &lt;i&gt;University of California, San Diego&lt;/i&gt;&lt;br /&gt;Yu-Chuan Su, &lt;i&gt;University of Texas, Austin&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;b&gt;&lt;u&gt;Natural Language Processing&lt;/u&gt;&lt;/b&gt;&lt;br /&gt;Jianpeng Cheng, The University of Edinburgh&lt;br /&gt;Kevin Clark, Stanford University&lt;br /&gt;Tim Rocktaschel, University College London&lt;br /&gt;&lt;br /&gt;&lt;b&gt;&lt;u&gt;Privacy and Security&lt;/u&gt;&lt;/b&gt;&lt;br /&gt;Romain Gay, &lt;i&gt;ENS - École Normale Supérieure&lt;/i&gt;&lt;br /&gt;Xi He, &lt;i&gt;Duke University&lt;/i&gt;&lt;br /&gt;Yupeng Zhang, &lt;i&gt;University of Maryland, College Park&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;b&gt;&lt;u&gt;Programming Languages and Software Engineering&lt;/u&gt;&lt;/b&gt;&lt;br /&gt;Christoffer Quist Adamsen, &lt;i&gt;Aarhus Universit&lt;/i&gt;y&lt;br /&gt;Muhammad Ali Gulzar, &lt;i&gt;University of California, Los Angeles&lt;/i&gt;&lt;br /&gt;Oded Padon, &lt;i&gt;Tel-Aviv University&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;b&gt;&lt;u&gt;Structured Data and Database Management&lt;/u&gt;&lt;/b&gt;&lt;br /&gt;Amir Shaikhha, &lt;i&gt;EPFL CS&lt;/i&gt;&lt;br /&gt;Jingbo Shang, &lt;i&gt;University of Illinois, Urbana-Champaign&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;b&gt;&lt;u&gt;Systems and Networking&lt;/u&gt;&lt;/b&gt;&lt;br /&gt;Ahmed M. Said Mohamed Tawfik Issa, &lt;i&gt;Georgia Institute of Technology&lt;/i&gt;&lt;br /&gt;Khanh Nguyen, &lt;i&gt;University of California, Irvine&lt;/i&gt;&lt;br /&gt;Radhika Mittal, &lt;i&gt;University of California, Berkeley&lt;/i&gt;&lt;br /&gt;Ryan Beckett, &lt;i&gt;Princeton University&lt;/i&gt;&lt;div class="feedflare"&gt;
&lt;a href="http://feeds.feedburner.com/~ff/blogspot/gJZg?a=nLPbXcp_GKc:xrs2APpm2fI:yIl2AUoC8zA"&gt;&lt;img src="http://feeds.feedburner.com/~ff/blogspot/gJZg?d=yIl2AUoC8zA" border="0"&gt;&lt;/img&gt;&lt;/a&gt;
&lt;/div&gt;&lt;img src="http://feeds.feedburner.com/~r/blogspot/gJZg/~4/nLPbXcp_GKc" height="1" width="1" alt=""/&gt;</content><link rel="replies" type="application/atom+xml" href="http://research.googleblog.com/feeds/6501613459508973924/comments/default" title="Post Comments" /><link rel="replies" type="text/html" href="http://research.googleblog.com/2017/04/announcing-2017-google-phd-fellows-for.html#comment-form" title="0 Comments" /><link rel="edit" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/6501613459508973924" /><link rel="self" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/6501613459508973924" /><link rel="alternate" type="text/html" href="http://feedproxy.google.com/~r/blogspot/gJZg/~3/nLPbXcp_GKc/announcing-2017-google-phd-fellows-for.html" title="Announcing the 2017 Google PhD Fellows for North America, Europe and the Middle East" /><author><name>Research Blog</name><uri>https://plus.google.com/101673966767287570260</uri><email>noreply@blogger.com</email><gd:image rel="http://schemas.google.com/g/2005#thumbnail" width="32" height="32" src="//lh5.googleusercontent.com/-GXTJ2sdBxaM/AAAAAAAAAAI/AAAAAAAAA14/2q3ThT_9JIY/s512-c/photo.jpg" /></author><thr:total>0</thr:total><gd:extendedProperty name="commentSource" value="1" /><gd:extendedProperty name="commentModerationMode" value="FILTERED_POSTMOD" /><feedburner:origLink>http://research.googleblog.com/2017/04/announcing-2017-google-phd-fellows-for.html</feedburner:origLink></entry><entry><id>tag:blogger.com,1999:blog-21224994.post-1857743015124551433</id><published>2017-04-07T11:00:00.000-07:00</published><updated>2017-04-07T16:46:16.918-07:00</updated><category scheme="http://www.blogger.com/atom/ns#" term="Chemistry" /><category scheme="http://www.blogger.com/atom/ns#" term="Deep Learning" /><category scheme="http://www.blogger.com/atom/ns#" term="Machine Learning" /><title type="text">Predicting Properties of Molecules with Machine Learning</title><content type="html">&lt;span class="byline-author"&gt;Posted by George Dahl, Research Scientist, Google Brain Team&lt;/span&gt;&lt;br /&gt;&lt;br /&gt;Recently there have been many exciting applications of machine learning (ML) to chemistry, particularly in chemical search problems, from &lt;a href="http://pubs.acs.org/doi/abs/10.1021/ci500747n"&gt;drug discovery&lt;/a&gt; and &lt;a href="https://tomkat.stanford.edu/research/designing-better-battery-machine-learning"&gt;battery design&lt;/a&gt; to finding better &lt;a href="http://www.nature.com/nmat/journal/v15/n10/abs/nmat4717.html"&gt;OLEDs&lt;/a&gt; and &lt;a href="http://pubs.rsc.org/is/content/articlelanding/2016/cp/c6cp05711j#!divAbstract"&gt;catalysts&lt;/a&gt;. Historically, chemists have used numerical approximations to Schrödinger’s equation, such as &lt;a href="https://en.wikipedia.org/wiki/Density_functional_theory"&gt;Density Functional Theory&lt;/a&gt; (DFT), in these sorts of chemical searches. However, the computational cost of these approximations limits the size of the search. In the hope of enabling larger searches, several research groups have created ML models to predict chemical properties using training data generated by DFT (e.g. &lt;a href="https://journals.aps.org/prl/abstract/10.1103/PhysRevLett.108.058301"&gt;Rupp et al.&lt;/a&gt; and &lt;a href="https://journals.aps.org/prl/abstract/10.1103/PhysRevLett.98.146401"&gt;Behler and Parrinello&lt;/a&gt;). Expanding upon this previous work, we have been applying various modern ML methods to the &lt;a href="http://www.quantum-machine.org/datasets/#qm9"&gt;QM9&lt;/a&gt; &lt;a href="http://www.nature.com/articles/sdata201422"&gt;benchmark&lt;/a&gt; –a public collection of molecules paired with DFT-computed electronic, thermodynamic, and vibrational properties.&lt;br /&gt;&lt;br /&gt;We have recently posted two papers describing our research in this area that grew out of a collaboration between the &lt;a href="http://g.co/brain"&gt;Google Brain team&lt;/a&gt;, the Google Accelerated Science team, &lt;a href="https://deepmind.com/"&gt;DeepMind&lt;/a&gt;, and the &lt;a href="http://www.chemie.unibas.ch/~anatole/"&gt;University of Basel&lt;/a&gt;. The &lt;a href="https://arxiv.org/abs/1702.05532"&gt;first paper&lt;/a&gt; includes a new featurization of molecules and a systematic assessment of a multitude of machine learning methods on the QM9 benchmark. After trying many existing approaches on this benchmark, we worked on improving the most promising deep neural network models. &lt;br /&gt;&lt;br /&gt;The resulting second paper, “&lt;a href="https://arxiv.org/abs/1704.01212"&gt;Neural Message Passing for Quantum Chemistry&lt;/a&gt;,” describes a model family called Message Passing Neural Networks (MPNNs), which are defined abstractly enough to include many previous neural net models that are invariant to graph symmetries. We developed novel variations within the MPNN family which significantly outperform all baseline methods on the QM9 benchmark, with improvements of nearly a factor of four on some targets. &lt;br /&gt;&lt;br /&gt;One reason molecular data is so interesting from a machine learning standpoint is that one natural representation of a molecule is as a graph with &lt;a href="https://en.wikipedia.org/wiki/Graph_theory"&gt;atoms as nodes and bonds as edges&lt;/a&gt;. Models that can leverage inherent symmetries in data will tend to generalize better — part of the success of convolutional neural networks on images is due to their ability to incorporate our prior knowledge about the invariances of image data (e.g. a picture of a dog shifted to the left is still a picture of a dog). Invariance to graph symmetries is a particularly desirable property for machine learning models that operate on graph data, and there has been a lot of interesting research in this area as well (e.g. &lt;a href="https://arxiv.org/abs/1511.05493"&gt;Li et al.&lt;/a&gt;, &lt;a href="http://papers.nips.cc/paper/5954-convolutional-networks-on-graphs-for-learning-molecular-fingerprints.pdf"&gt;Duvenaud et al.&lt;/a&gt;, &lt;a href="https://arxiv.org/abs/1603.00856"&gt;Kearnes et al.&lt;/a&gt;, &lt;a href="http://papers.nips.cc/paper/6081-convolutional-neural-networks-on-graphs-with-fast-localized-spectral-filtering.pdf"&gt;Defferrard et al.&lt;/a&gt;). However, despite this progress, much work remains. We would like to find the best versions of these models for chemistry (and other) applications and map out the connections between different models proposed in the literature.&lt;br /&gt;&lt;br /&gt;Our MPNNs set a new state of the art for predicting all 13 chemical properties in QM9. On this particular set of molecules, our model can predict 11 of these properties accurately enough to potentially be useful to chemists, but up to 300,000 times faster than it would take to simulate them using DFT. However, much work remains to be done before MPNNs can be of real practical use to chemists. In particular, MPNNs must be applied to a significantly more diverse set of molecules (e.g. larger or with a more varied set of heavy atoms) than exist in QM9. Of course, even with a realistic training set, generalization to very different molecules could still be poor. Overcoming both of these challenges will involve making progress on questions–such as generalization–that are at the heart of machine learning research.&lt;br /&gt;&lt;br /&gt;Predicting the properties of molecules is a practically important problem that both benefits from advanced machine learning techniques and presents interesting fundamental research challenges for learning algorithms. Eventually, such predictions could aid the design of new medicines and materials that benefit humanity. At Google, we feel that it’s important to disseminate our research and to help train new researchers in machine learning. As such, we are delighted that the first and second authors of our MPNN paper are &lt;a href="https://research.google.com/teams/brain/residency/"&gt;Google Brain residents&lt;/a&gt;. &lt;div class="feedflare"&gt;
&lt;a href="http://feeds.feedburner.com/~ff/blogspot/gJZg?a=6jNP7Hiyr4k:uGwuCO19rP8:yIl2AUoC8zA"&gt;&lt;img src="http://feeds.feedburner.com/~ff/blogspot/gJZg?d=yIl2AUoC8zA" border="0"&gt;&lt;/img&gt;&lt;/a&gt;
&lt;/div&gt;&lt;img src="http://feeds.feedburner.com/~r/blogspot/gJZg/~4/6jNP7Hiyr4k" height="1" width="1" alt=""/&gt;</content><link rel="replies" type="application/atom+xml" href="http://research.googleblog.com/feeds/1857743015124551433/comments/default" title="Post Comments" /><link rel="replies" type="text/html" href="http://research.googleblog.com/2017/04/predicting-properties-of-molecules-with.html#comment-form" title="0 Comments" /><link rel="edit" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/1857743015124551433" /><link rel="self" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/1857743015124551433" /><link rel="alternate" type="text/html" href="http://feedproxy.google.com/~r/blogspot/gJZg/~3/6jNP7Hiyr4k/predicting-properties-of-molecules-with.html" title="Predicting Properties of Molecules with Machine Learning" /><author><name>Research Blog</name><uri>https://plus.google.com/101673966767287570260</uri><email>noreply@blogger.com</email><gd:image rel="http://schemas.google.com/g/2005#thumbnail" width="32" height="32" src="//lh5.googleusercontent.com/-GXTJ2sdBxaM/AAAAAAAAAAI/AAAAAAAAA14/2q3ThT_9JIY/s512-c/photo.jpg" /></author><thr:total>0</thr:total><gd:extendedProperty name="commentSource" value="1" /><gd:extendedProperty name="commentModerationMode" value="FILTERED_POSTMOD" /><feedburner:origLink>http://research.googleblog.com/2017/04/predicting-properties-of-molecules-with.html</feedburner:origLink></entry><entry><id>tag:blogger.com,1999:blog-21224994.post-5091316152211278759</id><published>2017-04-06T15:17:00.000-07:00</published><updated>2017-04-07T17:03:30.376-07:00</updated><category scheme="http://www.blogger.com/atom/ns#" term="Collaboration" /><category scheme="http://www.blogger.com/atom/ns#" term="Machine Learning" /><category scheme="http://www.blogger.com/atom/ns#" term="On-device Learning" /><category scheme="http://www.blogger.com/atom/ns#" term="optimization" /><title type="text">Federated Learning: Collaborative Machine Learning without Centralized Training Data </title><content type="html">&lt;span class="byline-author"&gt;Posted by Brendan McMahan and Daniel Ramage, Research Scientists&lt;/span&gt;&lt;br /&gt;&lt;br /&gt;Standard machine learning approaches require centralizing the training data on one machine or in a datacenter. And Google has built one of the most secure and robust cloud infrastructures for processing this data to make our services better. Now for models trained from user interaction with mobile devices, we're introducing an additional approach: &lt;i&gt;Federated Learning&lt;/i&gt;. &lt;br /&gt;&lt;br /&gt;Federated Learning enables mobile phones to collaboratively learn a shared prediction model while keeping all the training data on device, decoupling the ability to do machine learning from the need to store the data in the cloud. This goes beyond the use of local models that make predictions on mobile devices (like the &lt;a href="https://developers.google.com/vision/"&gt;Mobile Vision API&lt;/a&gt; and &lt;a href="https://research.googleblog.com/2017/02/on-device-machine-intelligence.html"&gt;On-Device Smart Reply&lt;/a&gt;) by bringing model &lt;i&gt;training&lt;/i&gt; to the device as well.&lt;br /&gt;&lt;br /&gt;It works like this: your device downloads the current model, improves it by learning from data on your phone, and then summarizes the changes as a small focused update. Only this update to the model is sent to the cloud, using encrypted communication, where it is immediately averaged with other user updates to improve the shared model. All the training data remains on your device, and no individual updates are stored in the cloud. &lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://1.bp.blogspot.com/-K65Ed68KGXk/WOa9jaRWC6I/AAAAAAAABsM/gglycD_anuQSp-i67fxER1FOlVTulvV2gCLcB/s1600/FederatedLearning_FinalFiles_Flow%2BChart1.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="360" src="https://1.bp.blogspot.com/-K65Ed68KGXk/WOa9jaRWC6I/AAAAAAAABsM/gglycD_anuQSp-i67fxER1FOlVTulvV2gCLcB/s640/FederatedLearning_FinalFiles_Flow%2BChart1.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Your phone personalizes the model locally, based on your usage (A). Many users' updates are aggregated (B) to form a consensus change (C) to the shared model, after which the procedure is repeated.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;Federated Learning allows for smarter models, lower latency, and less power consumption, all while ensuring privacy. And this approach has another immediate benefit: in addition to providing an update to the shared model, the improved model on your phone can also be used immediately, powering experiences personalized by the way you use your phone.&lt;br /&gt;&lt;br /&gt;We're currently testing Federated Learning in &lt;a href="https://blog.google/products/search/gboard-now-on-android/"&gt;Gboard on Android&lt;/a&gt;, the Google Keyboard. When Gboard shows a suggested query, your phone locally stores information about the current context and whether you clicked the suggestion. Federated Learning processes that history on-device to suggest improvements to the next iteration of Gboard’s query suggestion model.&lt;br /&gt;&lt;div class="separator" style="clear: both; text-align: center;"&gt;&lt;a href="https://1.bp.blogspot.com/-W-husQJfa7s/WObDco6Ql0I/AAAAAAAABso/ERk3Q3mM2xILzEgMa0RMi5UJED7VDLYCACLcB/s1600/2017-04-06.gif" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"&gt;&lt;img border="0" height="358" src="https://1.bp.blogspot.com/-W-husQJfa7s/WObDco6Ql0I/AAAAAAAABso/ERk3Q3mM2xILzEgMa0RMi5UJED7VDLYCACLcB/s640/2017-04-06.gif" width="640" /&gt;&lt;/a&gt;&lt;/div&gt;To make Federated Learning possible, we had to overcome many algorithmic and technical challenges. In a typical machine learning system, an optimization algorithm like &lt;a href="https://en.wikipedia.org/wiki/Stochastic_gradient_descent"&gt;Stochastic Gradient Descent&lt;/a&gt; (SGD) runs on a large dataset partitioned homogeneously across servers in the cloud. Such highly iterative algorithms require low-latency, high-throughput connections to the training data. But in the Federated Learning setting, the data is distributed across millions of devices in a highly uneven fashion. In addition, these devices have significantly higher-latency, lower-throughput connections and are only intermittently available for training.&lt;br /&gt;&lt;br /&gt;These bandwidth and latency limitations motivate our &lt;a href="https://arxiv.org/abs/1602.05629"&gt;Federated Averaging algorithm&lt;/a&gt;, which can train deep networks using 10-100x less communication compared to a naively federated version of SGD. The key idea is to use the powerful processors in modern mobile devices to compute higher quality updates than simple gradient steps. Since it takes fewer iterations of high-quality updates to produce a good model, training can use much less communication. As upload speeds are typically &lt;a href="http://www.speedtest.net/reports/united-states/"&gt;much slower&lt;/a&gt; than download speeds, we also developed a novel way to reduce upload communication costs up to another 100x by &lt;a href="https://arxiv.org/abs/1610.05492"&gt;compressing updates&lt;/a&gt; using random rotations and quantization. While these approaches are focused on training deep networks, we've also &lt;a href="https://arxiv.org/abs/1610.02527"&gt;designed algorithms&lt;/a&gt; for high-dimensional sparse convex models which excel on problems like click-through-rate prediction.&lt;br /&gt;&lt;br /&gt;Deploying this technology to millions of heterogenous phones running Gboard requires a sophisticated technology stack. On device training uses a miniature version of &lt;a href="https://www.tensorflow.org/"&gt;TensorFlow&lt;/a&gt;. Careful scheduling ensures training happens only when the device is idle, plugged in, and on a free wireless connection, so there is no impact on the phone's performance.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://3.bp.blogspot.com/-sb40Lg5MchE/WOa92rSrXVI/AAAAAAAABsU/edvH01nn2SMCnJle8mLnHT_hCa-xXrnsACLcB/s1600/02_Personalization%2Bsleeping.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="360" src="https://3.bp.blogspot.com/-sb40Lg5MchE/WOa92rSrXVI/AAAAAAAABsU/edvH01nn2SMCnJle8mLnHT_hCa-xXrnsACLcB/s640/02_Personalization%2Bsleeping.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Your phone participates in Federated Learning only &lt;br /&gt;when it won't negatively impact your experience.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;The system then needs to communicate and aggregate the model updates in a secure, efficient, scalable, and fault-tolerant way. It's only the combination of research with this infrastructure that makes the benefits of Federated Learning possible.&lt;br /&gt;&lt;br /&gt;Federated learning works without the need to store user data in the cloud, but we're not stopping there. We've developed a &lt;a href="http://eprint.iacr.org/2017/281"&gt;Secure Aggregation protocol&lt;/a&gt; that uses cryptographic techniques so a coordinating server can only decrypt the average update if 100s or 1000s of users have participated — no individual phone's update can be inspected before averaging. It's the first protocol of its kind that is practical for deep-network-sized problems and real-world connectivity constraints. We designed Federated Averaging so the coordinating server only needs the average update, which allows Secure Aggregation to be used; however the protocol is general and can be applied to other problems as well. We're working hard on a production implementation of this protocol and expect to deploy it for Federated Learning applications in the near future.&lt;br /&gt;&lt;br /&gt;Our work has only scratched the surface of what is possible. Federated Learning can't solve all machine learning problems (for example, learning to &lt;a href="https://research.googleblog.com/2016/08/improving-inception-and-image.html"&gt;recognize different dog breeds&lt;/a&gt; by training on carefully labeled examples), and for many other models the necessary training data is already stored in the cloud (like training spam filters for Gmail). So Google will continue to advance the state-of-the-art for cloud-based ML, but we are also committed to ongoing research to expand the range of problems we can solve with Federated Learning. Beyond Gboard query suggestions, for example, we hope to improve the language models that power your keyboard based on what you actually type on your phone (which can have a style all its own) and photo rankings based on what kinds of photos people look at, share, or delete.&lt;br /&gt;&lt;br /&gt;Applying Federated Learning requires machine learning practitioners to adopt new tools and a new way of thinking: model development, training, and evaluation with no direct access to or labeling of raw data, with communication cost as a limiting factor. We believe the user benefits of Federated Learning make tackling the technical challenges worthwhile, and are publishing our work with hopes of a widespread conversation within the machine learning community.&lt;br /&gt;&lt;br /&gt;&lt;b&gt;Acknowledgements&lt;/b&gt;&lt;br /&gt;This post reflects the work of many people in Google Research, including Blaise Agüera y Arcas, Galen Andrew, Dave Bacon, Keith Bonawitz, Chris Brumme, Arlie Davis, Jac de Haan, Hubert Eichner, Wolfgang Grieskamp, Wei Huang, Vladimir Ivanov, Chloé Kiddon, Jakub Konečný, Nicholas Kong, Ben Kreuter, Alison Lentz, Stefano Mazzocchi, Sarvar Patel, Martin Pelikan, Aaron Segal, Karn Seth, Ananda Theertha Suresh, Iulia Turc, Felix Yu, Antonio Marcedone and our partners in the Gboard team.&lt;div class="feedflare"&gt;
&lt;a href="http://feeds.feedburner.com/~ff/blogspot/gJZg?a=9PENsme-eHg:iqaanPMka1Q:yIl2AUoC8zA"&gt;&lt;img src="http://feeds.feedburner.com/~ff/blogspot/gJZg?d=yIl2AUoC8zA" border="0"&gt;&lt;/img&gt;&lt;/a&gt;
&lt;/div&gt;&lt;img src="http://feeds.feedburner.com/~r/blogspot/gJZg/~4/9PENsme-eHg" height="1" width="1" alt=""/&gt;</content><link rel="replies" type="application/atom+xml" href="http://research.googleblog.com/feeds/5091316152211278759/comments/default" title="Post Comments" /><link rel="replies" type="text/html" href="http://research.googleblog.com/2017/04/federated-learning-collaborative.html#comment-form" title="0 Comments" /><link rel="edit" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/5091316152211278759" /><link rel="self" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/5091316152211278759" /><link rel="alternate" type="text/html" href="http://feedproxy.google.com/~r/blogspot/gJZg/~3/9PENsme-eHg/federated-learning-collaborative.html" title="Federated Learning: Collaborative Machine Learning without Centralized Training Data " /><author><name>Research Blog</name><uri>https://plus.google.com/101673966767287570260</uri><email>noreply@blogger.com</email><gd:image rel="http://schemas.google.com/g/2005#thumbnail" width="32" height="32" src="//lh5.googleusercontent.com/-GXTJ2sdBxaM/AAAAAAAAAAI/AAAAAAAAA14/2q3ThT_9JIY/s512-c/photo.jpg" /></author><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://1.bp.blogspot.com/-K65Ed68KGXk/WOa9jaRWC6I/AAAAAAAABsM/gglycD_anuQSp-i67fxER1FOlVTulvV2gCLcB/s72-c/FederatedLearning_FinalFiles_Flow%2BChart1.png" height="72" width="72" /><thr:total>0</thr:total><gd:extendedProperty name="commentSource" value="1" /><gd:extendedProperty name="commentModerationMode" value="FILTERED_POSTMOD" /><feedburner:origLink>http://research.googleblog.com/2017/04/federated-learning-collaborative.html</feedburner:origLink></entry><entry><id>tag:blogger.com,1999:blog-21224994.post-8551536041968103505</id><published>2017-04-06T09:00:00.000-07:00</published><updated>2017-04-06T09:33:16.507-07:00</updated><category scheme="http://www.blogger.com/atom/ns#" term="Publications" /><category scheme="http://www.blogger.com/atom/ns#" term="Security and Privacy" /><category scheme="http://www.blogger.com/atom/ns#" term="WWW" /><title type="text">Keeping fake listings off Google Maps</title><content type="html">&lt;span class="byline-author"&gt;Posted by Doug Grundman, Maps Anti-Abuse, and Kurt Thomas, Security &amp;amp; Anti-Abuse Research&lt;/span&gt;&lt;br /&gt;&lt;br /&gt;&lt;i&gt;(Crossposted on the &lt;a href="https://security.googleblog.com/2017/04/new-research-keeping-fake-listings-off.html"&gt;Google Security blog&lt;/a&gt;)&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;&lt;a href="https://www.google.com/business/"&gt;Google My Business&lt;/a&gt; enables millions of business owners to create listings and share information about their business on Google Maps and Search, making sure everything is up-to-date and accurate for their customers. Unfortunately, some actors attempt to abuse this service to register fake listings in order to defraud legitimate business owners, or to &lt;a href="https://www.nytimes.com/2016/01/31/business/fake-online-locksmiths-may-be-out-to-pick-your-pocket-too.html?_r=1"&gt;charge exorbitant service fees for services&lt;/a&gt;.&lt;br /&gt;&lt;br /&gt;Over a year ago, we teamed up with the University of California, San Diego to research the actors behind fake listings, in order to improve our products and keep our users safe. The full report, “&lt;a href="https://research.google.com/pubs/pub45976.html"&gt;Pinning Down Abuse on Google Maps&lt;/a&gt;”, will be presented tomorrow at the 2017 &lt;a href="http://www2017.com.au/"&gt;International World Wide Web Conference&lt;/a&gt;.&lt;br /&gt;&lt;br /&gt;Our study shows that fewer than 0.5% of local searches lead to fake listings. We’ve also improved how we verify new businesses, which has reduced the number of fake listings by 70% from its all-time peak back in June 2015.&lt;br /&gt;&lt;br /&gt;&lt;b&gt;What is a fake listing?&lt;/b&gt;&lt;br /&gt;For over a year, we tracked the bad actors behind fake listings.  Unlike email-based scams &lt;a href="https://security.googleblog.com/2015/09/new-research-underground-market-fueling.html"&gt;selling knock-off products online&lt;/a&gt;, local listing scams require physical proximity to potential victims. This fundamentally changes both the scale and types of abuse possible.&lt;br /&gt;&lt;br /&gt;Bad actors posing as locksmiths, plumbers, electricians, and other contractors were the most common source of abuse—roughly 2 out of 5 fake listings. The actors operating these fake listings would cycle through non-existent postal addresses and disposable VoIP phone numbers even as their listings were discovered and disabled. The purported addresses for these businesses were irrelevant as the contractors would travel directly to potential victims.&lt;br /&gt;&lt;br /&gt;Another 1 in 10 fake listings belonged to real businesses that bad actors had improperly claimed ownership over, such as hotels and restaurants. While making a reservation or ordering a meal was indistinguishable from the real thing, behind the scenes, the bad actors would deceive the actual business into paying referral fees for organic interest.&lt;br /&gt;&lt;br /&gt;&lt;b&gt;How does Google My Business verify information?&lt;/b&gt;&lt;br /&gt;Google My Business currently verifies the information provided by business owners before making it available to users. For freshly created listings, we physically mail a postcard to the new listings’ address to ensure the location really exists. For businesses changing owners, we make an automated call to the listing’s phone number to verify the change.&lt;br /&gt;&lt;div class="separator" style="clear: both; text-align: center;"&gt;&lt;a href="https://2.bp.blogspot.com/-deN6n5cLOfQ/WOVjIudB8QI/AAAAAAAABr0/5w-M6H2GToUVP2NelIqTOvXvELPHFUCZwCLcB/s1600/image00.png" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"&gt;&lt;img border="0" height="234" src="https://2.bp.blogspot.com/-deN6n5cLOfQ/WOVjIudB8QI/AAAAAAAABr0/5w-M6H2GToUVP2NelIqTOvXvELPHFUCZwCLcB/s640/image00.png" width="640" /&gt;&lt;/a&gt;&lt;/div&gt;Unfortunately, our research showed that these processes can be abused to get fake listings on Google Maps. Fake contractors would request hundreds of postcard verifications to non-existent suites at a single address, such as 123 Main St #456 and 123 Main St #789, or to stores that provided PO boxes. Alternatively, a phishing attack could maliciously repurpose freshly verified business listings by tricking  the legitimate owner into sharing verification information sent either by phone or postcard.&lt;br /&gt;&lt;b&gt;&lt;br /&gt;&lt;/b&gt;&lt;b&gt;Keeping deceptive businesses out — by the numbers&lt;/b&gt;&lt;br /&gt;Leveraging our study’s findings, we’ve made significant changes to how we verify addresses and are even &lt;a href="https://news.fastcompany.com/google-is-testing-out-a-merchant-verification-program-to-weed-out-fake-businesses-4020905"&gt;piloting an advanced verification process&lt;/a&gt; for locksmiths and plumbers. Improvements we’ve made include prohibiting bulk registrations at most addresses, preventing businesses from relocating impossibly far from their original address without additional verification, and detecting and ignoring intentionally mangled text in address fields designed to confuse our algorithms. We have also adapted our anti-spam machine learning systems to detect data discrepancies common to fake or deceptive listings.&lt;br /&gt;&lt;br /&gt;Combined, here’s how these defenses stack up:&lt;br /&gt;&lt;br /&gt;&lt;ul&gt;&lt;li&gt;We detect and disable 85% of fake listings before they even appear on Google Maps.&lt;/li&gt;&lt;li&gt;We’ve reduced the number of abusive listings by 70% from its peak back in June 2015.&lt;/li&gt;&lt;li&gt;We’ve also reduced the number of impressions to abusive listings by 70%.&lt;/li&gt;&lt;/ul&gt;&lt;br /&gt;As we’ve shown, verifying local information comes with a number of unique anti-abuse challenges. While fake listings may slip through our defenses from time to time, we are constantly improving our systems to better serve both users and business owners.&lt;br /&gt;&lt;br /&gt;&lt;div class="feedflare"&gt;
&lt;a href="http://feeds.feedburner.com/~ff/blogspot/gJZg?a=-M-TxFy2MYQ:oWni4bWyaZo:yIl2AUoC8zA"&gt;&lt;img src="http://feeds.feedburner.com/~ff/blogspot/gJZg?d=yIl2AUoC8zA" border="0"&gt;&lt;/img&gt;&lt;/a&gt;
&lt;/div&gt;&lt;img src="http://feeds.feedburner.com/~r/blogspot/gJZg/~4/-M-TxFy2MYQ" height="1" width="1" alt=""/&gt;</content><link rel="replies" type="application/atom+xml" href="http://research.googleblog.com/feeds/8551536041968103505/comments/default" title="Post Comments" /><link rel="replies" type="text/html" href="http://research.googleblog.com/2017/04/keeping-fake-listings-off-google-maps.html#comment-form" title="0 Comments" /><link rel="edit" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/8551536041968103505" /><link rel="self" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/8551536041968103505" /><link rel="alternate" type="text/html" href="http://feedproxy.google.com/~r/blogspot/gJZg/~3/-M-TxFy2MYQ/keeping-fake-listings-off-google-maps.html" title="Keeping fake listings off Google Maps" /><author><name>Research Blog</name><uri>https://plus.google.com/101673966767287570260</uri><email>noreply@blogger.com</email><gd:image rel="http://schemas.google.com/g/2005#thumbnail" width="32" height="32" src="//lh5.googleusercontent.com/-GXTJ2sdBxaM/AAAAAAAAAAI/AAAAAAAAA14/2q3ThT_9JIY/s512-c/photo.jpg" /></author><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://2.bp.blogspot.com/-deN6n5cLOfQ/WOVjIudB8QI/AAAAAAAABr0/5w-M6H2GToUVP2NelIqTOvXvELPHFUCZwCLcB/s72-c/image00.png" height="72" width="72" /><thr:total>0</thr:total><gd:extendedProperty name="commentSource" value="1" /><gd:extendedProperty name="commentModerationMode" value="FILTERED_POSTMOD" /><feedburner:origLink>http://research.googleblog.com/2017/04/keeping-fake-listings-off-google-maps.html</feedburner:origLink></entry><entry><id>tag:blogger.com,1999:blog-21224994.post-3534392145444604799</id><published>2017-04-05T01:00:00.000-07:00</published><updated>2017-04-05T01:00:17.184-07:00</updated><category scheme="http://www.blogger.com/atom/ns#" term="ACM" /><category scheme="http://www.blogger.com/atom/ns#" term="Awards" /><category scheme="http://www.blogger.com/atom/ns#" term="Publications" /><category scheme="http://www.blogger.com/atom/ns#" term="WWW" /><title type="text">And the award goes to...</title><content type="html">&lt;span class="byline-author"&gt;Posted by Evgeniy Gabrilovich, Senior Staff Research Scientist, Google Research, and WWW 2017 Technical Program Co-Chair&lt;/span&gt;&lt;br /&gt;&lt;br /&gt;Today, Google's &lt;a href="https://research.google.com/pubs/AndreiBroder.html"&gt;Andrei Broder&lt;/a&gt;, &lt;a href="https://scholar.google.com/citations?user=J_XhIsgAAAAJ"&gt;Ravi Kumar&lt;/a&gt;, &lt;a href="https://research.google.com/pubs/PrabhakarRaghavan.html"&gt;Prabhakar Raghavan&lt;/a&gt;, &lt;a href="http://dblp.uni-trier.de/pers/hd/r/Rajagopalan:Sridhar"&gt;Sridhar Rajagopalan&lt;/a&gt;, and &lt;a href="https://research.google.com/pubs/author8038.html"&gt;Andrew Tomkin&lt;/a&gt;s, along with their coauthors, Farzin Maghoul, Raymie Stata, and Janet Wiener, have received the prestigious 2017 Seoul Test of Time Award for their classic paper “&lt;a href="http://wwwconference.org/proceedings/www9/160/160.html"&gt;Graph Structure in the Web&lt;/a&gt;”. This award is given to the authors of a previous World Wide Web conference paper that has demonstrated significant scientific, technical, or social impact over the years. The first award, introduced in 2015, &lt;a href="https://research.googleblog.com/2015/05/sergey-and-larry-awarded-seoul-test-of.html"&gt;was given to Google founders Larry Page and Sergey Brin&lt;/a&gt;. &lt;br /&gt;&lt;br /&gt;Originally presented in 2000 at the 9th WWW conference in Amsterdam, “&lt;a href="http://wwwconference.org/proceedings/www9/160/160.html"&gt;Graph Structure in the Web&lt;/a&gt;” represents the seminal study of the structure of the World Wide Web. At the time of publication, it received the Best Paper Award from the WWW conference, and in the following 17 years proved to be highly influential, accumulating over &lt;a href="https://scholar.google.com/scholar?cites=9024688421763252029&amp;amp;as_sdt=2005&amp;amp;sciodt=0,5&amp;amp;hl=en"&gt;3,500 citations&lt;/a&gt;. &lt;br /&gt;&lt;br /&gt;The paper made two major contributions to the study of the structure of the Internet. First, it reported the results of a very large scale experiment to confirm that the &lt;a href="https://en.wikipedia.org/wiki/Directed_graph#Indegree_and_outdegree"&gt;indegree&lt;/a&gt; of Web nodes is distributed according to a power law. To wit, the probability that a node of the Web graph has &lt;i&gt;i&lt;/i&gt; incoming links is roughly proportional to &lt;i&gt;1/i&lt;sup&gt;2.1&lt;/sup&gt;&lt;/i&gt;. Second, in contrast to previous research that assumed the Web to be almost fully connected, “&lt;a href="http://wwwconference.org/proceedings/www9/160/160.html"&gt;Graph Structure in the Web&lt;/a&gt;” described a much more elaborate structure of the Web, which since then has been depicted with the iconic “bowtie” shape:&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://1.bp.blogspot.com/-1xICW7QwoEw/WOPWgOmrNEI/AAAAAAAABrU/_jI6HGrbEPQi2AOi4aoX6HGsJSk-XzFDACLcB/s1600/image01.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="578" src="https://1.bp.blogspot.com/-1xICW7QwoEw/WOPWgOmrNEI/AAAAAAAABrU/_jI6HGrbEPQi2AOi4aoX6HGsJSk-XzFDACLcB/s640/image01.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Original “bowtie” schematic from “Graph Structure in the Web”&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;The authors presented a refined model of the Web graph, and described several characteristic classes of Web pages:&lt;br /&gt;&lt;ul&gt;&lt;li&gt;the &lt;i&gt;strongly connected core component&lt;/i&gt;, where each page is reachable from any other page,&lt;/li&gt;&lt;li&gt;the so-called &lt;i&gt;IN&lt;/i&gt; and &lt;i&gt;OUT clusters&lt;/i&gt;, which only have unidirectional paths to or from the core,&lt;/li&gt;&lt;li&gt;&lt;i&gt;tendrils&lt;/i&gt; dangling from the two clusters, and tubes connecting the clusters while bypassing the core, and finally&lt;/li&gt;&lt;li&gt;&lt;i&gt;disconnected components&lt;/i&gt;, which are isolated from the rest of the graph.&lt;/li&gt;&lt;/ul&gt;Whereas the core component is fully connected and each node can be reached from any other node, Broder et al. discovered that as a whole the Web is much more loosely connected than previously believed, while the probability that any two given pages can be reached from one another is just under &lt;i&gt;1/4&lt;/i&gt;.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://1.bp.blogspot.com/-IPbpaM10Uts/WOPWpj8JKzI/AAAAAAAABrY/4BOGDXFmarAe-wy1hIQQB6ZWt0DMwQdKQCLcB/s1600/image00.jpg" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="480" src="https://1.bp.blogspot.com/-IPbpaM10Uts/WOPWpj8JKzI/AAAAAAAABrY/4BOGDXFmarAe-wy1hIQQB6ZWt0DMwQdKQCLcB/s640/image00.jpg" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Ravi Kumar, presenting the original paper in Amsterdam at WWW 2000&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;Curiously, the original study was done back in 1999 on two Altavista crawls having 200 million pages and 1.5 billion links. Today, Google indexes over 100 billion links merely within apps, and overall processes over 130 trillion web addresses in its web crawls.&lt;br /&gt;&lt;br /&gt;Over the years, the power law was found to be characteristic of many other Web-related phenomena, including the structure of social networks and the distribution of search query frequencies. The description of the macroscopic structure of the Web graph proposed by Broder et al. provided a solid mathematical foundation for numerous subsequent studies on crawling and searching the Web, which profoundly influenced the architecture of modern search engines.&lt;br /&gt;&lt;br /&gt;Hearty congratulations to all the authors on the well-deserved award! &lt;div class="feedflare"&gt;
&lt;a href="http://feeds.feedburner.com/~ff/blogspot/gJZg?a=K2lQUZJmfW0:Q_aENU8A3Lo:yIl2AUoC8zA"&gt;&lt;img src="http://feeds.feedburner.com/~ff/blogspot/gJZg?d=yIl2AUoC8zA" border="0"&gt;&lt;/img&gt;&lt;/a&gt;
&lt;/div&gt;&lt;img src="http://feeds.feedburner.com/~r/blogspot/gJZg/~4/K2lQUZJmfW0" height="1" width="1" alt=""/&gt;</content><link rel="replies" type="application/atom+xml" href="http://research.googleblog.com/feeds/3534392145444604799/comments/default" title="Post Comments" /><link rel="replies" type="text/html" href="http://research.googleblog.com/2017/04/and-award-goes-to.html#comment-form" title="0 Comments" /><link rel="edit" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/3534392145444604799" /><link rel="self" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/3534392145444604799" /><link rel="alternate" type="text/html" href="http://feedproxy.google.com/~r/blogspot/gJZg/~3/K2lQUZJmfW0/and-award-goes-to.html" title="And the award goes to..." /><author><name>Research Blog</name><uri>https://plus.google.com/101673966767287570260</uri><email>noreply@blogger.com</email><gd:image rel="http://schemas.google.com/g/2005#thumbnail" width="32" height="32" src="//lh5.googleusercontent.com/-GXTJ2sdBxaM/AAAAAAAAAAI/AAAAAAAAA14/2q3ThT_9JIY/s512-c/photo.jpg" /></author><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://1.bp.blogspot.com/-1xICW7QwoEw/WOPWgOmrNEI/AAAAAAAABrU/_jI6HGrbEPQi2AOi4aoX6HGsJSk-XzFDACLcB/s72-c/image01.png" height="72" width="72" /><thr:total>0</thr:total><gd:extendedProperty name="commentSource" value="1" /><gd:extendedProperty name="commentModerationMode" value="FILTERED_POSTMOD" /><feedburner:origLink>http://research.googleblog.com/2017/04/and-award-goes-to.html</feedburner:origLink></entry><entry><id>tag:blogger.com,1999:blog-21224994.post-749275968056489402</id><published>2017-04-03T10:00:00.000-07:00</published><updated>2017-04-03T10:00:23.496-07:00</updated><category scheme="http://www.blogger.com/atom/ns#" term="Algorithms" /><category scheme="http://www.blogger.com/atom/ns#" term="Computer Science" /><category scheme="http://www.blogger.com/atom/ns#" term="Google Cloud Platform" /><category scheme="http://www.blogger.com/atom/ns#" term="optimization" /><title type="text">Consistent Hashing with Bounded Loads</title><content type="html">&lt;span class="byline-author"&gt;Posted by Vahab Mirrokni, Principal Scientist, Morteza Zadimoghaddam, Research Scientist, NYC Algorithms Team&lt;/span&gt;&lt;br /&gt;&lt;br /&gt;Running a large-scale web service, such as content hosting, necessarily requires &lt;a href="https://en.wikipedia.org/wiki/Load_balancing_(computing)"&gt;load balancing&lt;/a&gt; — distributing clients &lt;i&gt;uniformly&lt;/i&gt; across multiple servers such that none get overloaded. Further, it is desirable to find an allocation that does not change very much over time in a &lt;i&gt;dynamic&lt;/i&gt; environment in which both clients and servers can be added or removed at any time. In other words, we need the allocation of clients to servers to be &lt;i&gt;consistent&lt;/i&gt; over time.&lt;br /&gt;&lt;br /&gt;In collaboration with &lt;a href="http://www.diku.dk/~mthorup/"&gt;Mikkel Thorup&lt;/a&gt;, a visiting researcher from university of Copenhagen, we developed a new efficient allocation algorithm for this problem with &lt;i&gt;tight guarantees&lt;/i&gt; on the maximum load of each server, and studied it theoretically and empirically. We then worked with our Cloud team to implement it in &lt;a href="https://cloud.google.com/pubsub/"&gt;Google Cloud Pub/Sub&lt;/a&gt;, a scalable event streaming service, and observed substantial improvement on uniformity of the load allocation (in terms of the maximum load assigned to servers) while maintaining consistency and stability objectives. In August 2016 we described our algorithm in the paper “&lt;a href="https://arxiv.org/abs/1608.01350"&gt;Consistent Hashing with Bounded Loads&lt;/a&gt;”, and shared it on ArXiv for potential use by the broader research community. &lt;br /&gt;&lt;br /&gt;Three months later, Andrew Rodland from &lt;a href="https://vimeo.com/"&gt;Vimeo&lt;/a&gt; informed us that he had found the paper, implemented it in &lt;a href="https://github.com/arodland/haproxy/commit/b02bed24daf64743cb9a571e93ed29ee4bc7efe7"&gt;haproxy&lt;/a&gt; (a widely-used piece of open source software), and used it for their load balancing project at Vimeo. The results were dramatic: applying these algorithmic ideas helped them decrease the cache bandwidth by a factor of almost 8, eliminating a scaling bottleneck. He recently summarized this story in a &lt;a href="https://medium.com/vimeo-engineering-blog/improving-load-balancing-with-a-new-consistent-hashing-algorithm-9f1bd75709ed"&gt;blog post&lt;/a&gt; detailing his use case. Needless to say, we were excited to learn that our theoretical research was not only put into application, but also that it was useful &lt;i&gt;and&lt;/i&gt; open-sourced. &lt;br /&gt;&lt;br /&gt;&lt;b&gt;Background&lt;/b&gt;&lt;br /&gt;While the concept of &lt;a href="https://en.wikipedia.org/wiki/Consistent_hashing"&gt;consistent hashing&lt;/a&gt; has been developed in the past to deal with load balancing in dynamic environments, a fundamental issue with all the previously developed schemes is that, in certain scenarios, they may result in sub-optimal load balancing on many servers. &lt;br /&gt;&lt;br /&gt;Additionally, both clients and servers may be added or removed periodically, and with such changes, we do not want to move too many clients. Thus, while the dynamic allocation algorithm has to always ensure a proper load balancing, it should also aim to minimize the number of clients moved after each change to the system. Such allocation problems become even more challenging when we face hard constraints on the capacity of each server - that is, each server has a capacity that the load may not exceed. Typically, we want capacities close to the average loads. &lt;br /&gt;&lt;br /&gt;In other words, we want to simultaneously achieve both &lt;i&gt;uniformity&lt;/i&gt; and &lt;i&gt;consistency&lt;/i&gt; in the resulting allocations. There is a vast amount of literature on solutions in the much simpler case where the set of servers is fixed and only the client set is updated, but in this post we discuss solutions that are relevant in the fully &lt;i&gt;dynamic&lt;/i&gt; case where both clients and servers can be added and removed. &lt;br /&gt;&lt;br /&gt;&lt;b&gt;The Algorithm&lt;/b&gt;&lt;br /&gt;We can think about the servers as bins and clients as balls to have a similar notation with well-studied &lt;a href="https://en.wikipedia.org/wiki/Balls_into_bins"&gt;balls-to-bins stochastic processes&lt;/a&gt;. The uniformity objective encourages all bins to have a load roughly equal to the average density (the number of balls divided by the number of bins). For some parameter ε, we set the capacity of each bin to either &lt;a href="https://en.wikipedia.org/wiki/Floor_and_ceiling_functions"&gt;floor or ceiling&lt;/a&gt; of the average load times (1+ε). This extra capacity allows us to design an allocation algorithm that meets the consistency objective in addition to the uniformity property. &lt;br /&gt;&lt;br /&gt;Imagine a given range of numbers overlaid on a circle. We apply a hash function to balls and a separate hash function to bins to obtain numbers in that range that correspond to positions on that circle. We then start allocating balls in a specific order independent of their hash values (let’s say based on their ID). Then each ball is moved clockwise and is assigned to the first bin with spare capacity. &lt;br /&gt;&lt;div class="separator" style="clear: both; text-align: center;"&gt;&lt;a href="https://3.bp.blogspot.com/-pgZ4b9H7VlM/WOJ91rDe_XI/AAAAAAAABqw/wIjtyPHheFgyHpXIqY4qNLhd_H9DnHsXACLcB/s1600/image00.png" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"&gt;&lt;img border="0" height="588" src="https://3.bp.blogspot.com/-pgZ4b9H7VlM/WOJ91rDe_XI/AAAAAAAABqw/wIjtyPHheFgyHpXIqY4qNLhd_H9DnHsXACLcB/s640/image00.png" width="640" /&gt;&lt;/a&gt;&lt;/div&gt;Consider the example above where 6 balls and 3 bins are assigned using two separate hash functions to random locations on the circle. For the sake of this instance, assume the capacity of each bin is set to 2. We start allocating balls in the increasing order of their ID values. Ball number 1 moves clockwise, and goes to bin C. Ball number 2 goes to A. Balls 3 and 4 go to bin B. Ball number 5 goes to bin C. Then ball number 6 moves clockwise and hits bin B first. However bin B has capacity 2 and already contains balls 3 and 4. So ball 6 keeps moving to reach bin C but that bin is also full. Finally, ball 6 ends up in bin A that has a spare slot for it.&lt;br /&gt;&lt;br /&gt;Upon any update in the system (ball or bin insertion/deletion), the allocation is recomputed to keep the uniformity objective. The art of the analysis is to show that a small update (a few number of insertions and deletions) results in minor changes in the state of the allocation and therefore the consistency objective is met. In &lt;a href="https://arxiv.org/abs/1608.01350"&gt;our paper&lt;/a&gt; we show that every ball removal or insertion in the system results in O(1/ε&lt;sup&gt;2&lt;/sup&gt;) movements of other balls. The most important thing about this upper bound is that it is independent of the total number of balls or bins in the system. So if the number of balls or bins are doubled, this bound will not change. Having an upper bound independent of the number of balls or bins introduces room for scalability as the consistency objective is not violated if we move to bigger instances. Simulations for the number of movements (relocations) per update is shown below when an update occurs on a bin/server. &lt;br /&gt;&lt;div class="separator" style="clear: both; text-align: center;"&gt;&lt;a href="https://3.bp.blogspot.com/-IJFOHvSomXY/WOJ-AecmDaI/AAAAAAAABq0/wVAwJd8jxNs7cT30aU0ek3_WpPzYYSO9ACLcB/s1600/image01.png" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"&gt;&lt;img border="0" height="505" src="https://3.bp.blogspot.com/-IJFOHvSomXY/WOJ-AecmDaI/AAAAAAAABq0/wVAwJd8jxNs7cT30aU0ek3_WpPzYYSO9ACLcB/s640/image01.png" width="640" /&gt;&lt;/a&gt;&lt;/div&gt;The red curve shows the average number of movements and the blue bars indicate the variance for different values of ε (the x-axis). The dashed curve is the upper bound suggested by our theoretical results which fits nicely as a prediction of the actual number of movements. Furthermore, for any value of ε, we know the load of each bin is at most (1+ε) times the average load. Below we see the load distribution of bins for different values of ε=0.1, ε=0.3 and ε=0.9.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://1.bp.blogspot.com/-PUdBAM5bDKk/WOJ-Hm3HAgI/AAAAAAAABq4/iREEzcJjdIQ7YYYE6bfGIEsbQALIozEKgCLcB/s1600/image02.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="532" src="https://1.bp.blogspot.com/-PUdBAM5bDKk/WOJ-Hm3HAgI/AAAAAAAABq4/iREEzcJjdIQ7YYYE6bfGIEsbQALIozEKgCLcB/s640/image02.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;The distribution of loads for several values of ε. The load distribution is nearly uniform covering all ranges of loads from 0 to (1+ε) times average, and many bins with load equal to (1+ε) times average.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;As one can see there is a tradeoff — a lower ε helps with uniformity but not with consistency, while larger ε values help with consistency. A lower ε will ensure that many loads will be equal to the hard capacity limit of (1+ε) times the average, and the rest have a decaying distribution.&lt;br /&gt;&lt;br /&gt;When providing content hosting services, one must be ready to face a variety of instances with different characteristics. This consistent hashing scheme is ideal for such scenarios as it performs well even for worst-case instances. &lt;br /&gt;&lt;br /&gt;While our internal results are exciting, we are even more pleased that the broader community found our solution useful enough to &lt;a href="https://github.com/arodland/haproxy"&gt;open-source&lt;/a&gt;, allowing anyone to use this algorithm. If you are interested in further details of this research, please see the &lt;a href="https://arxiv.org/abs/1608.01350"&gt;paper&lt;/a&gt; on ArXiv, and stay tuned for more research from the &lt;a href="https://research.google.com/teams/nycalg/"&gt;NYC Algorithms Team&lt;/a&gt;!&lt;br /&gt;&lt;br /&gt;&lt;b&gt;Acknowledgements:&lt;/b&gt;&lt;br /&gt;We would like to thank Alex Totok, Matt Gruskin, Sergey Kondratyev and Haakon Ringberg from the Google Cloud Pub/Sub team, and of course &lt;a href="http://www.diku.dk/~mthorup/"&gt;Mikkel Thorup&lt;/a&gt; for his invaluable contributions to this paper.&lt;div class="feedflare"&gt;
&lt;a href="http://feeds.feedburner.com/~ff/blogspot/gJZg?a=VGqk0FwT0XU:ACv5pTzKaJU:yIl2AUoC8zA"&gt;&lt;img src="http://feeds.feedburner.com/~ff/blogspot/gJZg?d=yIl2AUoC8zA" border="0"&gt;&lt;/img&gt;&lt;/a&gt;
&lt;/div&gt;&lt;img src="http://feeds.feedburner.com/~r/blogspot/gJZg/~4/VGqk0FwT0XU" height="1" width="1" alt=""/&gt;</content><link rel="replies" type="application/atom+xml" href="http://research.googleblog.com/feeds/749275968056489402/comments/default" title="Post Comments" /><link rel="replies" type="text/html" href="http://research.googleblog.com/2017/04/consistent-hashing-with-bounded-loads.html#comment-form" title="0 Comments" /><link rel="edit" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/749275968056489402" /><link rel="self" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/749275968056489402" /><link rel="alternate" type="text/html" href="http://feedproxy.google.com/~r/blogspot/gJZg/~3/VGqk0FwT0XU/consistent-hashing-with-bounded-loads.html" title="Consistent Hashing with Bounded Loads" /><author><name>Research Blog</name><uri>https://plus.google.com/101673966767287570260</uri><email>noreply@blogger.com</email><gd:image rel="http://schemas.google.com/g/2005#thumbnail" width="32" height="32" src="//lh5.googleusercontent.com/-GXTJ2sdBxaM/AAAAAAAAAAI/AAAAAAAAA14/2q3ThT_9JIY/s512-c/photo.jpg" /></author><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://3.bp.blogspot.com/-pgZ4b9H7VlM/WOJ91rDe_XI/AAAAAAAABqw/wIjtyPHheFgyHpXIqY4qNLhd_H9DnHsXACLcB/s72-c/image00.png" height="72" width="72" /><thr:total>0</thr:total><gd:extendedProperty name="commentSource" value="1" /><gd:extendedProperty name="commentModerationMode" value="FILTERED_POSTMOD" /><feedburner:origLink>http://research.googleblog.com/2017/04/consistent-hashing-with-bounded-loads.html</feedburner:origLink></entry><entry><id>tag:blogger.com,1999:blog-21224994.post-1933076705168904901</id><published>2017-03-30T10:00:00.000-07:00</published><updated>2017-03-30T10:12:42.455-07:00</updated><category scheme="http://www.blogger.com/atom/ns#" term="datasets" /><category scheme="http://www.blogger.com/atom/ns#" term="Machine Perception" /><category scheme="http://www.blogger.com/atom/ns#" term="Research" /><category scheme="http://www.blogger.com/atom/ns#" term="YouTube" /><title type="text">Announcing AudioSet: A Dataset for Audio Event Research</title><content type="html">&lt;span class="byline-author"&gt;Posted by Dan Ellis, Research Scientist, Sound Understanding Team&lt;/span&gt;&lt;br /&gt;&lt;br /&gt;Systems able to recognize sounds familiar to human listeners have a wide range of applications, from &lt;a href="https://research.googleblog.com/2017/03/adding-sound-effect-information-to.html"&gt;adding sound effect information to automatic video captions,&lt;/a&gt; to potentially allowing you to search videos for specific audio events.  Building Deep Learning systems to do this relies heavily on both a large quantity of computing (often from highly parallel GPUs), and also – and perhaps more importantly – on significant amounts of accurately-labeled training data. However, research in environmental sound recognition is limited by currently available public datasets.&lt;br /&gt;&lt;br /&gt;In order to address this, we recently released &lt;a href="http://g.co/audioset"&gt;AudioSet&lt;/a&gt;, a collection of over 2 million ten-second YouTube excerpts labeled with a vocabulary of 527 sound event categories, with at least 100 examples for each category. Announced in &lt;a href="https://research.google.com/pubs/pub45857.html"&gt;our paper&lt;/a&gt; at the &lt;a href="http://www.ieee-icassp2017.org/"&gt;IEEE International Conference on Acoustics, Speech, and Signal Processing&lt;/a&gt;, AudioSet provides a common, realistic-scale evaluation task for audio event detection and a starting point for a comprehensive vocabulary of sound events, designed to advance research into audio event detection and recognition.  &lt;br /&gt;&lt;div class="separator" style="clear: both; text-align: center;"&gt;&lt;a href="https://research.google.com/audioset/" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"&gt;&lt;img border="0" height="176" src="https://1.bp.blogspot.com/-V1eBvymMYK8/WN00mlaVHoI/AAAAAAAABqM/sNvS811qyBA2YHg2c0_W4Qfa3WkOHsW_ACLcB/s640/image01.png" width="640" /&gt;&lt;/a&gt;&lt;/div&gt;&lt;b&gt;Developing an Ontology&lt;/b&gt;&lt;br /&gt;When we started on this work last year, our first task was to define a vocabulary of sound classes that provided a consistent level of detail over the spectrum of sound events we planned to label.  Defining this ontology was necessary to avoid problems of ambiguity and synonyms; without this, we might end up trying to differentiate “Timpani” from “Kettle drum”, or “Water tap” from “Faucet”. Although a number of scientists have looked at &lt;a href="https://openi.nlm.nih.gov/detailedresult.php?img=PMC3871505_CIN2013-586138.001&amp;amp;req=4"&gt;how humans organize sound events&lt;/a&gt;, the few existing ontologies proposed have been small and partial.  To build our own, we searched the web for phrases like “Sounds, such as X and Y”, or “X, Y, and other sounds”.  This gave us a list of sound-related words which we manually sorted into a hierarchy of over 600 sound event classes ranging from “&lt;a href="https://research.google.com/audioset/ontology/child_speech_kid_speaking.html"&gt;Child speech&lt;/a&gt;” to “&lt;a href="https://research.google.com/audioset/ontology/ukulele.html"&gt;Ukulele&lt;/a&gt;” to “&lt;a href="https://research.google.com/audioset/ontology/boing.html"&gt;Boing&lt;/a&gt;”. To make our taxonomy as comprehensive as possible, we then looked at comparable lists of sound events  (for instance, the &lt;a href="https://serv.cusp.nyu.edu/projects/urbansounddataset/taxonomy.html"&gt;Urban Sound Taxonomy&lt;/a&gt;) to add significant classes we may have missed and to merge classes that weren't well defined or well distinguished. You can explore our ontology &lt;a href="https://research.google.com/audioset/ontology/index.html"&gt;here&lt;/a&gt;.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://4.bp.blogspot.com/-aFpmH4QCAvY/WN00tJ_O6-I/AAAAAAAABqQ/foXm-2kBjpkbc55-ui6yx_-N6EgYuq86ACLcB/s1600/image03.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="572" src="https://4.bp.blogspot.com/-aFpmH4QCAvY/WN00tJ_O6-I/AAAAAAAABqQ/foXm-2kBjpkbc55-ui6yx_-N6EgYuq86ACLcB/s640/image03.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;The top two levels of the AudioSet &lt;a href="https://research.google.com/audioset/ontology/index.html"&gt;ontology&lt;/a&gt;.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;&lt;b&gt;From Ontology to Labeled Data&lt;/b&gt;&lt;br /&gt;With our new ontology in hand, we were able to begin collecting human judgments of where the sound events occur. This, too, raises subtle problems: unlike the billions of well-composed photographs available online, people don’t typically produce “well-framed” sound recordings, much less provide them with captions. We decided to use 10 second sound snippets as our unit; anything shorter becomes very difficult to identify in isolation. We collected candidate snippets for each of our classes by taking random excerpts from YouTube videos whose metadata indicated they might contain the sound in question (“Dogs Barking for 10 Hours”).  Each snippet was presented to a human labeler with a small set of category names to be confirmed (“Do you hear a Bark?”).  Subsequently, we proposed snippets whose content was similar to examples that had already been manually verified to contain the class, thereby finding examples that were not discoverable from the metadata. Because some classes were much harder to find than others – particularly the &lt;a href="https://en.wikipedia.org/wiki/Onomatopoeia"&gt;onomatopoeia words&lt;/a&gt; like “Squish” and “Clink” – we adapted our segment proposal process to increase the sampling for those categories. For more details, see &lt;a href="https://research.google.com/pubs/pub45760.html"&gt;our paper&lt;/a&gt; on the matching technique. &lt;br /&gt;&lt;br /&gt;AudioSet provides the URLs of each video excerpt along with the sound classes that the raters confirmed as present, as well as precalculated audio features from the same classifier used to generate audio features for the &lt;a href="https://research.googleblog.com/2017/02/an-updated-youtube-8m-video.html"&gt;updated YouTube 8M Dataset&lt;/a&gt;. Below is a histogram of the number of examples per class:&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://1.bp.blogspot.com/-OZv7NYeee3M/WN00_ZBnRUI/AAAAAAAABqU/yCIVntSaHJYpFWIIJjDkaDcSokGKB_DBwCLcB/s1600/image02.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="602" src="https://1.bp.blogspot.com/-OZv7NYeee3M/WN00_ZBnRUI/AAAAAAAABqU/yCIVntSaHJYpFWIIJjDkaDcSokGKB_DBwCLcB/s640/image02.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;The total number of videos for selected classes in AudioSet.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;You can browse this data at the &lt;a href="https://g.co/audioset"&gt;AudioSet website&lt;/a&gt; which lets you view all the 2 million excerpts for all the classes.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://research.google.com/audioset/eval/violin_fiddle.html" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="320" src="https://3.bp.blogspot.com/-qxP4i1apvog/WN01Wx-L4-I/AAAAAAAABqY/v1wM6H7Nq5Muj2p0-GVUoZqKmCi3x-HUgCLcB/s640/image00.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;A few of the segments representing the class “&lt;a href="https://research.google.com/audioset/eval/violin_fiddle.html"&gt;Violin, fiddle&lt;/a&gt;”. &lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;&lt;b&gt;Quality Assessment&lt;/b&gt;&lt;br /&gt;Once we had a substantial set of human ratings, we conducted an internal Quality Assessment task where, for most of the classes, we checked 10 examples of excerpts that the annotators had labeled with that class. This revealed a significant number of classes with inaccurate labeling: some, like “Dribble” (uneven water flow) and “Roll” (a hard object moving by rotating) had been systematically confused (as basketball dribbling and drum rolls, respectively); some such as “Patter” (footsteps of small animals) and “Sidetone” (background sound on a telephony channel) were too difficult to label and/or find examples for, even with our content-based matching. We also looked at the behavior of a classifier trained on the entire dataset and found a number of frequent confusions indicating separate classes that were not really distinct, such as “Jingle” and “Tinkle”.&lt;br /&gt;&lt;br /&gt;To address these “problem” classes, we developed a re-rating process by labeling groups of classes that span (and thus highlight) common confusions, and created instructions for the labeler to be used with these sounds. This re-rating has led to multiple improvements – merged classes, expanded coverage, better descriptions – that we were able to incorporate in this release. This iterative process of labeling and assessment has been particularly effective in shaking out weaknesses in the ontology.&lt;br /&gt;&lt;br /&gt;&lt;b&gt;A Community Dataset&lt;/b&gt;&lt;br /&gt;By releasing AudioSet, we hope to provide a common, realistic-scale evaluation task for audio event detection, as well as a starting point for a comprehensive vocabulary of sound events. We would like to see a vibrant sound event research community develop, including through external efforts such as the &lt;a href="http://www.cs.tut.fi/sgn/arg/dcase2017/"&gt;DCASE challenge series&lt;/a&gt;. We will continue to improve the size, coverage, and accuracy of this data and plan to make a second release in the coming months when our rerating process is complete. We additionally encourage the research community to continue to refine our ontology, which we have open sourced on &lt;a href="https://github.com/audioset/ontology"&gt;GitHub&lt;/a&gt;.  We believe that with this common focus, sound event recognition will continue to advance and will allow machines to understand sounds similar to the way we do, enabling new and exciting applications.  &lt;br /&gt;&lt;br /&gt;&lt;b&gt;Acknowledgments:&lt;/b&gt;&lt;br /&gt;AudioSet is the work of Jort F. Gemmeke, Dan Ellis, Dylan Freedman, Shawn Hershey, Aren Jansen, Wade Lawrence, Channing Moore, Manoj Plakal, and Marvin Ritter, with contributions from Sami Abu-El-Hajia, Sourish Chaudhuri, Victor Gomes, Nisarg Kothari, Dick Lyon, Sobhan Naderi Parizi, Paul Natsev, Brian Patton, Rif A. Saurous, Malcolm Slaney, Ron Weiss, and Kevin Wilson.&lt;div class="feedflare"&gt;
&lt;a href="http://feeds.feedburner.com/~ff/blogspot/gJZg?a=mApb03g68ZM:na5wECJgIo8:yIl2AUoC8zA"&gt;&lt;img src="http://feeds.feedburner.com/~ff/blogspot/gJZg?d=yIl2AUoC8zA" border="0"&gt;&lt;/img&gt;&lt;/a&gt;
&lt;/div&gt;&lt;img src="http://feeds.feedburner.com/~r/blogspot/gJZg/~4/mApb03g68ZM" height="1" width="1" alt=""/&gt;</content><link rel="replies" type="application/atom+xml" href="http://research.googleblog.com/feeds/1933076705168904901/comments/default" title="Post Comments" /><link rel="replies" type="text/html" href="http://research.googleblog.com/2017/03/announcing-audioset-dataset-for-audio.html#comment-form" title="0 Comments" /><link rel="edit" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/1933076705168904901" /><link rel="self" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/1933076705168904901" /><link rel="alternate" type="text/html" href="http://feedproxy.google.com/~r/blogspot/gJZg/~3/mApb03g68ZM/announcing-audioset-dataset-for-audio.html" title="Announcing AudioSet: A Dataset for Audio Event Research" /><author><name>Research Blog</name><uri>https://plus.google.com/101673966767287570260</uri><email>noreply@blogger.com</email><gd:image rel="http://schemas.google.com/g/2005#thumbnail" width="32" height="32" src="//lh5.googleusercontent.com/-GXTJ2sdBxaM/AAAAAAAAAAI/AAAAAAAAA14/2q3ThT_9JIY/s512-c/photo.jpg" /></author><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://1.bp.blogspot.com/-V1eBvymMYK8/WN00mlaVHoI/AAAAAAAABqM/sNvS811qyBA2YHg2c0_W4Qfa3WkOHsW_ACLcB/s72-c/image01.png" height="72" width="72" /><thr:total>0</thr:total><gd:extendedProperty name="commentSource" value="1" /><gd:extendedProperty name="commentModerationMode" value="FILTERED_POSTMOD" /><feedburner:origLink>http://research.googleblog.com/2017/03/announcing-audioset-dataset-for-audio.html</feedburner:origLink></entry><entry><id>tag:blogger.com,1999:blog-21224994.post-9017009920101052463</id><published>2017-03-23T10:00:00.000-07:00</published><updated>2017-03-23T10:03:08.836-07:00</updated><category scheme="http://www.blogger.com/atom/ns#" term="accessibility" /><category scheme="http://www.blogger.com/atom/ns#" term="Machine Hearing" /><category scheme="http://www.blogger.com/atom/ns#" term="Machine Learning" /><category scheme="http://www.blogger.com/atom/ns#" term="YouTube" /><title type="text">Adding Sound Effect Information to YouTube Captions</title><content type="html">&lt;span class="byline-author"&gt;Posted by Sourish Chaudhuri, Software Engineer, Sound Understanding&lt;/span&gt;&lt;br /&gt;&lt;br /&gt;The effect of audio on our perception of the world can hardly be overstated. Its importance as a communication medium via speech is obviously the most familiar, but there is also significant information conveyed by ambient sounds. These ambient sounds create context that we instinctively respond to, like getting startled by sudden commotion, the use of music as a narrative element, or how laughter is used as an audience cue in sitcoms. &lt;br /&gt;&lt;br /&gt;Since 2009, &lt;a href="https://googleblog.blogspot.com/2009/11/automatic-captions-in-youtube.html"&gt;YouTube has provided automatic caption tracks&lt;/a&gt; for videos, focusing heavily on speech transcription in order to make the content hosted more accessible. However, without similar descriptions of the ambient sounds in videos, much of the information and impact of a video is not captured by speech transcription alone. To address this, &lt;a href="http://youtube-eng.googleblog.com/2017/03/visualizing-sound-effects.html"&gt;we announced&lt;/a&gt;&amp;nbsp;the addition of sound effect information to the automatic caption track in YouTube videos, enabling greater access to the richness of all the audio content.&lt;br /&gt;&lt;br /&gt;In this post, we discuss the backend system developed for this effort, a collaboration among the Accessibility, Sound Understanding and YouTube teams that used machine learning (ML) to enable the first ever automatic sound effect captioning system for YouTube.&lt;br /&gt;&lt;div class="separator" style="clear: both; text-align: center;"&gt;&lt;iframe allowfullscreen="" class="YOUTUBE-iframe-video" data-thumbnail-src="https://i.ytimg.com/vi/QGiK8DAZ9BA/0.jpg" frameborder="0" height="360" src="https://www.youtube.com/embed/QGiK8DAZ9BA?rel=0&amp;amp;start=15&amp;amp;end=55;feature=player_embedded" width="640"&gt;&lt;/iframe&gt;&lt;/div&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Click the CC button to see the sound effect captioning system in action.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;The application of ML – in this case, a &lt;a href="https://en.wikipedia.org/wiki/Deep_learning"&gt;Deep Neural Network&lt;/a&gt; (DNN) model – to the captioning task presented unique challenges. While the process of analyzing the time-domain audio signal of a video to detect various ambient sounds is similar to other well known classification problems (such as object detection in images), in a product setting the solution faces additional difficulties. In particular, given an arbitrary segment of audio, we need our models to be able to 1) detect the desired sounds, 2) temporally localize the sound in the segment and 3) effectively integrate it in the caption track, which may have parallel and independent speech recognition results.&lt;br /&gt;&lt;br /&gt;&lt;b&gt;A DNN Model for Ambient Sound&lt;/b&gt;&lt;br /&gt;The first challenge we faced in developing the model was the task of obtaining enough labeled data suitable for training our neural network. While labeled ambient sound information is difficult to come by, we were able to generate a large enough dataset for training using weakly labeled data. But of all the ambient sounds in a given video, which ones should we train our DNN to detect? &lt;br /&gt;&lt;br /&gt;For the initial launch of this feature, we chose [APPLAUSE], [MUSIC] and [LAUGHTER], prioritized based upon our analysis of human-created caption tracks that indicates that they are among the most frequent sounds that are manually captioned. While the sound space is obviously far richer and provides even more contextually relevant information than these three classes, the semantic information conveyed by these sound effects in the caption track is relatively unambiguous, as opposed to sounds like [RING] which raises the question of “what was it that rang – a bell, an alarm, a phone?”&lt;br /&gt;&lt;br /&gt;Much of our initial work on detecting these ambient sounds also included developing the infrastructure and analysis frameworks to enable scaling for future work, including both the detection of sound events and their integration into the automatic caption track. Investing in the development of this infrastructure has the added benefit of allowing us to easily incorporate more sound types in the future, as we expand our algorithms to understand a wider vocabulary of sounds (e.g. [RING], [KNOCK], [BARK]).  In doing so, we will be able to incorporate the detected sounds into the narrative to provide more relevant information (e.g. [PIANO MUSIC], [RAUCOUS APPLAUSE]) to viewers. &lt;br /&gt;&lt;br /&gt;&lt;b&gt;Dense Detections to Captions&lt;/b&gt;&lt;br /&gt;When a video is uploaded to YouTube, the sound effect recognition pipeline runs on the audio stream in the video. The DNN looks at short segments of audio and predicts whether that segment contains any one of the sound events of interest – since multiple sound effects can co-occur, our model makes a prediction at each time step for each of the sound effects. The segment window is then slid to the right (i.e. a slightly later point in time) and the model is used to make a prediction again, and so on till it reaches the end. This results in a dense stream the (likelihood of) presence of each of the sound events in our vocabulary at 100 frames per second.&lt;br /&gt;&lt;br /&gt;The dense prediction stream is not directly exposed to the user, of course, since that would result in captions flickering on and off, and because we know that a number of sound effects have some degree of temporal continuity when they occur; e.g. “music” and “applause” will usually be present for a few seconds at least. To incorporate this intuition, we smooth over the dense prediction stream using a modified &lt;a href="https://en.wikipedia.org/wiki/Viterbi_algorithm"&gt;Viterbi algorithm&lt;/a&gt; containing two states: ON and OFF, with the predicted segments for each sound effect corresponding to the ON state. The figure below provides an illustration of the process in going from the dense detections to the final segments determined to contain sound effects of interest.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://1.bp.blogspot.com/-4FWWZ_W3e7E/WNLmPZKZpoI/AAAAAAAABpw/87z6wa1L4moo0JMDkaocqvr8OhHQsI9NwCLcB/s1600/image00.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="146" src="https://1.bp.blogspot.com/-4FWWZ_W3e7E/WNLmPZKZpoI/AAAAAAAABpw/87z6wa1L4moo0JMDkaocqvr8OhHQsI9NwCLcB/s640/image00.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;(Left) The dense sequence of probabilities from our DNN for the occurrence over time of single sound category in a video. (Center) Binarized segments based on the modified Viterbi algorithm. (Right) The duration-based filter removes segments that are shorter in duration than desired for the class.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;A classification-based system such as this one will certainly have some errors, and needs to be able to trade off false positives against missed detections as per the product goals. For example, due to the weak labels in the training dataset, the model was often confused between events that tended to co-occur. For example, a segment labeled “laugh” would usually contain both speech and laughter and the model for “laugh” would have a hard time distinguishing them in test data. In our system, we allow further restrictions based on time spent in the ON state (i.e. do not determine sound X to be detected unless it was determined to be present for at least Y seconds) to push performance toward a desired point in the precision-recall curve.&lt;br /&gt;&lt;br /&gt;Once we were satisfied with the performance of our system in temporally localizing sound effect captions based on our offline evaluation metrics, we were faced with the following: how do we combine the sound effect and speech captions to create a single automatic caption track, and how (or when) do we present sound effect information to the user to make it most useful to them?&lt;br /&gt;&lt;br /&gt;&lt;b&gt;Adding Sound Effect Information into the Automatic Captions Track&lt;/b&gt;&lt;br /&gt;Once we had a system capable of accurately detecting and classifying the ambient sounds in video, we investigated how to convey that information to the viewer in an effective way. In collaboration with our User Experience (UX) research teams, we explored various design options and tested them in a qualitative pilot usability study. The participants of the study had different hearing levels and varying needs for captions. We asked participants a number of questions including whether it improved their overall experience, their ability to follow events in the video and extract relevant information from the caption track, to understand the effect of variables such as:&lt;br /&gt;&lt;ul&gt;&lt;li&gt;Using separate parts of the screen for speech and sound effect captions.&lt;/li&gt;&lt;li&gt;Interleaving the speech and sound effect captions as they occur.&lt;/li&gt;&lt;li&gt;Only showing sound effect captions at the end of sentences or when there is a pause in speech (even if they occurred in the middle of speech).&lt;/li&gt;&lt;li&gt;How hearing users perceive captions when watching with the sound off.&lt;/li&gt;&lt;/ul&gt;While it wasn’t surprising that almost all users appreciated the added sound effect information when it was accurate, we also paid specific attention to the feedback when the sound detection system made an error (a false positive when determining presence of a sound, or failing to detect an occurrence). This presented a surprising result: when sound effect information was incorrect, it did not detract from the participant’s experience in roughly 50% of the cases. Based upon participant feedback, the reasons for this appear to be:&lt;br /&gt;&lt;ul&gt;&lt;li&gt;Participants who could hear the audio were able to ignore the inaccuracies.&lt;/li&gt;&lt;li&gt;Participants who could not hear the audio interpreted the error as the presence of a sound event, and that they had not missed out on critical speech information.&lt;/li&gt;&lt;/ul&gt;Overall, users reported that they would be fine with the system making the occasional mistake as long as it was able to provide good information far more often than not. &lt;br /&gt;&lt;br /&gt;&lt;b&gt;Looking Forward&lt;/b&gt;&lt;br /&gt;Our work toward enabling automatic sound effect captions for YouTube videos and the initial rollout is a step toward making the richness of content in videos more accessible to our users who experience videos in different ways and in different environments that require captions. We’ve developed a framework to enrich the automatic caption track with sound effects, but there is still much to be done here. We hope that this will spur further work and discussion in the community around improving captions using not only automatic techniques, but also around ways to make creator-generated and community-contributed caption tracks richer (including perhaps, starting with the auto-captions) and better to further improve the viewing experience for our users.&lt;br /&gt;&lt;br /&gt;&lt;br /&gt;&lt;br /&gt;&lt;div class="feedflare"&gt;
&lt;a href="http://feeds.feedburner.com/~ff/blogspot/gJZg?a=ol6uIA42PKE:WyPDeuddy7A:yIl2AUoC8zA"&gt;&lt;img src="http://feeds.feedburner.com/~ff/blogspot/gJZg?d=yIl2AUoC8zA" border="0"&gt;&lt;/img&gt;&lt;/a&gt;
&lt;/div&gt;&lt;img src="http://feeds.feedburner.com/~r/blogspot/gJZg/~4/ol6uIA42PKE" height="1" width="1" alt=""/&gt;</content><link rel="replies" type="application/atom+xml" href="http://research.googleblog.com/feeds/9017009920101052463/comments/default" title="Post Comments" /><link rel="replies" type="text/html" href="http://research.googleblog.com/2017/03/adding-sound-effect-information-to.html#comment-form" title="0 Comments" /><link rel="edit" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/9017009920101052463" /><link rel="self" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/9017009920101052463" /><link rel="alternate" type="text/html" href="http://feedproxy.google.com/~r/blogspot/gJZg/~3/ol6uIA42PKE/adding-sound-effect-information-to.html" title="Adding Sound Effect Information to YouTube Captions" /><author><name>Research Blog</name><uri>https://plus.google.com/101673966767287570260</uri><email>noreply@blogger.com</email><gd:image rel="http://schemas.google.com/g/2005#thumbnail" width="32" height="32" src="//lh5.googleusercontent.com/-GXTJ2sdBxaM/AAAAAAAAAAI/AAAAAAAAA14/2q3ThT_9JIY/s512-c/photo.jpg" /></author><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://img.youtube.com/vi/QGiK8DAZ9BA/default.jpg" height="72" width="72" /><thr:total>0</thr:total><gd:extendedProperty name="commentSource" value="1" /><gd:extendedProperty name="commentModerationMode" value="FILTERED_POSTMOD" /><feedburner:origLink>http://research.googleblog.com/2017/03/adding-sound-effect-information-to.html</feedburner:origLink></entry><entry><id>tag:blogger.com,1999:blog-21224994.post-5305047312580606390</id><published>2017-03-20T10:00:00.000-07:00</published><updated>2017-03-20T15:00:37.566-07:00</updated><category scheme="http://www.blogger.com/atom/ns#" term="Google Brain" /><category scheme="http://www.blogger.com/atom/ns#" term="Machine Learning" /><category scheme="http://www.blogger.com/atom/ns#" term="Publications" /><category scheme="http://www.blogger.com/atom/ns#" term="Visualization" /><title type="text">Distill: Supporting Clarity in Machine Learning</title><content type="html">&lt;span class="byline-author"&gt;Posted by Shan Carter, Software Engineer and Chris Olah, Research Scientist, Google Brain Team&lt;/span&gt;&lt;br /&gt;&lt;br /&gt;Science isn't just about discovering new results. It’s also about human understanding. Scientists need to develop notations, analogies, visualizations, and explanations of ideas. This human dimension of science isn't a minor side project. It's deeply tied to the heart of science.&lt;br /&gt;&lt;br /&gt;That’s why, in collaboration with OpenAI, DeepMind, YC Research, and others, we’re excited to announce the launch of &lt;a href="http://distill.pub/about/"&gt;Distill&lt;/a&gt;, a new open science journal and ecosystem supporting human understanding of machine learning. Distill is an independent organization, dedicated to fostering a new segment of the research community.&lt;br /&gt;&lt;br /&gt;Modern web technology gives us powerful new tools for expressing this human dimension of science. We can create interactive diagrams and user interfaces the enable intuitive exploration of research ideas. Over the last few years we've seen &lt;a href="http://explorableexplanations.com/"&gt;many&lt;/a&gt; &lt;a href="http://cs.stanford.edu/people/karpathy/convnetjs/"&gt;incredible&lt;/a&gt; &lt;a href="http://www.r2d3.us/visual-intro-to-machine-learning-part-1/"&gt;demonstrations&lt;/a&gt; &lt;a href="http://colah.github.io/"&gt;of&lt;/a&gt; &lt;a href="https://acko.net/tv/toolsforthought/"&gt;this&lt;/a&gt; &lt;a href="https://affinelayer.com/pixsrv/"&gt;kind&lt;/a&gt; &lt;a href="http://cognitivemedium.com/tat/"&gt;of&lt;/a&gt; &lt;a href="http://worrydream.com/MediaForThinkingTheUnthinkable/"&gt;work&lt;/a&gt;.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://1.bp.blogspot.com/-sPav5EPIk5c/WNAC9SOmDJI/AAAAAAAABpU/UVIwVWMwdkowWqDHwAVaALkPH8Y4T_YpwCLcB/s1600/image01.gif" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" src="https://1.bp.blogspot.com/-sPav5EPIk5c/WNAC9SOmDJI/AAAAAAAABpU/UVIwVWMwdkowWqDHwAVaALkPH8Y4T_YpwCLcB/s1600/image01.gif" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;An interactive diagram explaining the Neural Turing Machine from &lt;a href="http://distill.pub/2016/augmented-rnns/"&gt;Olah &amp;amp; Carter, 2016&lt;/a&gt;.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;Unfortunately, while there are a plethora of conferences and journals in machine learning, there aren’t any research venues that are dedicated to publishing this kind of work. This is partly an issue of focus, and partly because traditional publication venues can't, by virtue of their medium, support  interactive visualizations. Without a venue to publish in, many significant contributions don’t count as “real academic contributions” and their authors can’t access the academic support structure.&lt;br /&gt;&lt;br /&gt;That’s why Distill aims to build an ecosystem to support this kind of work, starting with three pieces: a research journal, prizes recognizing outstanding work, and tools to facilitate the creation of interactive articles.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="http://distill.pub/about/" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="272" src="https://3.bp.blogspot.com/-7hqB8FW_mg4/WNADSCOSqFI/AAAAAAAABpY/AfztQoO7GfMGYplEzP9rQzE6Rv2NmlDawCLcB/s640/image00.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;&lt;a href="http://distill.pub/about/"&gt;Distill is an ecosystem to support clarity in Machine Learning&lt;/a&gt;.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;Led by a diverse steering committee of leaders from the machine learning and user interface communities, we are very excited to see where Distill will go. To learn more about Distill, see the &lt;a href="http://distill.pub/about/"&gt;overview page&lt;/a&gt; or read the &lt;a href="http://distill.pub/"&gt;latest articles&lt;/a&gt;.&lt;div class="feedflare"&gt;
&lt;a href="http://feeds.feedburner.com/~ff/blogspot/gJZg?a=5pSZMAqlvZg:imIHRt8t_jw:yIl2AUoC8zA"&gt;&lt;img src="http://feeds.feedburner.com/~ff/blogspot/gJZg?d=yIl2AUoC8zA" border="0"&gt;&lt;/img&gt;&lt;/a&gt;
&lt;/div&gt;&lt;img src="http://feeds.feedburner.com/~r/blogspot/gJZg/~4/5pSZMAqlvZg" height="1" width="1" alt=""/&gt;</content><link rel="replies" type="application/atom+xml" href="http://research.googleblog.com/feeds/5305047312580606390/comments/default" title="Post Comments" /><link rel="replies" type="text/html" href="http://research.googleblog.com/2017/03/distill-supporting-clarity-in-machine.html#comment-form" title="0 Comments" /><link rel="edit" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/5305047312580606390" /><link rel="self" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/5305047312580606390" /><link rel="alternate" type="text/html" href="http://feedproxy.google.com/~r/blogspot/gJZg/~3/5pSZMAqlvZg/distill-supporting-clarity-in-machine.html" title="Distill: Supporting Clarity in Machine Learning" /><author><name>Research Blog</name><uri>https://plus.google.com/101673966767287570260</uri><email>noreply@blogger.com</email><gd:image rel="http://schemas.google.com/g/2005#thumbnail" width="32" height="32" src="//lh5.googleusercontent.com/-GXTJ2sdBxaM/AAAAAAAAAAI/AAAAAAAAA14/2q3ThT_9JIY/s512-c/photo.jpg" /></author><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://1.bp.blogspot.com/-sPav5EPIk5c/WNAC9SOmDJI/AAAAAAAABpU/UVIwVWMwdkowWqDHwAVaALkPH8Y4T_YpwCLcB/s72-c/image01.gif" height="72" width="72" /><thr:total>0</thr:total><gd:extendedProperty name="commentSource" value="1" /><gd:extendedProperty name="commentModerationMode" value="FILTERED_POSTMOD" /><feedburner:origLink>http://research.googleblog.com/2017/03/distill-supporting-clarity-in-machine.html</feedburner:origLink></entry><entry><id>tag:blogger.com,1999:blog-21224994.post-2264335905192158785</id><published>2017-03-16T10:00:00.000-07:00</published><updated>2017-03-17T13:59:08.939-07:00</updated><category scheme="http://www.blogger.com/atom/ns#" term="Image Processing" /><category scheme="http://www.blogger.com/atom/ns#" term="Publications" /><title type="text">Announcing Guetzli: A New Open Source JPEG Encoder</title><content type="html">&lt;span class="byline-author"&gt;Posted by Robert Obryk and Jyrki Alakuijala, Software Engineers, Google Research Europe&lt;/span&gt;&lt;br /&gt;&lt;br /&gt;&lt;i&gt;(Cross-posted on the &lt;a href="http://opensource.googleblog.com/2017/03/guetzli-new-open-source-jpeg-encoder.html"&gt;Google Open Source Blog&lt;/a&gt;)&lt;/i&gt;&lt;br /&gt;&lt;br /&gt;At Google, we care about giving users the best possible online experience, both through our own services and products and by contributing new tools and industry standards for use by the online community. That’s why we’re excited to announce &lt;a href="https://github.com/google/guetzli/"&gt;Guetzli&lt;/a&gt;, &lt;a href="http://arxiv.org/abs/1703.04421"&gt;a new open source algorithm&lt;/a&gt; that creates high quality JPEG images with file sizes 35% smaller than currently available methods, enabling webmasters to create webpages that can load faster and use even less data.&lt;br /&gt;&lt;br /&gt;Guetzli [guɛtsli] — &lt;i&gt;cookie&lt;/i&gt; in Swiss German — is a JPEG encoder for digital images and web graphics that can enable faster online experiences by producing smaller JPEG files while still maintaining compatibility with existing browsers, image processing applications and the JPEG standard. From the practical viewpoint this is very similar to our &lt;a href="https://github.com/google/zopfli"&gt;Zopfli&lt;/a&gt; algorithm, which produces smaller &lt;a href="https://en.wikipedia.org/wiki/Portable_Network_Graphics"&gt;PNG&lt;/a&gt; and &lt;a href="https://en.wikipedia.org/wiki/Gzip"&gt;gzip&lt;/a&gt; files without needing to introduce a new format, and different than the techniques used in &lt;a href="https://research.googleblog.com/2016/09/image-compression-with-neural-networks.html"&gt;RNN-based image compression&lt;/a&gt;, &lt;a href="https://blog.google/products/google-plus/saving-you-bandwidth-through-machine-learning/"&gt;RAISR&lt;/a&gt;, and &lt;a href="https://developers.google.com/speed/webp/"&gt;WebP&lt;/a&gt;, which all need client changes for compression gains at internet scale. &lt;br /&gt;&lt;br /&gt;The visual quality of JPEG images is directly correlated to its multi-stage compression process: &lt;a href="https://en.wikipedia.org/wiki/YUV"&gt;color space transform&lt;/a&gt;, &lt;a href="https://en.wikipedia.org/wiki/Discrete_cosine_transform"&gt;discrete cosine transform&lt;/a&gt;, and &lt;a href="https://en.wikipedia.org/wiki/Quantization_(image_processing)"&gt;quantization&lt;/a&gt;. Guetzli specifically targets the quantization stage in which the more visual quality loss is introduced, the smaller the resulting file. Guetzli strikes a balance between minimal loss and file size by employing a search algorithm that tries to overcome the difference between the &lt;a href="https://en.wikipedia.org/wiki/Human_visual_system_model"&gt;psychovisual&lt;/a&gt; modeling of JPEG's format, and &lt;a href="https://github.com/google/butteraugli"&gt;Guetzli’s psychovisual model&lt;/a&gt;, which approximates color perception and visual masking in a more thorough and detailed way than what is achievable by simpler color transforms and the discrete cosine transform. However, while Guetzli creates smaller image file sizes, the tradeoff is that these search algorithms take significantly longer to create compressed images than currently available methods.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://4.bp.blogspot.com/-hO5YEeBXctA/WMrBKcJti6I/AAAAAAAABo0/-Y6uN8TLQ_owiS4iFmfS4O_n1e69spYhQCLcB/s1600/image00.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="212" src="https://4.bp.blogspot.com/-hO5YEeBXctA/WMrBKcJti6I/AAAAAAAABo0/-Y6uN8TLQ_owiS4iFmfS4O_n1e69spYhQCLcB/s640/image00.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Figure 1. 16x16 pixel synthetic example of  a phone line  hanging against a blue sky — traditionally a case where JPEG compression algorithms suffer from artifacts. Uncompressed original is on the left. Guetzli (on the right) shows less ringing artefacts than libjpeg (middle) and has a smaller file size.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;And while Guetzli produces smaller image file sizes without sacrificing quality, we additionally found that in &lt;a href="http://arxiv.org/abs/1703.04416"&gt;experiments&lt;/a&gt; where compressed image file sizes are kept constant that human raters consistently preferred the images Guetzli produced over libjpeg images, even when the libjpeg files were the same size or even slightly larger. We think this makes the slower compression a worthy tradeoff.&lt;br /&gt;&lt;table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td style="text-align: center;"&gt;&lt;a href="https://3.bp.blogspot.com/-TMh-GxiSmLg/WMrBh8KGCVI/AAAAAAAABo4/LGVnXB_jMDs-1IhrJ38cM39dpNchAEvNgCLcB/s1600/image01.png" imageanchor="1" style="margin-left: auto; margin-right: auto;"&gt;&lt;img border="0" height="268" src="https://3.bp.blogspot.com/-TMh-GxiSmLg/WMrBh8KGCVI/AAAAAAAABo4/LGVnXB_jMDs-1IhrJ38cM39dpNchAEvNgCLcB/s640/image01.png" width="640" /&gt;&lt;/a&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td class="tr-caption" style="text-align: center;"&gt;Figure 2. 20x24 pixel zoomed areas from a picture of a cat’s eye. Uncompressed original on the left. Guetzli (on the right) &lt;br /&gt;shows less ringing artefacts than libjpeg (middle) without requiring a larger file size.&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt;It is our hope that webmasters and graphic designers will find Guetzli useful and apply it to their photographic content, making users’ experience smoother on image-heavy websites in addition to reducing load times and bandwidth costs for mobile users. Last, we hope that the new explicitly psychovisual approach in Guetzli will inspire further image and video compression research.&lt;div class="feedflare"&gt;
&lt;a href="http://feeds.feedburner.com/~ff/blogspot/gJZg?a=Vx9m61_8MGM:5_etv-seylI:yIl2AUoC8zA"&gt;&lt;img src="http://feeds.feedburner.com/~ff/blogspot/gJZg?d=yIl2AUoC8zA" border="0"&gt;&lt;/img&gt;&lt;/a&gt;
&lt;/div&gt;&lt;img src="http://feeds.feedburner.com/~r/blogspot/gJZg/~4/Vx9m61_8MGM" height="1" width="1" alt=""/&gt;</content><link rel="replies" type="application/atom+xml" href="http://research.googleblog.com/feeds/2264335905192158785/comments/default" title="Post Comments" /><link rel="replies" type="text/html" href="http://research.googleblog.com/2017/03/announcing-guetzli-new-open-source-jpeg.html#comment-form" title="0 Comments" /><link rel="edit" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/2264335905192158785" /><link rel="self" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/2264335905192158785" /><link rel="alternate" type="text/html" href="http://feedproxy.google.com/~r/blogspot/gJZg/~3/Vx9m61_8MGM/announcing-guetzli-new-open-source-jpeg.html" title="Announcing Guetzli: A New Open Source JPEG Encoder" /><author><name>Research Blog</name><uri>https://plus.google.com/101673966767287570260</uri><email>noreply@blogger.com</email><gd:image rel="http://schemas.google.com/g/2005#thumbnail" width="32" height="32" src="//lh5.googleusercontent.com/-GXTJ2sdBxaM/AAAAAAAAAAI/AAAAAAAAA14/2q3ThT_9JIY/s512-c/photo.jpg" /></author><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://4.bp.blogspot.com/-hO5YEeBXctA/WMrBKcJti6I/AAAAAAAABo0/-Y6uN8TLQ_owiS4iFmfS4O_n1e69spYhQCLcB/s72-c/image00.png" height="72" width="72" /><thr:total>0</thr:total><gd:extendedProperty name="commentSource" value="1" /><gd:extendedProperty name="commentModerationMode" value="FILTERED_POSTMOD" /><feedburner:origLink>http://research.googleblog.com/2017/03/announcing-guetzli-new-open-source-jpeg.html</feedburner:origLink></entry><entry><id>tag:blogger.com,1999:blog-21224994.post-8761197246122169764</id><published>2017-03-15T13:30:00.000-07:00</published><updated>2017-03-16T09:21:53.876-07:00</updated><category scheme="http://www.blogger.com/atom/ns#" term="Natural Language Processing" /><category scheme="http://www.blogger.com/atom/ns#" term="Natural Language Understanding" /><category scheme="http://www.blogger.com/atom/ns#" term="open source" /><category scheme="http://www.blogger.com/atom/ns#" term="TensorFlow" /><title type="text">An Upgrade to SyntaxNet, New Models and a Parsing Competition</title><content type="html">&lt;span class="byline-author"&gt;Posted by David Weiss and Slav Petrov, Research Scientists&lt;/span&gt;&lt;br /&gt;&lt;br /&gt;At Google, we continuously improve the language understanding capabilities used in applications ranging from &lt;a href="https://research.googleblog.com/2015/11/computer-respond-to-this-email.html"&gt;generation of email responses&lt;/a&gt; to &lt;a href="https://research.googleblog.com/2016/09/a-neural-network-for-machine.html"&gt;translation&lt;/a&gt;. Last summer, we open-sourced &lt;a href="https://research.googleblog.com/2016/05/announcing-syntaxnet-worlds-most.html"&gt;SyntaxNet&lt;/a&gt;, a neural-network framework for analyzing and understanding the grammatical structure of sentences. Included in our release was &lt;a href="https://github.com/tensorflow/models/tree/master/syntaxnet"&gt;Parsey McParseface&lt;/a&gt;, a state-of-the-art model that we had trained for analyzing English, followed quickly by a collection of pre-trained models for 40 additional languages, which we dubbed &lt;a href="https://research.googleblog.com/2016/08/meet-parseys-cousins-syntax-for-40.html"&gt;Parsey's Cousins&lt;/a&gt;. While we were excited to share our research and to provide these resources to the broader community, building machine learning systems that work well for languages other than English remains an ongoing challenge. We are excited to announce a few new research resources, available now, that address this problem.&lt;br /&gt;&lt;br /&gt;&lt;b&gt;SyntaxNet Upgrade&lt;/b&gt;&lt;br /&gt;We are releasing a &lt;a href="https://github.com/tensorflow/models/tree/master/syntaxnet"&gt;major upgrade to SyntaxNet&lt;/a&gt;. This upgrade incorporates nearly a year’s worth of our research on multilingual language understanding, and is available to anyone interested in building systems for processing and understanding text. At the core of the upgrade is a &lt;a href="https://arxiv.org/pdf/1703.04474.pdf"&gt;new technology&lt;/a&gt; that enables learning of richly layered representations of input sentences. More specifically, the upgrade extends &lt;a href="https://www.tensorflow.org/"&gt;TensorFlow&lt;/a&gt; to allow joint modeling of multiple levels of linguistic structure, and to allow neural-network architectures to be created dynamically during processing of a sentence or document.&lt;br /&gt;&lt;br /&gt;Our upgrade makes it, for example, easy to build &lt;a href="https://github.com/tensorflow/models/tree/master/syntaxnet/g3doc/conll2017/paper.pdf"&gt;character-based models&lt;/a&gt; that learn to compose individual characters into words (e.g. ‘c-a-t’ spells ‘cat’). By doing so, the models can learn that words can be related to each other because they share common parts (e.g. ‘cats’ is the plural of ‘cat’ and shares the same stem; ‘wildcat’ is a type of ‘cat’). Parsey and Parsey’s Cousins, on the other hand, operated over sequences of words. As a result, they were forced to memorize words seen during training and relied mostly on the context to determine the grammatical function of previously unseen words. &lt;br /&gt;&lt;br /&gt;As an example, consider the following (meaningless but grammatically correct) sentence: &lt;br /&gt;&lt;div class="separator" style="clear: both; text-align: center;"&gt;&lt;a href="https://4.bp.blogspot.com/-FSHWoHulOiI/WMmeLjdv72I/AAAAAAAABoU/F79X7nfXWi4qhXwMx22gMma8IetZvU3ugCLcB/s1600/image01.png" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"&gt;&lt;img border="0" height="280" src="https://4.bp.blogspot.com/-FSHWoHulOiI/WMmeLjdv72I/AAAAAAAABoU/F79X7nfXWi4qhXwMx22gMma8IetZvU3ugCLcB/s640/image01.png" width="640" /&gt;&lt;/a&gt;&lt;/div&gt;This sentence was &lt;a href="https://en.wikipedia.org/wiki/Gostak"&gt;originally coined by Andrew Ingraham&lt;/a&gt; who explained: “You do not know what this means; nor do I. But if we assume that it is English, we know that the &lt;i&gt;doshes&lt;/i&gt; are &lt;i&gt;distimmed&lt;/i&gt; by the &lt;i&gt;gostak&lt;/i&gt;. We know too that one &lt;i&gt;distimmer&lt;/i&gt; of &lt;i&gt;doshes&lt;/i&gt; is a &lt;i&gt;gostak&lt;/i&gt;." Systematic patterns in &lt;a href="https://en.wikipedia.org/wiki/Morphology_(linguistics)"&gt;morphology&lt;/a&gt; and &lt;a href="https://en.wikipedia.org/wiki/Syntax"&gt;syntax&lt;/a&gt; allow us to guess the grammatical function of words even when they are completely novel: we understand that ‘doshes’ is the plural of the noun ‘dosh’ (similar to the ‘cats’ example above) or that ‘distim’ is the third person singular of the verb distim. Based on this analysis we can then derive the overall structure of this sentence even though we have never seen the words before.&lt;br /&gt;&lt;br /&gt;&lt;b&gt;ParseySaurus&lt;/b&gt;&lt;br /&gt;To showcase the new capabilities provided by our upgrade to SyntaxNet, we are releasing a set of new pretrained models called &lt;a href="https://github.com/tensorflow/models/tree/master/syntaxnet/g3doc/conll2017/README.md"&gt;ParseySaurus&lt;/a&gt;. These models use the character-based input representation mentioned above and are thus much better at predicting the meaning of new words based both on their spelling and how they are used in context. The ParseySaurus models are far more accurate than &lt;a href="https://research.googleblog.com/2016/08/meet-parseys-cousins-syntax-for-40.html"&gt;Parsey’s Cousins&lt;/a&gt; (reducing errors by as much as 25%), particularly for &lt;a href="https://en.wikipedia.org/wiki/Morphology_(linguistics)"&gt;morphologically-rich&lt;/a&gt; languages like Russian, or &lt;a href="https://en.wikipedia.org/wiki/Agglutinative_language"&gt;agglutinative languages&lt;/a&gt; like Turkish and Hungarian. In those languages there can be dozens of forms for each word and many of these forms might never be observed during training - even in a very large corpus.&lt;br /&gt;&lt;br /&gt;Consider the following fictitious Russian &lt;a href="https://en.wikipedia.org/wiki/Glokaya_kuzdra"&gt;sentence&lt;/a&gt;, where again the stems are meaningless, but the suffixes define an unambiguous interpretation of the sentence structure:&lt;br /&gt;&lt;div class="separator" style="clear: both; text-align: center;"&gt;&lt;a href="https://1.bp.blogspot.com/-rniANK5ZeLU/WMmeSKsf-4I/AAAAAAAABoY/7tbaPTLjSHs10IX9uGeuErPgjgNUHIRkgCLcB/s1600/image00.png" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"&gt;&lt;img border="0" height="172" src="https://1.bp.blogspot.com/-rniANK5ZeLU/WMmeSKsf-4I/AAAAAAAABoY/7tbaPTLjSHs10IX9uGeuErPgjgNUHIRkgCLcB/s640/image00.png" width="640" /&gt;&lt;/a&gt;&lt;/div&gt;Even though our Russian ParseySaurus model has never seen these words, it can correctly analyze the sentence by inspecting the character sequences which constitute each word. In doing so, the system can determine many properties of the words (notice how many more morphological features there are here than in the English example). To see the sentence as ParseySaurus does, here is a visualization of how the model analyzes this sentence:&lt;br /&gt;&lt;div class="separator" style="clear: both; text-align: center;"&gt;&lt;a href="https://4.bp.blogspot.com/-FcL_KZ3GZKE/WMmeXcCj55I/AAAAAAAABoc/FnOVRkvEfqwhJ5maDWgA54yDvJPBwQhjQCLcB/s1600/image02.png" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"&gt;&lt;img border="0" src="https://4.bp.blogspot.com/-FcL_KZ3GZKE/WMmeXcCj55I/AAAAAAAABoc/FnOVRkvEfqwhJ5maDWgA54yDvJPBwQhjQCLcB/s1600/image02.png" /&gt;&lt;/a&gt;&lt;/div&gt;Each square represents one node in the neural network graph, and lines show the connections between them. The left-side “tail” of the graph shows the model consuming the input as one long string of characters. These are intermittently passed to the right side, where the rich web of connections shows the model composing words into phrases and producing a syntactic parse. Check out the full-size rendering &lt;a href="https://github.com/tensorflow/models/tree/master/syntaxnet/g3doc/full-trace-image.png"&gt;here&lt;/a&gt;.&lt;br /&gt;&lt;br /&gt;&lt;b&gt;A Competition&lt;/b&gt;&lt;br /&gt;You might be wondering whether character-based modeling are all we need or whether there are other techniques that might be important. SyntaxNet has lots more to offer, like &lt;a href="https://en.wikipedia.org/wiki/Beam_search"&gt;beam search&lt;/a&gt; and different training objectives, but there are of course also many other possibilities. To find out what works well in practice we are helping co-organize, together with Charles University and other colleagues, a &lt;a href="http://universaldependencies.org/conll17/"&gt;multilingual parsing competition&lt;/a&gt; at this year’s &lt;a href="http://www.conll.org/"&gt;Conference on Computational Natural Language Learning&lt;/a&gt; (CoNLL) with the goal of building syntactic parsing systems that work well in real-world settings and for 45 different languages. &lt;br /&gt;&lt;br /&gt;The competition is made possible by the &lt;a href="http://universaldependencies.org/"&gt;Universal Dependencies&lt;/a&gt; (UD) initiative, whose goal is to develop cross-linguistically consistent treebanks. Because machine learned models can only be as good as the data that they have access to, we have been contributing data to UD &lt;a href="https://static.googleusercontent.com/media/research.google.com/en//pubs/archive/41377.pdf"&gt;since 2013&lt;/a&gt;. For the competition, we partnered with UD and &lt;a href="https://www.dfki.de/lt/"&gt;DFKI&lt;/a&gt; to build a new multilingual evaluation set consisting of 1000 sentences that have been translated into 20+ different languages and annotated by linguists with parse trees. This evaluation set is the first of its kind (in the past, each language had its own independent evaluation set) and will enable more consistent cross-lingual comparisons. Because the sentences have the same meaning and have been annotated according to the same guidelines, we will be able to get closer to answering the question of which languages might be harder to parse.&lt;br /&gt;&lt;br /&gt;We hope that the upgraded SyntaxNet framework and our the pre-trained ParseySaurus models will inspire researchers to participate in the competition. We have additionally created a &lt;a href="https://github.com/tensorflow/models/blob/master/syntaxnet/g3doc/CLOUD.md"&gt;tutorial&lt;/a&gt; showing how to load a &lt;a href="https://www.docker.com/"&gt;Docker&lt;/a&gt; image and train models on the &lt;a href="https://cloud.google.com/"&gt;Google Cloud Platform&lt;/a&gt;, to facilitate participation by smaller teams with limited resources. So, if you have an idea for making your own models with the SyntaxNet framework, &lt;a href="http://universaldependencies.org/conll17/registration.html"&gt;sign up to compete&lt;/a&gt;! We believe that the configurations that we are releasing are a good place to start, but we look forward to seeing how participants will be able to extend and improve these models or perhaps create better ones! &lt;br /&gt;&lt;br /&gt;Thanks to everyone involved who made this competition happen, including our collaborators at &lt;a href="https://ufal.mff.cuni.cz/udpipe"&gt;UD-Pipe&lt;/a&gt;, who provide another baseline implementation to make it easy to enter the competition. Happy parsing from the main developers, Chris Alberti, Daniel Andor, Ivan Bogatyy, Mark Omernick, Zora Tung and Ji Ma!&lt;div class="feedflare"&gt;
&lt;a href="http://feeds.feedburner.com/~ff/blogspot/gJZg?a=9xgmsNHAQak:O2F0Iq1BOn8:yIl2AUoC8zA"&gt;&lt;img src="http://feeds.feedburner.com/~ff/blogspot/gJZg?d=yIl2AUoC8zA" border="0"&gt;&lt;/img&gt;&lt;/a&gt;
&lt;/div&gt;&lt;img src="http://feeds.feedburner.com/~r/blogspot/gJZg/~4/9xgmsNHAQak" height="1" width="1" alt=""/&gt;</content><link rel="replies" type="application/atom+xml" href="http://research.googleblog.com/feeds/8761197246122169764/comments/default" title="Post Comments" /><link rel="replies" type="text/html" href="http://research.googleblog.com/2017/03/an-upgrade-to-syntaxnet-new-models-and.html#comment-form" title="0 Comments" /><link rel="edit" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/8761197246122169764" /><link rel="self" type="application/atom+xml" href="http://www.blogger.com/feeds/21224994/posts/default/8761197246122169764" /><link rel="alternate" type="text/html" href="http://feedproxy.google.com/~r/blogspot/gJZg/~3/9xgmsNHAQak/an-upgrade-to-syntaxnet-new-models-and.html" title="An Upgrade to SyntaxNet, New Models and a Parsing Competition" /><author><name>Research Blog</name><uri>https://plus.google.com/101673966767287570260</uri><email>noreply@blogger.com</email><gd:image rel="http://schemas.google.com/g/2005#thumbnail" width="32" height="32" src="//lh5.googleusercontent.com/-GXTJ2sdBxaM/AAAAAAAAAAI/AAAAAAAAA14/2q3ThT_9JIY/s512-c/photo.jpg" /></author><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://4.bp.blogspot.com/-FSHWoHulOiI/WMmeLjdv72I/AAAAAAAABoU/F79X7nfXWi4qhXwMx22gMma8IetZvU3ugCLcB/s72-c/image01.png" height="72" width="72" /><thr:total>0</thr:total><gd:extendedProperty name="commentSource" value="1" /><gd:extendedProperty name="commentModerationMode" value="FILTERED_POSTMOD" /><feedburner:origLink>http://research.googleblog.com/2017/03/an-upgrade-to-syntaxnet-new-models-and.html</feedburner:origLink></entry></feed>
