{"id":26425,"date":"2021-06-09T06:00:00","date_gmt":"2021-06-09T13:00:00","guid":{"rendered":"https:\/\/insidebigdata.com\/?p=26425"},"modified":"2021-06-10T09:29:08","modified_gmt":"2021-06-10T16:29:08","slug":"best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-may-2021","status":"publish","type":"post","link":"https:\/\/insidebigdata.com\/2021\/06\/09\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-may-2021\/","title":{"rendered":"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 May 2021"},"content":{"rendered":"\n<div class=\"wp-block-image is-style-default\"><figure class=\"alignright size-large is-resized\"><img decoding=\"async\" loading=\"lazy\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2013\/12\/arxiv.jpg\" alt=\"\" class=\"wp-image-6361\" width=\"287\" height=\"242\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2013\/12\/arxiv.jpg 450w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2013\/12\/arxiv-150x127.jpg 150w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2013\/12\/arxiv-300x253.jpg 300w\" sizes=\"(max-width: 287px) 100vw, 287px\" \/><\/figure><\/div>\n\n\n\n<p>In this recurring monthly feature, we filter recent research papers appearing on the <a rel=\"noreferrer noopener\" href=\"https:\/\/arxiv.org\/\" target=\"_blank\">arXiv.org<\/a> preprint server for compelling subjects relating to AI, machine learning and deep learning \u2013 from disciplines including statistics, mathematics and computer science \u2013 and provide you with a useful \u201cbest of\u201d list for the past month. Researchers from all over the world contribute to this repository as a prelude to the peer review process for publication in traditional journals. arXiv contains a veritable treasure trove of statistical learning methods you may use one day in the solution of data science problems. The articles listed below represent a small fraction of all articles appearing on the preprint server. They are listed in no particular order with a link to each paper along with a brief overview. Links to GitHub repos are provided when available. Especially relevant articles are marked with a \u201cthumbs up\u201d icon. Consider that these are academic research papers, typically geared toward graduate students, post docs, and seasoned professionals. They generally contain a high degree of mathematics so be prepared. Enjoy!<\/p>\n\n\n\n<p><a href=\"http:\/\/MLP-Mixer: An all-MLP Architecture for Vision\" target=\"_blank\" rel=\"noreferrer noopener\">MLP-Mixer: An all-MLP Architecture for Vision<\/a><\/p>\n\n\n\n<p>Convolutional Neural Networks (CNNs) are the go-to model for computer vision. Recently, attention-based networks, such as the Vision Transformer, have also become popular. This paper shows that while convolutions and attention are both sufficient for good performance, neither of them are necessary. Presented is MLP-Mixer, an architecture based exclusively on multi-layer perceptrons (MLPs). MLP-Mixer contains two types of layers: one with MLPs applied independently to image patches (i.e. &#8220;mixing&#8221; the per-location features), and one with MLPs applied across patches (i.e. &#8220;mixing&#8221; spatial information). When trained on large datasets, or with modern regularization schemes, MLP-Mixer attains competitive scores on image classification benchmarks, with pre-training and inference cost comparable to state-of-the-art models. <\/p>\n\n\n\n<div class=\"wp-block-image is-style-default\"><figure class=\"aligncenter size-large\"><img decoding=\"async\" loading=\"lazy\" width=\"688\" height=\"496\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May1.png\" alt=\"\" class=\"wp-image-26427\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May1.png 688w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May1-150x108.png 150w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May1-300x216.png 300w\" sizes=\"(max-width: 688px) 100vw, 688px\" \/><\/figure><\/div>\n\n\n\n<p><a href=\"https:\/\/arxiv.org\/pdf\/2105.05233v4.pdf\" target=\"_blank\" rel=\"noreferrer noopener\">Diffusion Models Beat GANs on Image Synthesis<\/a><\/p>\n\n\n\n<p>This paper shows that diffusion models can achieve image sample quality superior to the current state-of-the-art generative models. This is achieved with unconditional image synthesis by finding a better architecture through a series of ablations. For conditional image synthesis, we further improve sample quality with classifier guidance: a simple, compute-efficient method for trading off diversity for fidelity using gradients from a classifier. Achieved is an FID of 2.97 on ImageNet 128128, 4.59 on ImageNet 256256, and 7.72 on ImageNet 512512, and BigGAN-deep is matched even with as few as 25 forward passes per sample, all while maintaining better coverage of the distribution. Finally, it is found that classifier guidance combines well with upsampling diffusion models, further improving FID to 3.94 on ImageNet 256256 and 3.85 on ImageNet 512512. The code associated with this paper can be found <a href=\"https:\/\/github.com\/openai\/guided-diffusion\" target=\"_blank\" rel=\"noreferrer noopener\">HERE<\/a>.  <\/p>\n\n\n\n<div class=\"wp-block-image is-style-default\"><figure class=\"aligncenter size-large\"><img decoding=\"async\" loading=\"lazy\" width=\"702\" height=\"385\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May2.png\" alt=\"\" class=\"wp-image-26428\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May2.png 702w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May2-150x82.png 150w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May2-300x165.png 300w\" sizes=\"(max-width: 702px) 100vw, 702px\" \/><\/figure><\/div>\n\n\n\n<p><a href=\"https:\/\/arxiv.org\/pdf\/2105.08050v2.pdf\" target=\"_blank\" rel=\"noreferrer noopener\">Pay Attention to MLPs<\/a><\/p>\n\n\n\n<p>Transformers have become one of the most important architectural innovations in deep learning and have enabled many breakthroughs over the past few years. This paper proposes a simple network architecture, gMLP, based on MLPs with gating, and show that it can perform as well as Transformers in key language and vision applications. The comparisons show that self-attention is not critical for Vision Transformers, as gMLP can achieve the same accuracy. For BERT, the proposed model achieves parity with Transformers on pretraining perplexity and is better on some downstream NLP tasks. On finetuning tasks where gMLP performs worse, making the gMLP model substantially larger can close the gap with Transformers. In general, the experiments show that gMLP can scale as well as Transformers over increased data and compute.<\/p>\n\n\n\n<figure class=\"wp-block-image size-large is-style-default\"><img decoding=\"async\" loading=\"lazy\" width=\"704\" height=\"428\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May3.png\" alt=\"\" class=\"wp-image-26430\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May3.png 704w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May3-150x91.png 150w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May3-300x182.png 300w\" sizes=\"(max-width: 704px) 100vw, 704px\" \/><\/figure>\n\n\n\n<p><a href=\"https:\/\/arxiv.org\/pdf\/2105.09938v2.pdf\" target=\"_blank\" rel=\"noreferrer noopener\">Measuring Coding Challenge Competence With APPS<\/a><\/p>\n\n\n\n<div class=\"wp-block-image is-style-default\"><figure class=\"alignleft size-large\"><img decoding=\"async\" loading=\"lazy\" width=\"100\" height=\"100\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2019\/04\/ThumbsUP_shutterstock_325452782.jpg\" alt=\"\" class=\"wp-image-22440\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2019\/04\/ThumbsUP_shutterstock_325452782.jpg 100w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2019\/04\/ThumbsUP_shutterstock_325452782-50x50.jpg 50w\" sizes=\"(max-width: 100px) 100vw, 100px\" \/><\/figure><\/div>\n\n\n\n<p>While programming is one of the most broadly applicable skills in modern society, modern machine learning models still cannot code solutions to basic problems. Despite its importance, there has been surprisingly little work on evaluating code generation, and it can be difficult to accurately assess code generation performance rigorously. To meet this challenge, this paper introduces APPS, a benchmark for code generation. Unlike prior work in more restricted settings, this benchmark measures the ability of models to take an arbitrary natural language specification and generate satisfactory Python code. Similar to how companies assess candidate software developers, the proposed solution then evaluates models by checking their generated code on test cases. The benchmark includes 10,000 problems, which range from having simple one-line solutions to being substantial algorithmic challenges. Large language models are fined-tuned on both GitHub and a special training set, and it&#8217;s found that the prevalence of syntax errors is decreasing exponentially as models improve. Recent models such as GPT-Neo can pass approximately 20% of the test cases of introductory problems, so it&#8217;s found that machine learning models are now beginning to learn how to code. The code associated with this paper can be found <a href=\"https:\/\/github.com\/hendrycks\/apps\" target=\"_blank\" rel=\"noreferrer noopener\">HERE<\/a>. <\/p>\n\n\n\n<div class=\"wp-block-image is-style-default\"><figure class=\"aligncenter size-large\"><img decoding=\"async\" loading=\"lazy\" width=\"693\" height=\"363\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May4.png\" alt=\"\" class=\"wp-image-26431\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May4.png 693w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May4-150x79.png 150w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May4-300x157.png 300w\" sizes=\"(max-width: 693px) 100vw, 693px\" \/><\/figure><\/div>\n\n\n\n<p><a href=\"https:\/\/arxiv.org\/pdf\/2105.13462v1.pdf\" target=\"_blank\" rel=\"noreferrer noopener\">The Sobolev Regularization Effect of Stochastic Gradient Descent<\/a><\/p>\n\n\n\n<p>The multiplicative structure of parameters and input data in the first layer of neural networks is explored in this paper to build connection between the landscape of the loss function with respect to parameters and the landscape of the model function with respect to input data. By this connection, it is shown that flat minima regularize the gradient of the model function, which explains the good generalization performance of flat minima. Then, the paper goes beyond the flatness and consider high-order moments of the gradient noise, and show that Stochastic Gradient Dascent (SGD) tends to impose constraints on these moments by a linear stability analysis of SGD around global minima. Together with the multiplicative structure, it&#8217;s recognized that the Sobolev regularization effect of SGD, i.e. SGD regularizes the Sobolev seminorms of the model function with respect to the input data. Finally, bounds for generalization error and adversarial robustness are provided for solutions found by SGD under assumptions of the data distribution.<\/p>\n\n\n\n<figure class=\"wp-block-image size-large is-style-default\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"327\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May5.png\" alt=\"\" class=\"wp-image-26433\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May5.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May5-150x70.png 150w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May5-300x140.png 300w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure>\n\n\n\n<p><a href=\"https:\/\/arxiv.org\/pdf\/2105.11653v1.pdf\" target=\"_blank\" rel=\"noreferrer noopener\">Scaling Hierarchical Agglomerative Clustering to Billion-sized Datasets<\/a><\/p>\n\n\n\n<p>Hierarchical Agglomerative Clustering (HAC) is one of the oldest but still most widely used clustering methods. However, HAC is notoriously hard to scale to large data sets as the underlying complexity is at least quadratic in the number of data points and many algorithms to solve HAC are inherently sequential. This paper proposes Reciprocal Agglomerative Clustering (RAC), a distributed algorithm for HAC, that uses a novel strategy to efficiently merge clusters in parallel. The paper proves theoretically that RAC recovers the exact solution of HAC. Furthermore, under clusterability and balancedness assumption it&#8217;s shown provable speedups in total runtime due to the parallelism. It&#8217;s also shown that these speedups are achievable for certain probabilistic data models. In extensive experiments, it&#8217;s shown that this parallelism is achieved on real world data sets and that the proposed RAC algorithm can recover the HAC hierarchy on billions of data points connected by trillions of edges in less than an hour.<\/p>\n\n\n\n<div class=\"wp-block-image is-style-default\"><figure class=\"aligncenter size-large\"><img decoding=\"async\" loading=\"lazy\" width=\"583\" height=\"240\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May6.png\" alt=\"\" class=\"wp-image-26434\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May6.png 583w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May6-150x62.png 150w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May6-300x123.png 300w\" sizes=\"(max-width: 583px) 100vw, 583px\" \/><\/figure><\/div>\n\n\n\n<p><a href=\"https:\/\/arxiv.org\/pdf\/2105.10311v2.pdf\" target=\"_blank\" rel=\"noreferrer noopener\">Pretrained Language Models for Text Generation: A Survey<\/a><\/p>\n\n\n\n<p>Text generation has become one of the most important yet challenging tasks in natural language processing (NLP). The resurgence of deep learning has greatly advanced this field by neural generation models, especially the paradigm of pretrained language models (PLMs). This paper presents an overview of the major advances achieved in the topic of PLMs for text generation. As the preliminaries, the paper presents the general task definition and briefly describe the mainstream architectures of PLMs for text generation. As the core content, the paper discusses how to adapt existing PLMs to model different input data and satisfy special properties in the generated text. <\/p>\n\n\n\n<div class=\"wp-block-image is-style-default\"><figure class=\"aligncenter size-large\"><img decoding=\"async\" loading=\"lazy\" width=\"422\" height=\"279\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May7.png\" alt=\"\" class=\"wp-image-26436\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May7.png 422w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May7-150x99.png 150w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May7-300x198.png 300w\" sizes=\"(max-width: 422px) 100vw, 422px\" \/><\/figure><\/div>\n\n\n\n<p><a href=\"https:\/\/arxiv.org\/pdf\/2105.07033v1.pdf\" target=\"_blank\" rel=\"noreferrer noopener\">Cause and Effect: Concept-based Explanation of Neural Networks<\/a><\/p>\n\n\n\n<p>In many scenarios, human decisions are explained based on some high-level concepts. This paper takes a step in the interpretability of neural networks by examining their internal representation or neuron&#8217;s activations against concepts. A concept is characterized by a set of samples that have specific features in common. A framework is proposed to check the existence of a causal relationship between a concept (or its negation) and task classes. While the previous methods focus on the importance of a concept to a task class, the paper goes further and introduces four measures to quantitatively determine the order of causality. Through experiments, the effectiveness of the proposed method is demonstrated in explaining the relationship between a concept and the predictive behaviour of a neural network.<\/p>\n\n\n\n<div class=\"wp-block-image is-style-default\"><figure class=\"aligncenter size-large\"><img decoding=\"async\" loading=\"lazy\" width=\"444\" height=\"280\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May8.png\" alt=\"\" class=\"wp-image-26438\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May8.png 444w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May8-150x95.png 150w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/arXiv_2021_May8-300x189.png 300w\" sizes=\"(max-width: 444px) 100vw, 444px\" \/><\/figure><\/div>\n\n\n\n<p><em>Sign up for the free insideBIGDATA&nbsp;<a rel=\"noreferrer noopener\" href=\"http:\/\/insidebigdata.com\/newsletter\/\" target=\"_blank\">newsletter<\/a>.<\/em><\/p>\n\n\n\n<p><em>Join us on Twitter:&nbsp;@InsideBigData1 \u2013 <a href=\"https:\/\/twitter.com\/InsideBigData1\" target=\"_blank\" rel=\"noreferrer noopener\">https:\/\/twitter.com\/InsideBigData1<\/a><\/em><\/p>\n","protected":false},"excerpt":{"rendered":"<p>In this recurring monthly feature, we will filter all the recent research papers appearing in the arXiv.org preprint server for subjects relating to AI, machine learning and deep learning \u2013 from disciplines including statistics, mathematics and computer science \u2013 and provide you with a useful \u201cbest of\u201d list for the month.<\/p>\n","protected":false},"author":37,"featured_media":6361,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"jetpack_post_was_ever_published":false,"footnotes":""},"categories":[526,65,115,182,87,180,67,56,1],"tags":[437,741,133,264,277,95],"acf":[],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v20.6 - https:\/\/yoast.com\/wordpress\/plugins\/seo\/ -->\n<title>Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 May 2021 - insideBIGDATA<\/title>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/insidebigdata.com\/2021\/06\/09\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-may-2021\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 May 2021 - insideBIGDATA\" \/>\n<meta property=\"og:description\" content=\"In this recurring monthly feature, we will filter all the recent research papers appearing in the arXiv.org preprint server for subjects relating to AI, machine learning and deep learning \u2013 from disciplines including statistics, mathematics and computer science \u2013 and provide you with a useful \u201cbest of\u201d list for the month.\" \/>\n<meta property=\"og:url\" content=\"https:\/\/insidebigdata.com\/2021\/06\/09\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-may-2021\/\" \/>\n<meta property=\"og:site_name\" content=\"insideBIGDATA\" \/>\n<meta property=\"article:publisher\" content=\"http:\/\/www.facebook.com\/insidebigdata\" \/>\n<meta property=\"article:published_time\" content=\"2021-06-09T13:00:00+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2021-06-10T16:29:08+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2013\/12\/arxiv.jpg\" \/>\n\t<meta property=\"og:image:width\" content=\"450\" \/>\n\t<meta property=\"og:image:height\" content=\"380\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/jpeg\" \/>\n<meta name=\"author\" content=\"Daniel Gutierrez\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:creator\" content=\"@AMULETAnalytics\" \/>\n<meta name=\"twitter:site\" content=\"@insideBigData\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"Daniel Gutierrez\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"8 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"WebPage\",\"@id\":\"https:\/\/insidebigdata.com\/2021\/06\/09\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-may-2021\/\",\"url\":\"https:\/\/insidebigdata.com\/2021\/06\/09\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-may-2021\/\",\"name\":\"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 May 2021 - insideBIGDATA\",\"isPartOf\":{\"@id\":\"https:\/\/insidebigdata.com\/#website\"},\"datePublished\":\"2021-06-09T13:00:00+00:00\",\"dateModified\":\"2021-06-10T16:29:08+00:00\",\"author\":{\"@id\":\"https:\/\/insidebigdata.com\/#\/schema\/person\/2540da209c83a68f4f5922848f7376ed\"},\"breadcrumb\":{\"@id\":\"https:\/\/insidebigdata.com\/2021\/06\/09\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-may-2021\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/insidebigdata.com\/2021\/06\/09\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-may-2021\/\"]}]},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/insidebigdata.com\/2021\/06\/09\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-may-2021\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\/\/insidebigdata.com\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 May 2021\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/insidebigdata.com\/#website\",\"url\":\"https:\/\/insidebigdata.com\/\",\"name\":\"insideBIGDATA\",\"description\":\"Your Source for AI, Data Science, Deep Learning &amp; Machine Learning Strategies\",\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/insidebigdata.com\/?s={search_term_string}\"},\"query-input\":\"required name=search_term_string\"}],\"inLanguage\":\"en-US\"},{\"@type\":\"Person\",\"@id\":\"https:\/\/insidebigdata.com\/#\/schema\/person\/2540da209c83a68f4f5922848f7376ed\",\"name\":\"Daniel Gutierrez\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/insidebigdata.com\/#\/schema\/person\/image\/\",\"url\":\"https:\/\/secure.gravatar.com\/avatar\/5780282e7e567e2a502233e948464542?s=96&d=mm&r=g\",\"contentUrl\":\"https:\/\/secure.gravatar.com\/avatar\/5780282e7e567e2a502233e948464542?s=96&d=mm&r=g\",\"caption\":\"Daniel Gutierrez\"},\"description\":\"Daniel D. Gutierrez is a Data Scientist with Los Angeles-based AMULET Analytics, a service division of AMULET Development Corp. He's been involved with data science and Big Data long before it came in vogue, so imagine his delight when the Harvard Business Review recently deemed \\\"data scientist\\\" as the sexiest profession for the 21st century. Previously, he taught computer science and database classes at UCLA Extension for over 15 years, and authored three computer industry books on database technology. He also served as technical editor, columnist and writer at a major computer industry monthly publication for 7 years. Follow his data science musings at @AMULETAnalytics.\",\"sameAs\":[\"http:\/\/www.insidebigdata.com\",\"https:\/\/twitter.com\/@AMULETAnalytics\"],\"url\":\"https:\/\/insidebigdata.com\/author\/dangutierrez\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 May 2021 - insideBIGDATA","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/insidebigdata.com\/2021\/06\/09\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-may-2021\/","og_locale":"en_US","og_type":"article","og_title":"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 May 2021 - insideBIGDATA","og_description":"In this recurring monthly feature, we will filter all the recent research papers appearing in the arXiv.org preprint server for subjects relating to AI, machine learning and deep learning \u2013 from disciplines including statistics, mathematics and computer science \u2013 and provide you with a useful \u201cbest of\u201d list for the month.","og_url":"https:\/\/insidebigdata.com\/2021\/06\/09\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-may-2021\/","og_site_name":"insideBIGDATA","article_publisher":"http:\/\/www.facebook.com\/insidebigdata","article_published_time":"2021-06-09T13:00:00+00:00","article_modified_time":"2021-06-10T16:29:08+00:00","og_image":[{"width":450,"height":380,"url":"https:\/\/insidebigdata.com\/wp-content\/uploads\/2013\/12\/arxiv.jpg","type":"image\/jpeg"}],"author":"Daniel Gutierrez","twitter_card":"summary_large_image","twitter_creator":"@AMULETAnalytics","twitter_site":"@insideBigData","twitter_misc":{"Written by":"Daniel Gutierrez","Est. reading time":"8 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"WebPage","@id":"https:\/\/insidebigdata.com\/2021\/06\/09\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-may-2021\/","url":"https:\/\/insidebigdata.com\/2021\/06\/09\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-may-2021\/","name":"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 May 2021 - insideBIGDATA","isPartOf":{"@id":"https:\/\/insidebigdata.com\/#website"},"datePublished":"2021-06-09T13:00:00+00:00","dateModified":"2021-06-10T16:29:08+00:00","author":{"@id":"https:\/\/insidebigdata.com\/#\/schema\/person\/2540da209c83a68f4f5922848f7376ed"},"breadcrumb":{"@id":"https:\/\/insidebigdata.com\/2021\/06\/09\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-may-2021\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/insidebigdata.com\/2021\/06\/09\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-may-2021\/"]}]},{"@type":"BreadcrumbList","@id":"https:\/\/insidebigdata.com\/2021\/06\/09\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-may-2021\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/insidebigdata.com\/"},{"@type":"ListItem","position":2,"name":"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 May 2021"}]},{"@type":"WebSite","@id":"https:\/\/insidebigdata.com\/#website","url":"https:\/\/insidebigdata.com\/","name":"insideBIGDATA","description":"Your Source for AI, Data Science, Deep Learning &amp; Machine Learning Strategies","potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/insidebigdata.com\/?s={search_term_string}"},"query-input":"required name=search_term_string"}],"inLanguage":"en-US"},{"@type":"Person","@id":"https:\/\/insidebigdata.com\/#\/schema\/person\/2540da209c83a68f4f5922848f7376ed","name":"Daniel Gutierrez","image":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/insidebigdata.com\/#\/schema\/person\/image\/","url":"https:\/\/secure.gravatar.com\/avatar\/5780282e7e567e2a502233e948464542?s=96&d=mm&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/5780282e7e567e2a502233e948464542?s=96&d=mm&r=g","caption":"Daniel Gutierrez"},"description":"Daniel D. Gutierrez is a Data Scientist with Los Angeles-based AMULET Analytics, a service division of AMULET Development Corp. He's been involved with data science and Big Data long before it came in vogue, so imagine his delight when the Harvard Business Review recently deemed \"data scientist\" as the sexiest profession for the 21st century. Previously, he taught computer science and database classes at UCLA Extension for over 15 years, and authored three computer industry books on database technology. He also served as technical editor, columnist and writer at a major computer industry monthly publication for 7 years. Follow his data science musings at @AMULETAnalytics.","sameAs":["http:\/\/www.insidebigdata.com","https:\/\/twitter.com\/@AMULETAnalytics"],"url":"https:\/\/insidebigdata.com\/author\/dangutierrez\/"}]}},"jetpack_featured_media_url":"https:\/\/insidebigdata.com\/wp-content\/uploads\/2013\/12\/arxiv.jpg","jetpack_shortlink":"https:\/\/wp.me\/p9eA3j-6Sd","jetpack-related-posts":[{"id":21994,"url":"https:\/\/insidebigdata.com\/2019\/01\/16\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-december-2018\/","url_meta":{"origin":26425,"position":0},"title":"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 December 2018","date":"January 16, 2019","format":false,"excerpt":"In this recurring monthly feature, we will filter all the recent research papers appearing in the arXiv.org preprint server for subjects relating to AI, machine learning and deep learning \u2013 from disciplines including statistics, mathematics and computer science \u2013 and provide you with a useful \u201cbest of\u201d list for the\u2026","rel":"","context":"In &quot;AI Deep Learning&quot;","img":{"alt_text":"","src":"https:\/\/i0.wp.com\/insidebigdata.com\/wp-content\/uploads\/2013\/12\/arxiv.jpg?resize=350%2C200&ssl=1","width":350,"height":200},"classes":[]},{"id":22953,"url":"https:\/\/insidebigdata.com\/2019\/07\/18\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-june-2019\/","url_meta":{"origin":26425,"position":1},"title":"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 June 2019","date":"July 18, 2019","format":false,"excerpt":"In this recurring monthly feature, we will filter all the recent research papers appearing in the arXiv.org preprint server for subjects relating to AI, machine learning and deep learning \u2013 from disciplines including statistics, mathematics and computer science \u2013 and provide you with a useful \u201cbest of\u201d list for the\u2026","rel":"","context":"In &quot;AI Deep Learning&quot;","img":{"alt_text":"","src":"https:\/\/i0.wp.com\/insidebigdata.com\/wp-content\/uploads\/2013\/12\/arxiv.jpg?resize=350%2C200&ssl=1","width":350,"height":200},"classes":[]},{"id":21252,"url":"https:\/\/insidebigdata.com\/2018\/10\/17\/best-arxiv-org-ai-machine-learning-deep-learning-september-2018\/","url_meta":{"origin":26425,"position":2},"title":"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 September 2018","date":"October 17, 2018","format":false,"excerpt":"In this recurring monthly feature, we will filter all the recent research papers appearing in the arXiv.org preprint server for subjects relating to AI, machine learning and deep learning \u2013 from disciplines including statistics, mathematics and computer science \u2013 and provide you with a useful \u201cbest of\u201d list for the\u2026","rel":"","context":"In &quot;AI Deep Learning&quot;","img":{"alt_text":"","src":"https:\/\/i0.wp.com\/insidebigdata.com\/wp-content\/uploads\/2013\/12\/arxiv.jpg?resize=350%2C200&ssl=1","width":350,"height":200},"classes":[]},{"id":20910,"url":"https:\/\/insidebigdata.com\/2018\/08\/13\/best-arxiv-org-ai-machine-learning-deep-learning-july-2018\/","url_meta":{"origin":26425,"position":3},"title":"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 July 2018","date":"August 13, 2018","format":false,"excerpt":"In this recurring monthly feature, we will filter all the recent research papers appearing in the arXiv.org preprint server for subjects relating to AI, machine learning and deep learning \u2013 from disciplines including statistics, mathematics and computer science \u2013 and provide you with a useful \u201cbest of\u201d list for the\u2026","rel":"","context":"In &quot;AI Deep Learning&quot;","img":{"alt_text":"","src":"https:\/\/i0.wp.com\/insidebigdata.com\/wp-content\/uploads\/2013\/12\/arxiv.jpg?resize=350%2C200&ssl=1","width":350,"height":200},"classes":[]},{"id":22263,"url":"https:\/\/insidebigdata.com\/2019\/03\/15\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-february-2019\/","url_meta":{"origin":26425,"position":4},"title":"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 February 2019","date":"March 15, 2019","format":false,"excerpt":"In this recurring monthly feature, we will filter all the recent research papers appearing in the arXiv.org preprint server for subjects relating to AI, machine learning and deep learning \u2013 from disciplines including statistics, mathematics and computer science \u2013 and provide you with a useful \u201cbest of\u201d list for the\u2026","rel":"","context":"In &quot;AI Deep Learning&quot;","img":{"alt_text":"","src":"https:\/\/i0.wp.com\/insidebigdata.com\/wp-content\/uploads\/2013\/12\/arxiv.jpg?resize=350%2C200&ssl=1","width":350,"height":200},"classes":[]},{"id":22431,"url":"https:\/\/insidebigdata.com\/2019\/04\/09\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-march-2019\/","url_meta":{"origin":26425,"position":5},"title":"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 March 2019","date":"April 9, 2019","format":false,"excerpt":"In this recurring monthly feature, we will filter all the recent research papers appearing in the arXiv.org preprint server for subjects relating to AI, machine learning and deep learning \u2013 from disciplines including statistics, mathematics and computer science \u2013 and provide you with a useful \u201cbest of\u201d list for the\u2026","rel":"","context":"In &quot;AI Deep Learning&quot;","img":{"alt_text":"","src":"https:\/\/i0.wp.com\/insidebigdata.com\/wp-content\/uploads\/2013\/12\/arxiv.jpg?resize=350%2C200&ssl=1","width":350,"height":200},"classes":[]}],"_links":{"self":[{"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/posts\/26425"}],"collection":[{"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/users\/37"}],"replies":[{"embeddable":true,"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/comments?post=26425"}],"version-history":[{"count":0,"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/posts\/26425\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/media\/6361"}],"wp:attachment":[{"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/media?parent=26425"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/categories?post=26425"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/tags?post=26425"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}