{"id":27358,"date":"2021-10-15T06:00:00","date_gmt":"2021-10-15T13:00:00","guid":{"rendered":"https:\/\/insidebigdata.com\/?p=27358"},"modified":"2021-10-16T10:04:05","modified_gmt":"2021-10-16T17:04:05","slug":"best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-september-2021","status":"publish","type":"post","link":"https:\/\/insidebigdata.com\/2021\/10\/15\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-september-2021\/","title":{"rendered":"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 September 2021"},"content":{"rendered":"\n<div class=\"wp-block-image is-style-default\"><figure class=\"alignright size-full is-resized\"><img decoding=\"async\" loading=\"lazy\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2013\/12\/arxiv.jpg\" alt=\"\" class=\"wp-image-6361\" width=\"239\" height=\"202\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2013\/12\/arxiv.jpg 450w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2013\/12\/arxiv-150x127.jpg 150w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2013\/12\/arxiv-300x253.jpg 300w\" sizes=\"(max-width: 239px) 100vw, 239px\" \/><\/figure><\/div>\n\n\n\n<p>In this recurring monthly feature, we filter recent research papers appearing on the <a rel=\"noreferrer noopener\" href=\"https:\/\/arxiv.org\/\" target=\"_blank\">arXiv.org<\/a> preprint server for compelling subjects relating to AI, machine learning and deep learning \u2013 from disciplines including statistics, mathematics and computer science \u2013 and provide you with a useful \u201cbest of\u201d list for the past month. Researchers from all over the world contribute to this repository as a prelude to the peer review process for publication in traditional journals. arXiv contains a veritable treasure trove of statistical learning methods you may use one day in the solution of data science problems. The articles listed below represent a small fraction of all articles appearing on the preprint server. They are listed in no particular order with a link to each paper along with a brief overview. Links to GitHub repos are provided when available. Especially relevant articles are marked with a \u201cthumbs up\u201d icon. Consider that these are academic research papers, typically geared toward graduate students, post docs, and seasoned professionals. They generally contain a high degree of mathematics so be prepared. Enjoy!<\/p>\n\n\n\n<p><a href=\"https:\/\/arxiv.org\/pdf\/2109.05070v1.pdf\" target=\"_blank\" rel=\"noreferrer noopener\">Instance-Conditioned GAN<\/a><\/p>\n\n\n\n<p>Generative Adversarial Networks (GANs) can generate near photo realistic images in narrow domains such as human faces. Yet, modeling complex distributions of datasets such as ImageNet and COCO-Stuff remains challenging in unconditional settings. This paper takes inspiration from kernel density estimation techniques and introduces a non-parametric approach to modeling distributions of complex datasets. The data manifold is partitioned into a mixture of overlapping neighborhoods described by a datapoint and its nearest neighbors, and a model is introduced called instance-conditioned GAN (IC-GAN), which learns the distribution around each datapoint. Experimental results on ImageNet and COCO-Stuff show that IC-GAN significantly improves over unconditional models and unsupervised data partitioning baselines. It&#8217;s also shown that IC-GAN can effortlessly transfer to datasets not seen during training by simply changing the conditioning instances, and still generate realistic images. The GitHub repo associated with this paper can be found <a href=\"https:\/\/github.com\/facebookresearch\/ic_gan\" target=\"_blank\" rel=\"noreferrer noopener\">HERE<\/a>. <\/p>\n\n\n\n<div class=\"wp-block-image is-style-default\"><figure class=\"aligncenter size-large\"><img decoding=\"async\" loading=\"lazy\" width=\"690\" height=\"396\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/10\/arXiv_2021_09_1.png\" alt=\"\" class=\"wp-image-27362\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/10\/arXiv_2021_09_1.png 690w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/10\/arXiv_2021_09_1-150x86.png 150w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/10\/arXiv_2021_09_1-300x172.png 300w\" sizes=\"(max-width: 690px) 100vw, 690px\" \/><\/figure><\/div>\n\n\n\n<p><a href=\"https:\/\/arxiv.org\/pdf\/2109.10852v1.pdf\" target=\"_blank\" rel=\"noreferrer noopener\">Pix2seq: A Language Modeling Framework for Object Detection<\/a><\/p>\n\n\n\n<p>This paper presents Pix2Seq, a simple and generic framework for object detection. Unlike existing approaches that explicitly integrate prior knowledge about the task, Google researchers including Geoffrey Hinton, simply cast object detection as a language modeling task conditioned on the observed pixel inputs. Object descriptions (e.g., bounding boxes and class labels) are expressed as sequences of discrete tokens, and a neural net is trained to perceive the image and generate the desired sequence. The approach is based mainly on the intuition that if a neural net knows about where and what the objects are, we just need to teach it how to read them out. Beyond the use of task-specific data augmentations, the approach makes minimal assumptions about the task, yet it achieves competitive results on the challenging COCO dataset, compared to highly specialized and well optimized detection algorithms. The GitHub repo associated with this paper can be found <a href=\"https:\/\/github.com\/gaopengcuhk\/Stable-Pix2Seq\" target=\"_blank\" rel=\"noreferrer noopener\">HERE<\/a>.<\/p>\n\n\n\n<figure class=\"wp-block-image size-large is-style-default\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"342\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/10\/arXiv_2021_09_2.png\" alt=\"\" class=\"wp-image-27365\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/10\/arXiv_2021_09_2.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/10\/arXiv_2021_09_2-150x73.png 150w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/10\/arXiv_2021_09_2-300x147.png 300w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure>\n\n\n\n<p><a href=\"https:\/\/arxiv.org\/pdf\/2109.10686v1.pdf\" target=\"_blank\" rel=\"noreferrer noopener\">Scale Efficiently: Insights from Pre-training and Fine-tuning Transformers<\/a><\/p>\n\n\n\n<p>There remain many open questions pertaining to the scaling behavior of Transformer architectures. These scaling decisions and findings can be critical, as training runs often come with an associated computational cost which have both financial and\/or environmental impact. The goal of this paper is to present scaling insights from pretraining and finetuning Transformers. While Kaplan et al. presents a comprehensive study of the scaling behavior of Transformer language models, the scope is only on the upstream (pretraining) loss. Therefore, it is still unclear if these set of findings transfer to downstream task within the context of the pretrain-finetune paradigm. The key findings of this paper are as follows: (1) aside from only the model size, model shape matters for downstream fine-tuning, (2) scaling protocols operate differently at different compute regions, (3) widely adopted T5-base and T5-large sizes are Pareto-inefficient. To this end, improved scaling protocols are presented whereby the redesigned models achieve similar downstream fine-tuning quality while having 50\\% fewer parameters and training 40\\% faster compared to the widely adopted T5-base model. <\/p>\n\n\n\n<div class=\"wp-block-image is-style-default\"><figure class=\"aligncenter size-large\"><img decoding=\"async\" loading=\"lazy\" width=\"689\" height=\"429\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/10\/arXiv_2021_09_3.png\" alt=\"\" class=\"wp-image-27367\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/10\/arXiv_2021_09_3.png 689w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/10\/arXiv_2021_09_3-150x93.png 150w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/10\/arXiv_2021_09_3-300x187.png 300w\" sizes=\"(max-width: 689px) 100vw, 689px\" \/><\/figure><\/div>\n\n\n\n<p><a href=\"https:\/\/arxiv.org\/pdf\/2109.08668v1.pdf\" target=\"_blank\" rel=\"noreferrer noopener\">Primer: Searching for Efficient Transformers for Language Modeling<\/a><\/p>\n\n\n\n<div class=\"wp-block-image is-style-default\"><figure class=\"alignleft size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"100\" height=\"100\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2019\/04\/ThumbsUP_shutterstock_325452782.jpg\" alt=\"\" class=\"wp-image-22440\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2019\/04\/ThumbsUP_shutterstock_325452782.jpg 100w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2019\/04\/ThumbsUP_shutterstock_325452782-50x50.jpg 50w\" sizes=\"(max-width: 100px) 100vw, 100px\" \/><\/figure><\/div>\n\n\n\n<p>Large Transformer models have been central to recent advances in natural language processing. The training and inference costs of these models, however, have grown rapidly and become prohibitively expensive. The paper aims to reduce the costs of Transformers by searching for a more efficient variant. Compared to previous approaches, the search is performed at a lower level, over the primitives that define a Transformer TensorFlow program. An architecture is identified, named Primer, that has a smaller training cost than the original Transformer and other variants for auto-regressive language modeling. Primer&#8217;s improvements can be mostly attributed to two simple modifications: squaring ReLU activations and adding a depthwise convolution layer after each Q, K, and V projection in self-attention. The GitHub repo associated with this paper can be found <a href=\"https:\/\/github.com\/google-research\/google-research\/tree\/master\/primer\" target=\"_blank\" rel=\"noreferrer noopener\">HERE<\/a>. <\/p>\n\n\n\n<div class=\"wp-block-image is-style-default\"><figure class=\"aligncenter size-large\"><img decoding=\"async\" loading=\"lazy\" width=\"694\" height=\"361\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/10\/arXiv_2021_09_4.png\" alt=\"\" class=\"wp-image-27369\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/10\/arXiv_2021_09_4.png 694w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/10\/arXiv_2021_09_4-150x78.png 150w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/10\/arXiv_2021_09_4-300x156.png 300w\" sizes=\"(max-width: 694px) 100vw, 694px\" \/><\/figure><\/div>\n\n\n\n<p><a href=\"https:\/\/arxiv.org\/pdf\/2109.03144v1.pdf\" target=\"_blank\" rel=\"noreferrer noopener\">PP-OCRv2: Bag of Tricks for Ultra Lightweight OCR System<\/a><\/p>\n\n\n\n<p>Optical Character Recognition (OCR) systems have been widely used in various of application scenarios. Designing an OCR system is still a challenging task. This paper proposes a robust OCR system, i.e. PP-OCRv2. A bag of tricks is introduced to train a better text detector and a better text recognizer, which include Collaborative Mutual Learning (CML), CopyPaste, Lightweight CPUNetwork (LCNet), Unified-Deep Mutual Learning (U-DML) and Enhanced CTCLoss. Experiments on real data show that the precision of PP-OCRv2 is 7% higher than PP-OCR under the same inference cost. It is also comparable to the server models of the PP-OCR which uses ResNet series as backbones. All of the above mentioned models are open-sourced and the code is available in the GitHub repository PaddleOCR which is powered by PaddlePaddle. The GitHub report associated with this paper can be found <a href=\"https:\/\/github.com\/PaddlePaddle\/PaddleOCR\" target=\"_blank\" rel=\"noreferrer noopener\">HERE<\/a>. <\/p>\n\n\n\n<div class=\"wp-block-image is-style-default\"><figure class=\"aligncenter size-large\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"359\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/10\/arXiv_2021_09_5.png\" alt=\"\" class=\"wp-image-27372\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/10\/arXiv_2021_09_5.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/10\/arXiv_2021_09_5-150x77.png 150w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/10\/arXiv_2021_09_5-300x154.png 300w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n\n\n<p><a href=\"https:\/\/arxiv.org\/pdf\/2109.13157v1.pdf\" target=\"_blank\" rel=\"noreferrer noopener\">Visual Anomaly Detection for Images: A Survey<\/a><\/p>\n\n\n\n<div class=\"wp-block-image is-style-default\"><figure class=\"alignleft size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"100\" height=\"100\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2019\/04\/ThumbsUP_shutterstock_325452782.jpg\" alt=\"\" class=\"wp-image-22440\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2019\/04\/ThumbsUP_shutterstock_325452782.jpg 100w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2019\/04\/ThumbsUP_shutterstock_325452782-50x50.jpg 50w\" sizes=\"(max-width: 100px) 100vw, 100px\" \/><\/figure><\/div>\n\n\n\n<p>Visual anomaly detection is an important and challenging problem in the field of machine learning and computer vision. This problem has attracted a considerable amount of attention in relevant research communities. Especially in recent years, the development of deep learning has sparked an increasing interest in the visual anomaly detection problem and brought a great variety of novel methods. This paper provides a comprehensive survey of the classical and deep learning-based approaches for visual anomaly detection in the literature. The relevant approaches are grouped in view of their underlying principles and discuss their assumptions, advantages, and disadvantages carefully. The aim is to help the researchers to understand the common principles of visual anomaly detection approaches and identify promising research directions in this field.<\/p>\n\n\n\n<p><a href=\"https:\/\/arxiv.org\/pdf\/2109.10431v1.pdf\" target=\"_blank\" rel=\"noreferrer noopener\">Fairness without Imputation: A Decision Tree Approach for Fair Prediction with Missing Values<\/a><\/p>\n\n\n\n<p>This paper investigates the fairness concerns of training a machine learning model using data with missing values. Even though there are a number of fairness intervention methods in the literature, most of them require a complete training set as input. In practice, data can have missing values, and data missing patterns can depend on group attributes (e.g. gender or race). Simply applying off-the-shelf fair learning algorithms to an imputed dataset may lead to an unfair model. This paper theoretically analyzes different sources of discrimination risks when training with an imputed dataset. An integrated approach is proposed based on decision trees that does not require a separate process of imputation and learning. Instead, a tree is trained with missing incorporated as attribute (MIA), which does not require explicit imputation, and a fairness-regularized objective function is optimized. The approach outperforms existing fairness intervention methods applied to an imputed dataset, through several experiments on real-world datasets.<\/p>\n\n\n\n<div class=\"wp-block-image is-style-default\"><figure class=\"aligncenter size-large\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"285\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/10\/arXiv_2021_09_6a.png\" alt=\"\" class=\"wp-image-27378\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/10\/arXiv_2021_09_6a.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/10\/arXiv_2021_09_6a-150x61.png 150w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/10\/arXiv_2021_09_6a-300x122.png 300w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n\n\n<p><a href=\"https:\/\/arxiv.org\/pdf\/2109.09265v1.pdf\" target=\"_blank\" rel=\"noreferrer noopener\">Merlion: A Machine Learning Library for Time Series<\/a><\/p>\n\n\n\n<p>This paper introduces Merlion, an open-source machine learning library for time series. It features a unified interface for many commonly used models and datasets for anomaly detection and forecasting on both univariate and multivariate time series, along with standard pre\/post-processing layers. It has several modules to improve ease-of-use, including visualization, anomaly score calibration to improve interpetability, AutoML for hyperparameter tuning and model selection, and model ensembling. Merlion also provides a unique evaluation framework that simulates the live deployment and re-training of a model in production. This library aims to provide engineers and researchers a one-stop solution to rapidly develop models for their specific time series needs and benchmark them across multiple time series datasets. This technical report highlights Merlion&#8217;s architecture and major functionalities, and a report of benchmark numbers across different baseline models and ensembles.<\/p>\n\n\n\n<div class=\"wp-block-image is-style-default\"><figure class=\"aligncenter size-large\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"362\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/10\/arXiv_2021_09_7.png\" alt=\"\" class=\"wp-image-27380\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/10\/arXiv_2021_09_7.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/10\/arXiv_2021_09_7-150x78.png 150w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/10\/arXiv_2021_09_7-300x155.png 300w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n\n\n<p><a href=\"https:\/\/arxiv.org\/pdf\/2109.09014v1.pdf\" target=\"_blank\" rel=\"noreferrer noopener\">A Machine Learning Pipeline to Examine Political Bias with Congressional Speeches<\/a><\/p>\n\n\n\n<p>Computational methods to model political bias in social media involve several challenges due to heterogeneity, high-dimensional, multiple modalities, and the scale of the data. Political bias in social media has been studied in multiple viewpoints like media bias, political ideology, echo chambers, and controversies using machine learning pipelines. Most of the current methods rely heavily on the manually-labeled ground-truth data for the underlying political bias prediction tasks. Limitations of such methods include human-intensive labeling, labels related to only a specific problem, and the inability to determine the near future bias state of a social media conversation. This paper addresses such problems and give machine learning approaches to study political bias in two ideologically diverse social media forums: Gab and Twitter without the availability of human-annotated data. <\/p>\n\n\n\n<p><em>Sign up for the free insideBIGDATA&nbsp;<a rel=\"noreferrer noopener\" href=\"http:\/\/insidebigdata.com\/newsletter\/\" target=\"_blank\">newsletter<\/a>.<\/em><\/p>\n\n\n\n<p><em>Join us on Twitter:&nbsp;@InsideBigData1 \u2013 <a href=\"https:\/\/twitter.com\/InsideBigData1\" target=\"_blank\" rel=\"noreferrer noopener\">https:\/\/twitter.com\/InsideBigData1<\/a><\/em><\/p>\n","protected":false},"excerpt":{"rendered":"<p>In this recurring monthly feature, we will filter all the recent research papers appearing in the arXiv.org preprint server for subjects relating to AI, machine learning and deep learning \u2013 from disciplines including statistics, mathematics and computer science \u2013 and provide you with a useful \u201cbest of\u201d list for the month.<\/p>\n","protected":false},"author":37,"featured_media":6361,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"jetpack_post_was_ever_published":false,"footnotes":""},"categories":[526,65,115,68,87,180,67,56,84,1],"tags":[437,324,741,133,264,277,96],"acf":[],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v20.6 - https:\/\/yoast.com\/wordpress\/plugins\/seo\/ -->\n<title>Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 September 2021 - insideBIGDATA<\/title>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/insidebigdata.com\/2021\/10\/15\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-september-2021\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 September 2021 - insideBIGDATA\" \/>\n<meta property=\"og:description\" content=\"In this recurring monthly feature, we will filter all the recent research papers appearing in the arXiv.org preprint server for subjects relating to AI, machine learning and deep learning \u2013 from disciplines including statistics, mathematics and computer science \u2013 and provide you with a useful \u201cbest of\u201d list for the month.\" \/>\n<meta property=\"og:url\" content=\"https:\/\/insidebigdata.com\/2021\/10\/15\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-september-2021\/\" \/>\n<meta property=\"og:site_name\" content=\"insideBIGDATA\" \/>\n<meta property=\"article:publisher\" content=\"http:\/\/www.facebook.com\/insidebigdata\" \/>\n<meta property=\"article:published_time\" content=\"2021-10-15T13:00:00+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2021-10-16T17:04:05+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2013\/12\/arxiv.jpg\" \/>\n\t<meta property=\"og:image:width\" content=\"450\" \/>\n\t<meta property=\"og:image:height\" content=\"380\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/jpeg\" \/>\n<meta name=\"author\" content=\"Daniel Gutierrez\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:creator\" content=\"@AMULETAnalytics\" \/>\n<meta name=\"twitter:site\" content=\"@insideBigData\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"Daniel Gutierrez\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"9 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"WebPage\",\"@id\":\"https:\/\/insidebigdata.com\/2021\/10\/15\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-september-2021\/\",\"url\":\"https:\/\/insidebigdata.com\/2021\/10\/15\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-september-2021\/\",\"name\":\"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 September 2021 - insideBIGDATA\",\"isPartOf\":{\"@id\":\"https:\/\/insidebigdata.com\/#website\"},\"datePublished\":\"2021-10-15T13:00:00+00:00\",\"dateModified\":\"2021-10-16T17:04:05+00:00\",\"author\":{\"@id\":\"https:\/\/insidebigdata.com\/#\/schema\/person\/2540da209c83a68f4f5922848f7376ed\"},\"breadcrumb\":{\"@id\":\"https:\/\/insidebigdata.com\/2021\/10\/15\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-september-2021\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/insidebigdata.com\/2021\/10\/15\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-september-2021\/\"]}]},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/insidebigdata.com\/2021\/10\/15\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-september-2021\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\/\/insidebigdata.com\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 September 2021\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/insidebigdata.com\/#website\",\"url\":\"https:\/\/insidebigdata.com\/\",\"name\":\"insideBIGDATA\",\"description\":\"Your Source for AI, Data Science, Deep Learning &amp; Machine Learning Strategies\",\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/insidebigdata.com\/?s={search_term_string}\"},\"query-input\":\"required name=search_term_string\"}],\"inLanguage\":\"en-US\"},{\"@type\":\"Person\",\"@id\":\"https:\/\/insidebigdata.com\/#\/schema\/person\/2540da209c83a68f4f5922848f7376ed\",\"name\":\"Daniel Gutierrez\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/insidebigdata.com\/#\/schema\/person\/image\/\",\"url\":\"https:\/\/secure.gravatar.com\/avatar\/5780282e7e567e2a502233e948464542?s=96&d=mm&r=g\",\"contentUrl\":\"https:\/\/secure.gravatar.com\/avatar\/5780282e7e567e2a502233e948464542?s=96&d=mm&r=g\",\"caption\":\"Daniel Gutierrez\"},\"description\":\"Daniel D. Gutierrez is a Data Scientist with Los Angeles-based AMULET Analytics, a service division of AMULET Development Corp. He's been involved with data science and Big Data long before it came in vogue, so imagine his delight when the Harvard Business Review recently deemed \\\"data scientist\\\" as the sexiest profession for the 21st century. Previously, he taught computer science and database classes at UCLA Extension for over 15 years, and authored three computer industry books on database technology. He also served as technical editor, columnist and writer at a major computer industry monthly publication for 7 years. Follow his data science musings at @AMULETAnalytics.\",\"sameAs\":[\"http:\/\/www.insidebigdata.com\",\"https:\/\/twitter.com\/@AMULETAnalytics\"],\"url\":\"https:\/\/insidebigdata.com\/author\/dangutierrez\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 September 2021 - insideBIGDATA","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/insidebigdata.com\/2021\/10\/15\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-september-2021\/","og_locale":"en_US","og_type":"article","og_title":"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 September 2021 - insideBIGDATA","og_description":"In this recurring monthly feature, we will filter all the recent research papers appearing in the arXiv.org preprint server for subjects relating to AI, machine learning and deep learning \u2013 from disciplines including statistics, mathematics and computer science \u2013 and provide you with a useful \u201cbest of\u201d list for the month.","og_url":"https:\/\/insidebigdata.com\/2021\/10\/15\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-september-2021\/","og_site_name":"insideBIGDATA","article_publisher":"http:\/\/www.facebook.com\/insidebigdata","article_published_time":"2021-10-15T13:00:00+00:00","article_modified_time":"2021-10-16T17:04:05+00:00","og_image":[{"width":450,"height":380,"url":"https:\/\/insidebigdata.com\/wp-content\/uploads\/2013\/12\/arxiv.jpg","type":"image\/jpeg"}],"author":"Daniel Gutierrez","twitter_card":"summary_large_image","twitter_creator":"@AMULETAnalytics","twitter_site":"@insideBigData","twitter_misc":{"Written by":"Daniel Gutierrez","Est. reading time":"9 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"WebPage","@id":"https:\/\/insidebigdata.com\/2021\/10\/15\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-september-2021\/","url":"https:\/\/insidebigdata.com\/2021\/10\/15\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-september-2021\/","name":"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 September 2021 - insideBIGDATA","isPartOf":{"@id":"https:\/\/insidebigdata.com\/#website"},"datePublished":"2021-10-15T13:00:00+00:00","dateModified":"2021-10-16T17:04:05+00:00","author":{"@id":"https:\/\/insidebigdata.com\/#\/schema\/person\/2540da209c83a68f4f5922848f7376ed"},"breadcrumb":{"@id":"https:\/\/insidebigdata.com\/2021\/10\/15\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-september-2021\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/insidebigdata.com\/2021\/10\/15\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-september-2021\/"]}]},{"@type":"BreadcrumbList","@id":"https:\/\/insidebigdata.com\/2021\/10\/15\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-september-2021\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/insidebigdata.com\/"},{"@type":"ListItem","position":2,"name":"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 September 2021"}]},{"@type":"WebSite","@id":"https:\/\/insidebigdata.com\/#website","url":"https:\/\/insidebigdata.com\/","name":"insideBIGDATA","description":"Your Source for AI, Data Science, Deep Learning &amp; Machine Learning Strategies","potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/insidebigdata.com\/?s={search_term_string}"},"query-input":"required name=search_term_string"}],"inLanguage":"en-US"},{"@type":"Person","@id":"https:\/\/insidebigdata.com\/#\/schema\/person\/2540da209c83a68f4f5922848f7376ed","name":"Daniel Gutierrez","image":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/insidebigdata.com\/#\/schema\/person\/image\/","url":"https:\/\/secure.gravatar.com\/avatar\/5780282e7e567e2a502233e948464542?s=96&d=mm&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/5780282e7e567e2a502233e948464542?s=96&d=mm&r=g","caption":"Daniel Gutierrez"},"description":"Daniel D. Gutierrez is a Data Scientist with Los Angeles-based AMULET Analytics, a service division of AMULET Development Corp. He's been involved with data science and Big Data long before it came in vogue, so imagine his delight when the Harvard Business Review recently deemed \"data scientist\" as the sexiest profession for the 21st century. Previously, he taught computer science and database classes at UCLA Extension for over 15 years, and authored three computer industry books on database technology. He also served as technical editor, columnist and writer at a major computer industry monthly publication for 7 years. Follow his data science musings at @AMULETAnalytics.","sameAs":["http:\/\/www.insidebigdata.com","https:\/\/twitter.com\/@AMULETAnalytics"],"url":"https:\/\/insidebigdata.com\/author\/dangutierrez\/"}]}},"jetpack_featured_media_url":"https:\/\/insidebigdata.com\/wp-content\/uploads\/2013\/12\/arxiv.jpg","jetpack_shortlink":"https:\/\/wp.me\/p9eA3j-77g","jetpack-related-posts":[{"id":21994,"url":"https:\/\/insidebigdata.com\/2019\/01\/16\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-december-2018\/","url_meta":{"origin":27358,"position":0},"title":"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 December 2018","date":"January 16, 2019","format":false,"excerpt":"In this recurring monthly feature, we will filter all the recent research papers appearing in the arXiv.org preprint server for subjects relating to AI, machine learning and deep learning \u2013 from disciplines including statistics, mathematics and computer science \u2013 and provide you with a useful \u201cbest of\u201d list for the\u2026","rel":"","context":"In &quot;AI Deep Learning&quot;","img":{"alt_text":"","src":"https:\/\/i0.wp.com\/insidebigdata.com\/wp-content\/uploads\/2013\/12\/arxiv.jpg?resize=350%2C200&ssl=1","width":350,"height":200},"classes":[]},{"id":22953,"url":"https:\/\/insidebigdata.com\/2019\/07\/18\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-june-2019\/","url_meta":{"origin":27358,"position":1},"title":"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 June 2019","date":"July 18, 2019","format":false,"excerpt":"In this recurring monthly feature, we will filter all the recent research papers appearing in the arXiv.org preprint server for subjects relating to AI, machine learning and deep learning \u2013 from disciplines including statistics, mathematics and computer science \u2013 and provide you with a useful \u201cbest of\u201d list for the\u2026","rel":"","context":"In &quot;AI Deep Learning&quot;","img":{"alt_text":"","src":"https:\/\/i0.wp.com\/insidebigdata.com\/wp-content\/uploads\/2013\/12\/arxiv.jpg?resize=350%2C200&ssl=1","width":350,"height":200},"classes":[]},{"id":22431,"url":"https:\/\/insidebigdata.com\/2019\/04\/09\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-march-2019\/","url_meta":{"origin":27358,"position":2},"title":"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 March 2019","date":"April 9, 2019","format":false,"excerpt":"In this recurring monthly feature, we will filter all the recent research papers appearing in the arXiv.org preprint server for subjects relating to AI, machine learning and deep learning \u2013 from disciplines including statistics, mathematics and computer science \u2013 and provide you with a useful \u201cbest of\u201d list for the\u2026","rel":"","context":"In &quot;AI Deep Learning&quot;","img":{"alt_text":"","src":"https:\/\/i0.wp.com\/insidebigdata.com\/wp-content\/uploads\/2013\/12\/arxiv.jpg?resize=350%2C200&ssl=1","width":350,"height":200},"classes":[]},{"id":22263,"url":"https:\/\/insidebigdata.com\/2019\/03\/15\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-february-2019\/","url_meta":{"origin":27358,"position":3},"title":"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 February 2019","date":"March 15, 2019","format":false,"excerpt":"In this recurring monthly feature, we will filter all the recent research papers appearing in the arXiv.org preprint server for subjects relating to AI, machine learning and deep learning \u2013 from disciplines including statistics, mathematics and computer science \u2013 and provide you with a useful \u201cbest of\u201d list for the\u2026","rel":"","context":"In &quot;AI Deep Learning&quot;","img":{"alt_text":"","src":"https:\/\/i0.wp.com\/insidebigdata.com\/wp-content\/uploads\/2013\/12\/arxiv.jpg?resize=350%2C200&ssl=1","width":350,"height":200},"classes":[]},{"id":21252,"url":"https:\/\/insidebigdata.com\/2018\/10\/17\/best-arxiv-org-ai-machine-learning-deep-learning-september-2018\/","url_meta":{"origin":27358,"position":4},"title":"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 September 2018","date":"October 17, 2018","format":false,"excerpt":"In this recurring monthly feature, we will filter all the recent research papers appearing in the arXiv.org preprint server for subjects relating to AI, machine learning and deep learning \u2013 from disciplines including statistics, mathematics and computer science \u2013 and provide you with a useful \u201cbest of\u201d list for the\u2026","rel":"","context":"In &quot;AI Deep Learning&quot;","img":{"alt_text":"","src":"https:\/\/i0.wp.com\/insidebigdata.com\/wp-content\/uploads\/2013\/12\/arxiv.jpg?resize=350%2C200&ssl=1","width":350,"height":200},"classes":[]},{"id":20910,"url":"https:\/\/insidebigdata.com\/2018\/08\/13\/best-arxiv-org-ai-machine-learning-deep-learning-july-2018\/","url_meta":{"origin":27358,"position":5},"title":"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 July 2018","date":"August 13, 2018","format":false,"excerpt":"In this recurring monthly feature, we will filter all the recent research papers appearing in the arXiv.org preprint server for subjects relating to AI, machine learning and deep learning \u2013 from disciplines including statistics, mathematics and computer science \u2013 and provide you with a useful \u201cbest of\u201d list for the\u2026","rel":"","context":"In &quot;AI Deep Learning&quot;","img":{"alt_text":"","src":"https:\/\/i0.wp.com\/insidebigdata.com\/wp-content\/uploads\/2013\/12\/arxiv.jpg?resize=350%2C200&ssl=1","width":350,"height":200},"classes":[]}],"_links":{"self":[{"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/posts\/27358"}],"collection":[{"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/users\/37"}],"replies":[{"embeddable":true,"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/comments?post=27358"}],"version-history":[{"count":0,"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/posts\/27358\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/media\/6361"}],"wp:attachment":[{"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/media?parent=27358"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/categories?post=27358"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/tags?post=27358"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}