{"id":22319,"date":"2019-03-21T06:30:44","date_gmt":"2019-03-21T13:30:44","guid":{"rendered":"https:\/\/insidebigdata.com\/?p=22319"},"modified":"2019-03-22T08:58:09","modified_gmt":"2019-03-22T15:58:09","slug":"scaling-production-ai-enterprises","status":"publish","type":"post","link":"https:\/\/insidebigdata.com\/2019\/03\/21\/scaling-production-ai-enterprises\/","title":{"rendered":"Scaling Production AI"},"content":{"rendered":"<p><em>Barbara Murphy, VP of Marketing, WekaIO, explores how as as AI production models grow larger and more intricate, server architecture can get more and more complex. Explore how tools like GPUs and more are moving the dial on AI.\u00a0<\/em><\/p>\n<p>The use of machine learning (ML) to derive insights, automate repetitive tasks or to enhance human decision making has reached exponential growth levels. The technology is being utilized broadly to detect fraud, automate driving tasks, scan medical images, predict manufacturing failures, enhance financial trading strategies, and many more.<\/p>\n<div id=\"attachment_22248\" style=\"width: 327px\" class=\"wp-caption alignright\"><img aria-describedby=\"caption-attachment-22248\" decoding=\"async\" loading=\"lazy\" class=\" wp-image-22248\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2019\/03\/Barbara-Murphy-hi-res-e1551998824321.jpg\" alt=\"production AI\" width=\"317\" height=\"239\" \/><p id=\"caption-attachment-22248\" class=\"wp-caption-text\"><em>Barbara Murphy, VP of Marketing, WekaIO<\/em><\/p><\/div>\n<p>While ML is still considered a relatively nascent technology, across industries it is a race to be the first across the line, whether the first to showcase a fully autonomous vehicle or successfully develop automated healthcare systems. Time to market is measured in how fast machine learning models can be trained, and how big the data set that is driving the training model. Put simply this is a balance of two key factors, data and compute. More data equates to more accurate models and more compute equates to faster training.<\/p>\n<p>It is not uncommon for model development to start out with a simple infrastructure model, a single GPU server with a single view to the data set. In this case, the data set is often small enough to fit in the locally attached storage inside the GPU server providing great performance and fast completion of an epoch (the time it takes to run a single machine learning model training run).<\/p>\n<div id=\"attachment_22322\" style=\"width: 184px\" class=\"wp-caption alignright\"><img aria-describedby=\"caption-attachment-22322\" decoding=\"async\" loading=\"lazy\" class=\"wp-image-22322\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2019\/03\/Simlified-dataflow.jpg\" alt=\"production AI\" width=\"174\" height=\"69\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2019\/03\/Simlified-dataflow.jpg 116w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2019\/03\/Simlified-dataflow-110x45.jpg 110w\" sizes=\"(max-width: 174px) 100vw, 174px\" \/><p id=\"caption-attachment-22322\" class=\"wp-caption-text\"><em>Figure 1: Simplified data flow from storage to GPU server (Graphic: WekaIO)<\/em><\/p><\/div>\n<p>As AI production models grow larger and more complex, it requires a server architecture that looks much like high performance computing (HPC), with workloads scaled across many servers and distributed processing across the server infrastructure. Figure 2 provides an outline of a clustered GPU server environment, in this distributed environment the data flow becomes significantly more complex, with many different servers requiring access to the same data set at the same time, and the actual flow of data could be mixed between multiple servers. However, it is critical that a single view of the data set is preserved across the clustered server infrastructure.<\/p>\n<div id=\"attachment_22323\" style=\"width: 125px\" class=\"wp-caption alignleft\"><img aria-describedby=\"caption-attachment-22323\" decoding=\"async\" loading=\"lazy\" class=\"wp-image-22323\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2019\/03\/deep-learning-model.jpg\" alt=\"production AI\" width=\"115\" height=\"249\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2019\/03\/deep-learning-model.jpg 102w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2019\/03\/deep-learning-model-69x150.jpg 69w\" sizes=\"(max-width: 115px) 100vw, 115px\" \/><p id=\"caption-attachment-22323\" class=\"wp-caption-text\"><em>Figure 2. Deep learning model, single data set serving multiple GPU servers. (Graphic: WekaIO)<\/em><\/p><\/div>\n<p><strong><u>The Challenge of Scale<\/u><\/strong><\/p>\n<p>For complex tasks like <a href=\"https:\/\/insidebigdata.com\/?s=autonomous+driving\" target=\"_blank\" rel=\"noopener\">autonomous driving<\/a>, it is not uncommon for data sets to grow to tens or hundreds of petabytes for a model. With data sets this large, it requires a storage system that can serve up a single view of the data to all servers in the cluster, but the sheer size of the data set precludes the use of locally attached storage. The architecture has to rely on networked storage to deliver the scale of data required to train the models.<\/p>\n<p>Scaling a large data set to single machine is relatively straightforward, the challenge is how to deliver comparable performance to a complex data flow described in figure 2. All servers need to ingest data from a single source, placing immense pressure on the shared storage system. In the real world, performance generally deteriorates as the workload scales largely due to the bottleneck created from the data preprocessing and network latencies.<\/p>\n<p>Infrastructure choices have a significant impact on the performance and scalability of a DL workflow. Model complexity, catalog data size, and input type (such as images and text) will impact key elements of a solution, including the number of GPUs, servers, network interconnects, and storage type (local disk or shared). The more complex the environment, the greater the need to balance components. HPE has integrated best-in-class components into its server infrastructure for DL, including GPUs from NVIDIA, 100 Gbps InfiniBand networking from Mellanox, and the Matrix high-performance shared file system software from WekaIO.<\/p>\n<p><strong><u>Benchmarking Performance at Scale<\/u><\/strong><\/p>\n<p>HPE set out to investigate the storage impact of scaling data training from a single GPU server to a clustered multi-node environment utilizing WekaIO Matrix shared file system. The purpose of the benchmark was to measure the impact (performance degradation) the external shared storage system would have when scaled from a single GPU server to a multi-node environment.<\/p>\n<p>The team leveraged the HPE Apollo 6500 Gen10 GPU servers, each with 8 NVIDIA Tesla V100 SXM2 16GB GPUs per server. The performance benchmark scaled from a single Apollo 6500 GPU server to four, for a total of 32 GPU complex.<\/p>\n<p>The external storage system leveraged the HPE ProLiant DL360 servers, interconnected with Mellanox 100Gbps EDR networking, running WekaIO Matrix flash-optimized software. Independent I\/O benchmarks had previously verified that this storage solution is capable of delivering over 30GBytes\/second of sequential 1MB reads and over 2.5 million IOPS for small 4K random reads. The infrastructure is capable of scaling to hundreds of storage nodes and tens of petabytes in a single namespace, more than sufficient to handle the even the largest DL training data sets.<\/p>\n<div id=\"attachment_22324\" style=\"width: 441px\" class=\"wp-caption aligncenter\"><img aria-describedby=\"caption-attachment-22324\" decoding=\"async\" loading=\"lazy\" class=\"wp-image-22324 size-full\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2019\/03\/deep-learning-benchmark.jpg\" alt=\"production AI\" width=\"431\" height=\"425\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2019\/03\/deep-learning-benchmark.jpg 431w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2019\/03\/deep-learning-benchmark-150x148.jpg 150w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2019\/03\/deep-learning-benchmark-300x296.jpg 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2019\/03\/deep-learning-benchmark-50x50.jpg 50w\" sizes=\"(max-width: 431px) 100vw, 431px\" \/><p id=\"caption-attachment-22324\" class=\"wp-caption-text\"><em>Figure 3. Deep Learning benchmark environment leveraging HPE reference architecture<\/em><\/p><\/div>\n<p>The performance tests were conducted using the popular ImageNet data set stored in a standard Record format across a variety of popular benchmarks. The complete list is outlined in table 1, along with the batch size used. The same batch sizes were used for both training and inference.<\/p>\n<table>\n<tbody>\n<tr>\n<td>Benchmark<\/td>\n<td>ResNet152<\/td>\n<td>ResNet50<\/td>\n<td>Google Net<\/td>\n<td>VGG16<\/td>\n<td>Inception-V4<\/td>\n<\/tr>\n<tr>\n<td>Batch Size<\/td>\n<td>128<\/td>\n<td>256<\/td>\n<td>128<\/td>\n<td>128<\/td>\n<td>128<\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<p><em>Table 1. Training benchmarks conducted and batch size used.<\/em><\/p>\n<p><span style=\"font-family: -apple-system, BlinkMacSystemFont, 'Segoe UI', Roboto, Oxygen-Sans, Ubuntu, Cantarell, 'Helvetica Neue', sans-serif;\">The training data results, outlined in figure 4, clearly show that HPE Apollo 6500 Systems with varying quantities of NVIDIA SXM2 16 GB Tesla V100 GPUs scale near-linear performance across a variety of benchmark tests. The testing results show that using one server with eight GPUs delivers comparable performance to using two servers with four GPUs each, or four servers with two GPUs each. In other words, the storage system can easily handle a scale-out as well as a scale-up deployment.<\/span><\/p>\n<div id=\"attachment_22325\" style=\"width: 405px\" class=\"wp-caption aligncenter\"><img aria-describedby=\"caption-attachment-22325\" decoding=\"async\" loading=\"lazy\" class=\"wp-image-22325 size-full\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2019\/03\/scaling-performance-from-GPU.jpg\" alt=\"production AI\" width=\"395\" height=\"232\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2019\/03\/scaling-performance-from-GPU.jpg 395w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2019\/03\/scaling-performance-from-GPU-150x88.jpg 150w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2019\/03\/scaling-performance-from-GPU-300x176.jpg 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2019\/03\/scaling-performance-from-GPU-280x165.jpg 280w\" sizes=\"(max-width: 395px) 100vw, 395px\" \/><p id=\"caption-attachment-22325\" class=\"wp-caption-text\"><em>Figure 4. Scaling performance from a single GPU to 32 GPUs, across 4 Apollo 6500 Systems<\/em><\/p><\/div>\n<p>Worth noting is the ease of use of a configuration leveraging WekaIO Matrix. Due to the removed dependence on data locality, it is very easy to add new Apollo 6500 Systems to a test environment using a shared filesystem architecture. No additional data copying is needed to prepare additional clients, which helps to create an agile and flexible architecture for such a rapidly growing use case. As compute requirements increase, WekaIO Matrix and Mellanox networking provide excellent performance and ease of scaling.<\/p>\n<p><em>For full details on the testing completed by HPE, <a href=\"https:\/\/www.weka.io\/promo\/hpe-accelerate-performance-production-ai-tech-white-paper-mar-2019\" target=\"_blank\" rel=\"noopener\">see here<\/a>, and download the complete white paper.<\/em><\/p>\n<p><em>Get the first whitepaper in this series from HPE, WekaIO, NVIDIA and Mellanox at\u00a0<a href=\"https:\/\/www.weka.io\/promo\/hpe-ai-tech-white-paper-oct-2018\/\" target=\"_blank\" rel=\"noopener\">Time to Value and AI Insights.<\/a><\/em><\/p>\n<p><em>Get more information on WekaIO at <a href=\"http:\/\/www.weka.io\" target=\"_blank\" rel=\"noopener\">www.weka.io,\u00a0<\/a>or find out more details on the solutions from HPE and WekaIO <a href=\"https:\/\/www.weka.io\/solutions\/hpe\/\" target=\"_blank\" rel=\"noopener\">here.<\/a><\/em><\/p>\n","protected":false},"excerpt":{"rendered":"<p>As AI models grow larger and more complex, it requires a server architecture that looks much like high performance computing (HPC), with workloads scaled across many servers and distributed processing across the server infrastructure. Barbara Murphy, VP of Marketing, WekaIO, explores how as AI production models grow larger and more intricate,\u00a0server architecture gets more complex. Explore how tools like GPU clusters and more are moving the dial forward on AI.\u00a0<\/p>\n","protected":false},"author":10513,"featured_media":22248,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"jetpack_post_was_ever_published":false,"footnotes":""},"categories":[71,87,180,67,56,57,1],"tags":[437,538,95,731],"acf":[],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v20.6 - https:\/\/yoast.com\/wordpress\/plugins\/seo\/ -->\n<title>Production AI: A Guide to Scaling for Today&#039;s Enterprises<\/title>\n<meta name=\"description\" content=\"WekaIO&#039;s Barbara Murphy explores how as AI production models grow larger and more intricate,\u00a0 server architecture gets more complex to enable scalability.\" \/>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/insidebigdata.com\/2019\/03\/21\/scaling-production-ai-enterprises\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Production AI: A Guide to Scaling for Today&#039;s Enterprises\" \/>\n<meta property=\"og:description\" content=\"WekaIO&#039;s Barbara Murphy explores how as AI production models grow larger and more intricate,\u00a0 server architecture gets more complex to enable scalability.\" \/>\n<meta property=\"og:url\" content=\"https:\/\/insidebigdata.com\/2019\/03\/21\/scaling-production-ai-enterprises\/\" \/>\n<meta property=\"og:site_name\" content=\"insideBIGDATA\" \/>\n<meta property=\"article:publisher\" content=\"http:\/\/www.facebook.com\/insidebigdata\" \/>\n<meta property=\"article:published_time\" content=\"2019-03-21T13:30:44+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2019-03-22T15:58:09+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2019\/03\/Barbara-Murphy-hi-res-e1551998824321.jpg\" \/>\n\t<meta property=\"og:image:width\" content=\"400\" \/>\n\t<meta property=\"og:image:height\" content=\"300\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/jpeg\" \/>\n<meta name=\"author\" content=\"Editorial Team\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:creator\" content=\"@insideBigData\" \/>\n<meta name=\"twitter:site\" content=\"@insideBigData\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"Editorial Team\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"6 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"WebPage\",\"@id\":\"https:\/\/insidebigdata.com\/2019\/03\/21\/scaling-production-ai-enterprises\/\",\"url\":\"https:\/\/insidebigdata.com\/2019\/03\/21\/scaling-production-ai-enterprises\/\",\"name\":\"Production AI: A Guide to Scaling for Today's Enterprises\",\"isPartOf\":{\"@id\":\"https:\/\/insidebigdata.com\/#website\"},\"datePublished\":\"2019-03-21T13:30:44+00:00\",\"dateModified\":\"2019-03-22T15:58:09+00:00\",\"author\":{\"@id\":\"https:\/\/insidebigdata.com\/#\/schema\/person\/2949e412c144601cdbcc803bd234e1b9\"},\"description\":\"WekaIO's Barbara Murphy explores how as AI production models grow larger and more intricate,\u00a0 server architecture gets more complex to enable scalability.\",\"breadcrumb\":{\"@id\":\"https:\/\/insidebigdata.com\/2019\/03\/21\/scaling-production-ai-enterprises\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/insidebigdata.com\/2019\/03\/21\/scaling-production-ai-enterprises\/\"]}]},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/insidebigdata.com\/2019\/03\/21\/scaling-production-ai-enterprises\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\/\/insidebigdata.com\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Scaling Production AI\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/insidebigdata.com\/#website\",\"url\":\"https:\/\/insidebigdata.com\/\",\"name\":\"insideBIGDATA\",\"description\":\"Your Source for AI, Data Science, Deep Learning &amp; Machine Learning Strategies\",\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/insidebigdata.com\/?s={search_term_string}\"},\"query-input\":\"required name=search_term_string\"}],\"inLanguage\":\"en-US\"},{\"@type\":\"Person\",\"@id\":\"https:\/\/insidebigdata.com\/#\/schema\/person\/2949e412c144601cdbcc803bd234e1b9\",\"name\":\"Editorial Team\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/insidebigdata.com\/#\/schema\/person\/image\/\",\"url\":\"https:\/\/secure.gravatar.com\/avatar\/e137ce7ea40e38bd4d25bb7860cfe3e4?s=96&d=mm&r=g\",\"contentUrl\":\"https:\/\/secure.gravatar.com\/avatar\/e137ce7ea40e38bd4d25bb7860cfe3e4?s=96&d=mm&r=g\",\"caption\":\"Editorial Team\"},\"sameAs\":[\"http:\/\/www.insidebigdata.com\"],\"url\":\"https:\/\/insidebigdata.com\/author\/editorial\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Production AI: A Guide to Scaling for Today's Enterprises","description":"WekaIO's Barbara Murphy explores how as AI production models grow larger and more intricate,\u00a0 server architecture gets more complex to enable scalability.","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/insidebigdata.com\/2019\/03\/21\/scaling-production-ai-enterprises\/","og_locale":"en_US","og_type":"article","og_title":"Production AI: A Guide to Scaling for Today's Enterprises","og_description":"WekaIO's Barbara Murphy explores how as AI production models grow larger and more intricate,\u00a0 server architecture gets more complex to enable scalability.","og_url":"https:\/\/insidebigdata.com\/2019\/03\/21\/scaling-production-ai-enterprises\/","og_site_name":"insideBIGDATA","article_publisher":"http:\/\/www.facebook.com\/insidebigdata","article_published_time":"2019-03-21T13:30:44+00:00","article_modified_time":"2019-03-22T15:58:09+00:00","og_image":[{"width":400,"height":300,"url":"https:\/\/insidebigdata.com\/wp-content\/uploads\/2019\/03\/Barbara-Murphy-hi-res-e1551998824321.jpg","type":"image\/jpeg"}],"author":"Editorial Team","twitter_card":"summary_large_image","twitter_creator":"@insideBigData","twitter_site":"@insideBigData","twitter_misc":{"Written by":"Editorial Team","Est. reading time":"6 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"WebPage","@id":"https:\/\/insidebigdata.com\/2019\/03\/21\/scaling-production-ai-enterprises\/","url":"https:\/\/insidebigdata.com\/2019\/03\/21\/scaling-production-ai-enterprises\/","name":"Production AI: A Guide to Scaling for Today's Enterprises","isPartOf":{"@id":"https:\/\/insidebigdata.com\/#website"},"datePublished":"2019-03-21T13:30:44+00:00","dateModified":"2019-03-22T15:58:09+00:00","author":{"@id":"https:\/\/insidebigdata.com\/#\/schema\/person\/2949e412c144601cdbcc803bd234e1b9"},"description":"WekaIO's Barbara Murphy explores how as AI production models grow larger and more intricate,\u00a0 server architecture gets more complex to enable scalability.","breadcrumb":{"@id":"https:\/\/insidebigdata.com\/2019\/03\/21\/scaling-production-ai-enterprises\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/insidebigdata.com\/2019\/03\/21\/scaling-production-ai-enterprises\/"]}]},{"@type":"BreadcrumbList","@id":"https:\/\/insidebigdata.com\/2019\/03\/21\/scaling-production-ai-enterprises\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/insidebigdata.com\/"},{"@type":"ListItem","position":2,"name":"Scaling Production AI"}]},{"@type":"WebSite","@id":"https:\/\/insidebigdata.com\/#website","url":"https:\/\/insidebigdata.com\/","name":"insideBIGDATA","description":"Your Source for AI, Data Science, Deep Learning &amp; Machine Learning Strategies","potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/insidebigdata.com\/?s={search_term_string}"},"query-input":"required name=search_term_string"}],"inLanguage":"en-US"},{"@type":"Person","@id":"https:\/\/insidebigdata.com\/#\/schema\/person\/2949e412c144601cdbcc803bd234e1b9","name":"Editorial Team","image":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/insidebigdata.com\/#\/schema\/person\/image\/","url":"https:\/\/secure.gravatar.com\/avatar\/e137ce7ea40e38bd4d25bb7860cfe3e4?s=96&d=mm&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/e137ce7ea40e38bd4d25bb7860cfe3e4?s=96&d=mm&r=g","caption":"Editorial Team"},"sameAs":["http:\/\/www.insidebigdata.com"],"url":"https:\/\/insidebigdata.com\/author\/editorial\/"}]}},"jetpack_featured_media_url":"https:\/\/insidebigdata.com\/wp-content\/uploads\/2019\/03\/Barbara-Murphy-hi-res-e1551998824321.jpg","jetpack_shortlink":"https:\/\/wp.me\/p9eA3j-5NZ","jetpack-related-posts":[{"id":22246,"url":"https:\/\/insidebigdata.com\/2019\/03\/14\/ai-goes-mainstream\/","url_meta":{"origin":22319,"position":0},"title":"AI Goes Mainstream","date":"March 14, 2019","format":false,"excerpt":"According to a recent Gartner survey, Artificial intelligence (AI) learning has moved from a specialized field into mainstream business use with 37 percent of respondents reporting their enterprises either had deployed AI or would do so shortly. WekaIO's Barbara Murphy explores the path of artificial intelligence from the fringe to\u2026","rel":"","context":"In &quot;Enterprise&quot;","img":{"alt_text":"production AI","src":"https:\/\/i0.wp.com\/insidebigdata.com\/wp-content\/uploads\/2019\/03\/Barbara-Murphy-hi-res-e1551998824321.jpg?resize=350%2C200&ssl=1","width":350,"height":200},"classes":[]},{"id":23025,"url":"https:\/\/insidebigdata.com\/2019\/08\/08\/infrastructure-machine-learning-workloads\/","url_meta":{"origin":22319,"position":1},"title":"Why You Need a Modern Infrastructure to Accelerate AI and ML Workloads","date":"August 8, 2019","format":false,"excerpt":"Recent years have seen a boom in the generation of data from a variety of sources: connected devices, IoT, analytics, healthcare, smartphones, and much more. This data management problem is particularly acute in the areas of Artificial Intelligence (AI) and Machine Learning (ML) workloads. This guest article from WekaIO highlights\u2026","rel":"","context":"In &quot;AI Deep Learning&quot;","img":{"alt_text":"","src":"https:\/\/i0.wp.com\/insidebigdata.com\/wp-content\/uploads\/2019\/07\/shutterstock_561931702.jpg?resize=350%2C200&ssl=1","width":350,"height":200},"classes":[]},{"id":22292,"url":"https:\/\/insidebigdata.com\/2019\/03\/18\/ai-critical-measures-time-value-insights\/","url_meta":{"origin":22319,"position":2},"title":"AI Critical Measures: Time to Value and Insights","date":"March 18, 2019","format":false,"excerpt":"AI is a game changer for industries today but achieving AI success contains two critical factors to consider \u2014 time to value and time to insights.\u00a0 Time to value is the metric that looks at the time it takes to realize the value of a product, solution or offering. Time\u2026","rel":"","context":"In &quot;Enterprise&quot;","img":{"alt_text":"","src":"https:\/\/i0.wp.com\/insidebigdata.com\/wp-content\/uploads\/2019\/03\/Charla-e1552670361849.jpg?resize=350%2C200&ssl=1","width":350,"height":200},"classes":[]},{"id":25267,"url":"https:\/\/insidebigdata.com\/2020\/11\/19\/penguin-computing-announces-originai-powered-by-wekaio\/","url_meta":{"origin":22319,"position":3},"title":"Penguin Computing Announces OriginAI Powered by WekaIO","date":"November 19, 2020","format":false,"excerpt":"Penguin Computing, a division of SMART Global Holdings, Inc. (NASDAQ: SGH) and a leader in high-performance computing (HPC), artificial intelligence (AI), and enterprise data center solutions, announced that it has partnered with WekaIO\u2122 (Weka) to provide NVIDIA GPU-Powered OriginAI, a comprehensive, end-to-end solution for data center AI that maximizes the\u2026","rel":"","context":"In &quot;AI Deep Learning&quot;","img":{"alt_text":"","src":"","width":0,"height":0},"classes":[]},{"id":21331,"url":"https:\/\/insidebigdata.com\/2018\/10\/25\/time-value-ai-deep-learning-insights\/","url_meta":{"origin":22319,"position":4},"title":"Accelerate Time to Value and AI Insights","date":"October 25, 2018","format":false,"excerpt":"In this edition of Industry Perspectives, HPE explores how reducing the cycle time for inferencing helps to accelerate time to market for deep learning and AI insights and solutions.\u00a0","rel":"","context":"In &quot;AI Deep Learning&quot;","img":{"alt_text":"deep learning","src":"https:\/\/i0.wp.com\/insidebigdata.com\/wp-content\/uploads\/2018\/10\/shutterstock_1096541144.jpg?resize=350%2C200&ssl=1","width":350,"height":200},"classes":[]},{"id":22491,"url":"https:\/\/insidebigdata.com\/2019\/04\/16\/prepare-for-production-ai-with-the-hpe-ai-data-node\/","url_meta":{"origin":22319,"position":5},"title":"Prepare for Production AI with the HPE AI Data Node","date":"April 16, 2019","format":false,"excerpt":"https:\/\/www.youtube.com\/watch?v=xDJvLCwgASA In this video from GTC 2019 in San Jose, Harvey Skinner, Distinguished Technologist, discusses the advent of production AI and how the HPE AI Data Node offers a building block for AI storage. Commercial enterprises have been investigating and exploring how AI can improve their business. Now they\u2019re ready\u2026","rel":"","context":"In &quot;AI Deep Learning&quot;","img":{"alt_text":"","src":"https:\/\/i0.wp.com\/img.youtube.com\/vi\/xDJvLCwgASA\/0.jpg?resize=350%2C200","width":350,"height":200},"classes":[]}],"_links":{"self":[{"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/posts\/22319"}],"collection":[{"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/users\/10513"}],"replies":[{"embeddable":true,"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/comments?post=22319"}],"version-history":[{"count":0,"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/posts\/22319\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/media\/22248"}],"wp:attachment":[{"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/media?parent=22319"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/categories?post=22319"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/tags?post=22319"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}