{"id":34133,"date":"2023-12-19T02:59:00","date_gmt":"2023-12-19T10:59:00","guid":{"rendered":"https:\/\/insidebigdata.com\/?p=34133"},"modified":"2023-12-20T20:28:10","modified_gmt":"2023-12-21T04:28:10","slug":"insidebigdata-ai-news-briefs-bulletin-board","status":"publish","type":"post","link":"https:\/\/insidebigdata.com\/2023\/12\/19\/insidebigdata-ai-news-briefs-bulletin-board\/","title":{"rendered":"insideBIGDATA AI News Briefs Bulletin Board"},"content":{"rendered":"\n<p>Welcome insideBIGDATA AI News Briefs Bulletin Board, our timely new feature bringing you the latest industry insights and perspectives surrounding the field of AI including deep learning, large language models, generative AI, and transformers. We\u2019re working tirelessly to dig up the most timely and curious tidbits underlying the day\u2019s most popular technologies. We know this field is advancing rapidly and we want to bring you a regular resource to keep you informed and state-of-the-art. The news bites are constantly being added in reverse date order (most recent on top). With our bulletin board you can check back often to see what&#8217;s happening in our rapidly accelerating industry. Click <a href=\"https:\/\/insidebigdata.com\/category\/channels\/ai-news-briefs\/\" target=\"_blank\" rel=\"noreferrer noopener\">HERE<\/a> to check out previous &#8220;AI News Briefs&#8221; round-ups. <\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n\n<p>[12\/20\/2023] OpenAI is <a href=\"https:\/\/techcrunch.com\/2023\/12\/18\/openai-buffs-safety-team-and-gives-board-veto-power-on-risky-ai\/\" target=\"_blank\" rel=\"noreferrer noopener\">expanding its internal safety processes<\/a> to fend off the threat of harmful AI. A new \u201csafety advisory group\u201d will sit above the technical teams and make recommendations to leadership, and the board has been granted veto power \u2014 of course, whether it will actually use it is another question entirely.<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n\n<p>[12\/19\/2023] Salesforce <a href=\"https:\/\/www.emergingtechbrew.com\/stories\/2023\/12\/15\/salesforce-ai-copilot-race\" target=\"_blank\" rel=\"noreferrer noopener\">announced<\/a>&nbsp;it will start making available Einstein Copilot, its new all-purpose conversational AI assistant for the office. Salesforce aims to enhance its users\u2019 customer service options and offer a new level of efficiency to sales and marketing efforts via a new interface and customizable AI models. AI-powered CRM systems like this will enable Salesforce clients to create more personalized and effective interactions without extensive work developing their own AI models. Salesforce is expecting this product will help streamline the integration of one of its key advantages: the silos of enterprise data its clients maintain on the platform. Ultimately, every so-called data aggregator like Salesforce will need to adapt to this trend to compete with hyperscalers like Microsoft and Google.<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n<div class=\"wp-block-image\">\n<figure class=\"alignright size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"241\" height=\"169\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/OpenAI_logo.png\" alt=\"\" class=\"wp-image-33984\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/OpenAI_logo.png 241w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/OpenAI_logo-150x105.png 150w\" sizes=\"(max-width: 241px) 100vw, 241px\" \/><\/figure><\/div>\n\n\n<p>[12\/19\/2023] OpenAI <a href=\"https:\/\/platform.openai.com\/docs\/guides\/prompt-engineering\" target=\"_blank\" rel=\"noreferrer noopener\">Guide to Prompt Engineering<\/a>: The guide shares 6 strategies and tactics for getting better results from LLMs like GPT-4. The methods described can sometimes be deployed in combination for greater effect. The company encourages experimentation to find the methods that work best for you. Some of the examples demonstrated in the guide currently work only with the most capable model,&nbsp;<code>gpt-4<\/code>. In general, if you find that a model fails at a task and a more capable model is available, it&#8217;s often worth trying again with the more capable model.<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n<div class=\"wp-block-image\">\n<figure class=\"alignleft size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"295\" height=\"76\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/DeepMind_logo.png\" alt=\"\" class=\"wp-image-34236\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/DeepMind_logo.png 295w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/DeepMind_logo-150x39.png 150w\" sizes=\"(max-width: 295px) 100vw, 295px\" \/><\/figure><\/div>\n\n\n<p>[12\/17\/2023] NEW from Google DeepMind! <a href=\"https:\/\/deepmind.google\/discover\/blog\/funsearch-making-new-discoveries-in-mathematical-sciences-using-large-language-models\/\" target=\"_blank\" rel=\"noreferrer noopener\">FunSearch: Making new discoveries in mathematical sciences using Large Language Model<\/a> &#8211; Google DeepMind used FunSearch,  a unique LLM framework, to solve a famous unsolved problem in pure mathematics &#8211; the cap set problem using a unique pairing of a pre-trained LLM with an automated evaluator for iterative solution development. FunSearch is built on a modified version of Google&#8217;s PaLM 2, termed Codey, optimized for code generation. It fills in missing solution components in a Python-sketched problem. The cap set problem involves finding the largest set of points in a high-dimensional grid where no three points align linearly. It&#8217;s a complex issue representing a broader class of problems in extremal combinatorics. The approach involved pairing a pre-trained LLM with an evaluator for program generation and refinement. FunSearch first generates a range of potential solutions. The evaluator then rigorously filters these solutions, retaining only the most accurate and viable ones. This process iteratively refines the solutions, enhancing their reliability and applicability. FunSearch outperformed existing computational methods. Access the research paper <a href=\"https:\/\/storage.googleapis.com\/deepmind-media\/DeepMind.com\/Blog\/funsearch-making-new-discoveries-in-mathematical-sciences-using-large-language-models\/Mathematical-discoveries-from-program-search-with-large-language-models.pdf\" target=\"_blank\" rel=\"noreferrer noopener\">HERE<\/a>. <\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n<div class=\"wp-block-image\">\n<figure class=\"alignright size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"283\" height=\"172\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/LLM360_logo.png\" alt=\"\" class=\"wp-image-34232\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/LLM360_logo.png 283w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/LLM360_logo-150x91.png 150w\" sizes=\"(max-width: 283px) 100vw, 283px\" \/><\/figure><\/div>\n\n\n<p>[12\/17\/2023] <a href=\"https:\/\/www.llm360.ai\/\" target=\"_blank\" rel=\"noreferrer noopener\">LLM360<\/a> introduced <a href=\"https:\/\/www.llm360.ai\/blog\/introducing-llm360-fully-transparent-open-source-llms.html\" target=\"_blank\" rel=\"noreferrer noopener\">Amber-7B and Crystalcoder-7B LLMs<\/a>, offering full transparency in Large Language Model (LLM) training with comprehensive open-source release including training data and code. Existing LLMs lack transparency in training processes, limiting the AI community&#8217;s ability to assess reliability, biases, and replicability. This obscurity hinders collaborative progress and thorough understanding of LLM behaviors. LLM360 tackles this by releasing two models with all training components &#8211; including 1.3T and 1.4T token datasets, training code, intermediate checkpoints, and detailed logs. Training employs AdamW optimizer, mixed-precision techniques, and thorough data mix analysis for nuanced pre-training. Amber-7B and CrystalCoder-7B demonstrate robust performance on benchmarks like ARC and MMLU. Specifics include training on diverse datasets (e.g., RefinedWeb, StarCoder), achieving 582.4k tokens per second throughput, and detailed analysis of model behaviors like memorization across training stages. Download the research paper: &#8220;<a href=\"https:\/\/arxiv.org\/abs\/2312.06550\" target=\"_blank\" rel=\"noreferrer noopener\">LLM360: Towards Fully Transparent Open-Source LLMs<\/a>.&#8221; LLM 360, a collaboration between Petuum, MBZUAI and Cerebras, is dedicated to advancing the field of AI by providing comprehensive access to large language models.<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n\n<p>[12\/17\/2023] COOL PAPER ALERT! <strong>Weight subcloning: direct initialization of transformers using larger pretrained ones <\/strong><\/p>\n\n\n\n<p><strong>Paper page: <\/strong><a href=\"https:\/\/huggingface.co\/papers\/2312.09299\" target=\"_blank\" rel=\"noreferrer noopener\">https:\/\/huggingface.co\/papers\/2312.09299<\/a><\/p>\n\n\n\n<p><strong>Abstract: <\/strong>Training large transformer models from scratch for a target task requires lots of data and is computationally demanding. The usual practice of transfer learning overcomes this challenge by initializing the model with weights of a pretrained model of the same size and specification to increase the convergence and training speed. However, what if no pretrained model of the required size is available? In this paper, we introduce a simple yet effective technique to transfer the knowledge of a pretrained model to smaller variants. Our approach called weight subcloning expedites the training of scaled-down transformers by initializing their weights from larger pretrained models. Weight subcloning involves an operation on the pretrained model to obtain the equivalent initialized scaled-down model. It consists of two key steps: first, we introduce neuron importance ranking to decrease the embedding dimension per layer in the pretrained model. Then, we remove blocks from the transformer model to match the number of layers in the scaled-down network. The result is a network ready to undergo training, which gains significant improvements in training speed compared to random initialization. For instance, we achieve 4x faster training for vision transformers in image classification and language models designed for next token prediction.<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n<div class=\"wp-block-image\">\n<figure class=\"alignleft size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"300\" height=\"147\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/Google_GeminiPro_logo.png\" alt=\"\" class=\"wp-image-34229\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/Google_GeminiPro_logo.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/Google_GeminiPro_logo-150x74.png 150w\" sizes=\"(max-width: 300px) 100vw, 300px\" \/><\/figure><\/div>\n\n\n<p>[12\/16\/2023] Google announced <a href=\"https:\/\/blog.google\/technology\/ai\/gemini-api-developers-cloud\/\" target=\"_blank\" rel=\"noreferrer noopener\">Gemini Pro<\/a> developer access &#8211; its advanced artificial intelligence program, Gemini, which was unveiled last week, will now be available in a preview version for users of its AI Studio programming tool and Vertex AI, a fully managed programming tool for enterprises running on Google Cloud. Gemini will also be integrated into Duet AI, Google&#8217;s AI-enhanced coding tool, in the coming weeks. The announcement highlighted Gemini&#8217;s training on Google&#8217;s custom AI chip, the Tensor Processing Unit (TPU), and the release of TPU v5p, offering four times the performance of the existing v4 chips.<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n<div class=\"wp-block-image\">\n<figure class=\"alignleft size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"241\" height=\"169\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/OpenAI_logo.png\" alt=\"\" class=\"wp-image-33984\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/OpenAI_logo.png 241w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/OpenAI_logo-150x105.png 150w\" sizes=\"(max-width: 241px) 100vw, 241px\" \/><\/figure><\/div>\n\n\n<p>[12\/16\/2023] OpenAI Partnership with <a href=\"https:\/\/openai.com\/blog\/axel-springer-partnership\" target=\"_blank\" rel=\"noreferrer noopener\">Axel Springer<\/a> to deepen beneficial use of AI in journalism. The detail allows ChatGPT to summarize its current articles, including gated content. The deal is not exclusive and spans several years, enabling ChatGPT users worldwide to access summaries of global news from Axel Springer&#8217;s brands, with links to full articles for transparency purposes. The alignment contrasts with other media companies like CNN, the New York Times, and Disney, who have restricted their content from AI data scrapers. This collaboration is the first of its kind and aims to explore AI-enabled journalism and its potential to enhance journalism&#8217;s quality and business model.<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n\n<p>[12\/15\/2023] The <a href=\"https:\/\/octoml.ai\/blog\/mixtral-8x7b-instruct-model-now-available-on-octoai-text-gen-solution\/\">Mistral\u2019s Mixtral 8x7B Instruct large language model<\/a> (LLM) is now available on the <a href=\"https:\/\/octoml.ai\/\" target=\"_blank\" rel=\"noreferrer noopener\">OctoAI<\/a> Text Gen Solution. Users can benefit from competitive quality to GPT 3.5, the flexibility of open source software, and a 4x lower price per token than GPT 3.5. Details released by Mistral AI this week provided confirmation that Mixtral implements a sparse Mixture of Experts (MoE) architecture, and provided competitive comparisons that showed Mixtral outperforming both Llama 2 70B and GPT 3.5 in several LLM benchmarks. MoE models use conditional computing to limit the number of parameters used in generating each token, lowering the computational needs for training and inference.&nbsp;<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n\n<p>[12\/15\/2023] Meet Natasha at&nbsp;<a href=\"https:\/\/www.builder.ai\/natasha\" target=\"_blank\" rel=\"noreferrer noopener\">www.builder.ai\/natasha<\/a>&nbsp;and see how this AI slowly starts to evolve into a contextual companion across the journey and beyond. Developed by Builder.ai<sup>\u00ae<\/sup>, an AI-powered composable software platform for every idea and company on the planet. The AI-powered assembly line fuses together Lego-like reusable features, using Building Blocks\u2122 automation to reduce human effort, leveraging a verified network of experts to vastly extend development capabilities, and producing apps at an exceptionally high success rate that are multitudes cheaper and faster than traditional software development.<\/p>\n\n\n\n<figure class=\"wp-block-embed is-type-video is-provider-youtube wp-block-embed-youtube wp-embed-aspect-16-9 wp-has-aspect-ratio\"><div class=\"wp-block-embed__wrapper\">\n<iframe loading=\"lazy\"  id=\"_ytid_80465\"  width=\"480\" height=\"270\"  data-origwidth=\"480\" data-origheight=\"270\" src=\"https:\/\/www.youtube.com\/embed\/C8Xo9XFDoYQ?enablejsapi=1&#038;autoplay=0&#038;cc_load_policy=0&#038;cc_lang_pref=&#038;iv_load_policy=1&#038;loop=0&#038;modestbranding=0&#038;rel=1&#038;fs=1&#038;playsinline=0&#038;autohide=2&#038;theme=dark&#038;color=red&#038;controls=1&#038;\" class=\"__youtube_prefs__  epyt-is-override  no-lazyload\" title=\"YouTube player\"  allow=\"accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture\" allowfullscreen data-no-lazy=\"1\" data-skipgform_ajax_framebjll=\"\"><\/iframe>\n<\/div><\/figure>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n<div class=\"wp-block-image\">\n<figure class=\"alignright size-full is-resized\"><img decoding=\"async\" loading=\"lazy\" width=\"600\" height=\"243\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2016\/08\/Intel_updated-logo.png\" alt=\"\" class=\"wp-image-15818\" style=\"aspect-ratio:2.4691358024691357;width:296px;height:auto\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2016\/08\/Intel_updated-logo.png 600w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2016\/08\/Intel_updated-logo-300x122.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2016\/08\/Intel_updated-logo-150x61.png 150w\" sizes=\"(max-width: 600px) 100vw, 600px\" \/><\/figure><\/div>\n\n\n<p><a href=\"https:\/\/www.intel.com\/\" target=\"_blank\" rel=\"noreferrer noopener\">Intel<\/a>&nbsp;unveiled new computer chips, including <a href=\"https:\/\/habana.ai\/\" target=\"_blank\" rel=\"noreferrer noopener\">Gaudi3<\/a>, a chip for generative AI software. Gaudi3 will launch next year and will compete with rival chips from&nbsp;NVIDIA&nbsp;and&nbsp;AMD&nbsp;that power large AI models. The most prominent AI models, like OpenAI\u2019s ChatGPT, run on NVIDIA GPUs in the cloud. It\u2019s one reason NVIDIA stock has been up nearly 230% year to date while Intel shares have risen 68%. And it\u2019s why companies like AMD and, now Intel, have announced chips that they hope will attract AI companies away from NVIDIA\u2019s dominant position in the market. While the company was shy on details, Gaudi3 will compete with NVIDIA\u2019s H100, the main choice among companies that build huge farms of the chips to power AI applications, and AMD\u2019s forthcoming MI300X, when it starts shipping to customers in 2024. Intel has been building Gaudi chips since 2019, when it bought a chip developer called Habana Labs.<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n\n<p>[12\/15\/2023] Check out <a href=\"https:\/\/www.channel1.ai\/\" target=\"_blank\" rel=\"noreferrer noopener\">Channel 1<\/a> for &#8220;AI native news&#8221; bringing trusted news sources to the world by AI generated multilingual reporters. This is unreal! <\/p>\n\n\n\n<figure class=\"wp-block-embed is-type-video is-provider-youtube wp-block-embed-youtube wp-embed-aspect-16-9 wp-has-aspect-ratio\"><div class=\"wp-block-embed__wrapper\">\n<iframe loading=\"lazy\"  id=\"_ytid_78851\"  width=\"480\" height=\"270\"  data-origwidth=\"480\" data-origheight=\"270\" src=\"https:\/\/www.youtube.com\/embed\/ecHioH8fawE?enablejsapi=1&#038;autoplay=0&#038;cc_load_policy=0&#038;cc_lang_pref=&#038;iv_load_policy=1&#038;loop=0&#038;modestbranding=0&#038;rel=1&#038;fs=1&#038;playsinline=0&#038;autohide=2&#038;theme=dark&#038;color=red&#038;controls=1&#038;\" class=\"__youtube_prefs__  epyt-is-override  no-lazyload\" title=\"YouTube player\"  allow=\"accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture\" allowfullscreen data-no-lazy=\"1\" data-skipgform_ajax_framebjll=\"\"><\/iframe>\n<\/div><\/figure>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n\n<p>[12\/14\/2023] Here are some compelling predictions for 2024 from execs at <a href=\"https:\/\/lightning.ai\/\" target=\"_blank\" rel=\"noreferrer noopener\">Lightning AI<\/a>. <\/p>\n\n\n\n<p>According to Lightning AI CEO Will Falcon, in 2024:<\/p>\n\n\n\n<ul>\n<li>Language models will have the same capability as they do now, for 1\/10 of the parameter count<\/li>\n\n\n\n<li>Language models will need 1\/10 of the data for the same performance<\/li>\n\n\n\n<li>Transformer will not be the leading architecture, especially in the lower parameter count models<\/li>\n\n\n\n<li>Systems that allow for multimodal AI will be predominant<\/li>\n\n\n\n<li>RL \/ DPO will enter the mainstream for open source models, alignment recipes (current moat) will be unlocked<\/li>\n\n\n\n<li>Boundaries between pre-training and alignment will start to blur: next token&nbsp;prediction&nbsp;on large corpora will not be the sole strategy<\/li>\n\n\n\n<li>Curriculum will start to be part of the (pre-)training recipe<\/li>\n<\/ul>\n\n\n\n<p>And according to Lightning AI CTO Luca Antiga, In 2024<\/p>\n\n\n\n<ul>\n<li>RL applied to in-context learning will lead to effective agents<\/li>\n\n\n\n<li>Diffusion will enter the language \/ code space<\/li>\n\n\n\n<li>Companies will adopt AI meaningfully without data science \/ research teams<\/li>\n\n\n\n<li>There will be a hockey stick rise of companies including AI in their operations<\/li>\n<\/ul>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n\n<p>[12\/14\/2023] OpenAI\u2019s&nbsp;<a href=\"https:\/\/openai.com\/blog\/introducing-superalignment\" target=\"_blank\" rel=\"noreferrer noopener\">Superalignment<\/a>&nbsp;team just published its first research showing a GPT-2-level model can be used to supervise GPT-4 and recover strong (GPT-3.5-level) performance. This research has unlocked a new approach to the central challenge of aligning future superhuman models while making iterative empirical progress today. Additionally, OpenAI&nbsp;<a href=\"https:\/\/openai.notion.site\/Superalignment-Fast-Grants-and-OpenAI-Generalization-Prizes-fd12c66a286a4cbc9dc0f2fef1c62e92\" target=\"_blank\" rel=\"noreferrer noopener\">launched<\/a>&nbsp;a $10M Superalignment Fast Grants program, in partnership with Eric Schmidt, to support technical research towards ensuring superhuman AI systems are aligned and safe. See below my signature for more details.&nbsp;Please see the&nbsp;<a href=\"https:\/\/cdn.openai.com\/papers\/weak-to-strong-generalization.pdf\" target=\"_blank\" rel=\"noreferrer noopener\">Generalization paper<\/a>&nbsp;and blog posts on&nbsp;<a href=\"https:\/\/openai.com\/research\/weak-to-strong-generalization\" target=\"_blank\" rel=\"noreferrer noopener\">the research<\/a>&nbsp;and&nbsp;<a href=\"https:\/\/openai.com\/blog\/superalignment-fast-grants\" target=\"_blank\" rel=\"noreferrer noopener\">fast grants<\/a>&nbsp;for more details.&nbsp;<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"350\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/MetaAI_Seamless_logo.jpg\" alt=\"\" class=\"wp-image-34192\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/MetaAI_Seamless_logo.jpg 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/MetaAI_Seamless_logo-300x150.jpg 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/MetaAI_Seamless_logo-150x75.jpg 150w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/MetaAI_Seamless_logo-1024x512.jpg 1024w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n\n<p>[12\/13\/2023] FEATURED REPO NEWS &#8211; <a href=\"https:\/\/github.com\/facebookresearch\/seamless_communication\" target=\"_blank\" rel=\"noreferrer noopener\">Seamless<\/a> is a family of AI models from Meta AI that enable more natural and authentic communication across languages. SeamlessM4T is a massive multilingual multimodal machine translation model supporting around 100 languages. SeamlessM4T serves as foundation for SeamlessExpressive, a model that preserves elements of prosody and voice style across languages and SeamlessStreaming, a model supporting simultaneous translation and streaming ASR for around 100 languages. SeamlessExpressive and SeamlessStreaming are combined into Seamless, a unified model featuring multilinguality, real-time and expressive translations.<\/p>\n\n\n\n<p>SeamlessM4T v2 updates the UnitY2 framework from its predecessor and is pre-trained on 4.5M hours of unlabeled audio, and fine tuned on 114,800 hours of automatically aligned data. The architecture is optimized for lower latency, particularly in speech generation, making it more responsive and suitable for real-time applications.<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n\n<p>[12\/13\/2023] At their recent first ever developer conference <a href=\"https:\/\/www.modul.ar\/modcon\" target=\"_blank\" rel=\"noreferrer noopener\">ModCon 2023<\/a>, Modular announced <a href=\"https:\/\/www.modular.com\/max\">Modular Accelerated Xecution (MAX)<\/a>: An integrated, composable suite of products that simplify your AI infrastructure and give you everything you need to deploy low-latency, high-throughput generative and traditional inference pipelines into production. MAX will be available in a free, non-commercial Developer Edition and a paid, commercial Enterprise Edition in early 2024. Check out the ModCon keynote below:<\/p>\n\n\n\n<figure class=\"wp-block-embed is-type-video is-provider-youtube wp-block-embed-youtube wp-embed-aspect-16-9 wp-has-aspect-ratio\"><div class=\"wp-block-embed__wrapper\">\n<iframe loading=\"lazy\" title=\"ModCon23 Keynote Livestream\" width=\"500\" height=\"281\" src=\"https:\/\/www.youtube.com\/embed\/VKxNGFhpYQc?feature=oembed\" frameborder=\"0\" allow=\"accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share\" allowfullscreen><\/iframe>\n<\/div><\/figure>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n\n<p>[12\/13\/2023] Microsoft&nbsp;<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/blog\/phi-2-the-surprising-power-of-small-language-models\/\" target=\"_blank\" rel=\"noreferrer noopener\">launched Phi-2<\/a>, an a 2.7 billion-parameter AI model that matches or outperforms models up to 25x larger.<\/p>\n\n\n\n<p><em>&#8220;The release of Microsoft&#8217;s Phi-2 is a significant milestone,&#8221; said Victor Botev, CTO&nbsp;and co-founder at&nbsp;Iris.ai. &#8220;Microsoft has managed to challenge traditional scaling laws with a smaller-scale model that focuses on &#8216;textbook-quality&#8217; data. It&#8217;s a testament to the fact that there&#8217;s more to AI than just increasing the size of the model.<\/em><\/p>\n\n\n\n<p><em>Microsoft has cited \u201ctraining data curation\u201d as key to Phi-2 performing on par with models 25x larger.&nbsp; While it\u2019s unclear what data and how the model was trained on it, there are a range of innovations that can allow models to do more with less. If the data itself is well structured and promotes reasoning, there is less scope for any model to hallucinate. Coding language can also be used as the training data, as it is more reason-based than text.&nbsp;<\/em><\/p>\n\n\n\n<p><em>We must use domain-specific, structured knowledge to make sure language models ingest, process, and reproduce information on a factual basis. Taking this further, knowledge graphs can assess and demonstrate the steps a language model takes to arrive at its outputs, essentially generating a possible chain of thoughts. The less room for interpretation in this training means models are more likely to be guided to factually accurate answers. They will also require fewer parameters to generate better-reasoned responses.<\/em><\/p>\n\n\n\n<p><em>AI will be transformational for businesses and society, but first it has to&nbsp;be cost-effective. Ever-increasing parameter counts are not financially feasible and have huge implications for energy efficiency. Smaller models with high performance like Phi-2 represent the way forward.\u201d<\/em><\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n<div class=\"wp-block-image\">\n<figure class=\"alignright size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"241\" height=\"166\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/Grok_logo.png\" alt=\"\" class=\"wp-image-34186\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/Grok_logo.png 241w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/Grok_logo-150x103.png 150w\" sizes=\"(max-width: 241px) 100vw, 241px\" \/><\/figure><\/div>\n\n\n<p>[12\/12\/2023] Elon Musk\u2019s X started offering access to its new LLM (Grok) in the US. Designed to compete with other major AI models like OpenAI&#8217;s ChatGPT, <a href=\"https:\/\/grok.x.ai\/\" target=\"_blank\" rel=\"noreferrer noopener\">Grok<\/a> stands out for its integration with social media application X,  allowing it real-time access to information from the platform. This feature gives Grok an edge over other AI models that generally rely on older internet data. Grok is available for X\u2019s US Premium Plus subscribers.<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n<div class=\"wp-block-image\">\n<figure class=\"alignleft size-full is-resized\"><img decoding=\"async\" loading=\"lazy\" width=\"427\" height=\"154\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/Google_NotebookLM_logo.png\" alt=\"\" class=\"wp-image-34184\" style=\"aspect-ratio:2.772727272727273;width:341px;height:auto\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/Google_NotebookLM_logo.png 427w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/Google_NotebookLM_logo-300x108.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/Google_NotebookLM_logo-150x54.png 150w\" sizes=\"(max-width: 427px) 100vw, 427px\" \/><\/figure><\/div>\n\n\n<p>[12\/12\/2023] Google&#8217;s AI-powered writing assistant,<a href=\"https:\/\/notebooklm.google\/\" target=\"_blank\" rel=\"noreferrer noopener\"> NotebookLM<\/a>, is transitioning from an experimental phase to an official service&nbsp;with significant upgrades. Initially introduced as &#8220;Project Tailwind&#8221; at Google I\/O 2023, a new kind of notebook designed to help people learn faster, NotebookLM aims to organize notes by summarizing content and highlighting key topics and questions for better understanding. It\u2019s Google&#8217;s endeavor to reimagine what notetaking software might look like if you designed it from scratch knowing that you would have a powerful language model at its core: hence the LM.&nbsp;The latest version runs on Google&#8217;s advanced AI model, Gemini Pro, which enhances the tool&#8217;s reasoning skills and document comprehension.<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n<div class=\"wp-block-image\">\n<figure class=\"alignright size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"298\" height=\"112\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/Mistral_AI_logo.png\" alt=\"\" class=\"wp-image-34182\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/Mistral_AI_logo.png 298w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/Mistral_AI_logo-150x56.png 150w\" sizes=\"(max-width: 298px) 100vw, 298px\" \/><\/figure><\/div>\n\n\n<p>[12\/12\/2023] French startup <a href=\"https:\/\/mistral.ai\/\" target=\"_blank\" rel=\"noreferrer noopener\">Mistral AI<\/a> announced it completed a Series A funding round, raising \u20ac385 million (approximately $415 million), valuing the company at around $2 billion. The company, co-founded by Google DeepMind and Meta alumni, focuses on developing foundational models with an open technology approach.<\/p>\n\n\n\n<p>The start-up, best known for its Mistral 7B model and advocacy for regulatory exemptions for foundational models, has recently released <a href=\"https:\/\/mistral.ai\/news\/mixtral-of-experts\/\" target=\"_blank\" rel=\"noreferrer noopener\">Mixtral 8x7B and Mistral-medium models<\/a>, both available through its newly launched developer platform. While Mixtral 8x7B is accessible as a free download, Mistral-medium is exclusive to the paid API platform, reflecting the company&#8217;s strategy to monetize its AI models.<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n\n<p>[12\/11\/2023] Here is my favorite paper so far from NeurIPS 2023 happening this week: &#8220;<a href=\"https:\/\/arxiv.org\/pdf\/2304.15004.pdf\" target=\"_blank\" rel=\"noreferrer noopener\">Are Emergent Abilities of Large Language Models a Mirage?<\/a>&#8221; It looks deep into GenAI model explainability and interpretability. The primary author, Rylan Schaeffer, did an oral poster presentation today that was quite compelling. The paper was voted one of only two &#8220;Outstanding Main Track Papers.&#8221; <\/p>\n\n\n\n<p><strong>Abstract:<\/strong><\/p>\n\n\n\n<p>Recent work claims that large language models display emergent abilities, abilities not present in smaller-scale models that are present in larger-scale models. What makes emergent abilities intriguing is two-fold: their sharpness, transitioning seemingly instantaneously from not present to present, and their unpredictability, appearing at seemingly unforeseeable model scales. Here, we present an alternative explanation for emergent abilities: that for a particular task and model family, when analyzing fixed model outputs, emergent abilities appear due the researcher\u2019s choice of metric rather than due to fundamental changes in model behavior with scale. Specifically, nonlinear or discontinuous metrics produce apparent emergent abilities, whereas linear or continuous metrics produce smooth, continuous, predictable changes in model performance. We present our alternative explanation in a simple mathematical model, then test it in three complementary ways: we (1) make, test and confirm three predictions on the effect of metric choice using the InstructGPT\/GPT-3 family on tasks with claimed emergent abilities, (2) make, test and confirm two predictions about metric choices in a meta-analysis of emergent abilities on BIG-Bench; and (3) show how to choose metrics to produce never-before-seen seemingly emergent abilities in multiple vision tasks across diverse deep networks. Via all three analyses, we provide evidence that alleged emergent abilities evaporate with different metrics or with better statistics, and may not be a fundamental property of scaling AI models.<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n\n<p>[12\/11\/2023] Google is exploring a new application called &#8220;<a href=\"https:\/\/www.cnbc.com\/2023\/12\/08\/google-weighing-project-ellmann-uses-gemini-ai-to-tell-life-stories.html\" target=\"_blank\" rel=\"noreferrer noopener\">Project Ellman<\/a>,&#8221; which proposes to use its new Gemini AI models to create personalized life stories for users by analyzing images from Google Photos and text from Google Search. The conceptual project was named after the literary critic Richard David Ellmann, and envisions a chatbot that knows everything about a person&#8217;s life, weaving a narrative from various data sources like photos and public internet information.<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n<div class=\"wp-block-image\">\n<figure class=\"alignright size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"292\" height=\"72\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/NeurIPS2023_logo.png\" alt=\"\" class=\"wp-image-34144\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/NeurIPS2023_logo.png 292w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/NeurIPS2023_logo-150x37.png 150w\" sizes=\"(max-width: 292px) 100vw, 292px\" \/><\/figure><\/div>\n\n\n<p>[12\/10\/2023] The<a href=\"https:\/\/neurips.cc\/\" target=\"_blank\" rel=\"noreferrer noopener\"> NeurIPS<\/a> (Neural Information Processing Systems) 2023 Conference starts today and runs through Dec. 16 in New Orleans. This is the premiere AI research conference of the year and has become one of my favorite events. The academic vibe is invigorating and I am looking forward to many of the sessions and research papers featured on agenda. The NeurIPS purpose is to foster the exchange of research advances in Artificial Intelligence and Machine Learning. <\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n<div class=\"wp-block-image\">\n<figure class=\"alignright size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"300\" height=\"158\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/Google_Gemini_logo.png\" alt=\"\" class=\"wp-image-34147\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/Google_Gemini_logo.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/Google_Gemini_logo-150x79.png 150w\" sizes=\"(max-width: 300px) 100vw, 300px\" \/><\/figure><\/div>\n\n\n<p>[12\/8\/2023] Google just unveiled&nbsp;<a href=\"https:\/\/blog.google\/technology\/ai\/google-gemini-ai\/\" target=\"_blank\" rel=\"noreferrer noopener\">Gemini<\/a> 1.0, its largest and most capable AI model. Gemini was trained using the company&#8217;s custom-designed AI accelerators, Cloud TPU v4 and v5e.&nbsp;Built natively to be multimodal, it\u2019s the first step in the Gemini-era of models. Gemini is optimized in three sizes &#8211; Ultra, Pro, and Nano. In benchmark tests, Gemini outperforms OpenAI&#8217;s GPT-4 in 30 of 32 tests, particularly in multimodal understanding and Python code generation.<\/p>\n\n\n\n<p>Each model targets specific applications. Gemini Ultra, is able to perform complex tasks in data centers and enterprise applications, harnessing the full power of Google&#8217;s AI capabilities. Gemini Pro serves a wider array of AI services, integrating seamlessly with Google&#8217;s own AI service, Bard. Lastly, Gemini Nano has two versions: Nano-1 with 1.8 billion parameters and Nano-2 with 3.25 billion parameters. These models are specifically engineered for on-device operations, with a focus on optimizing performance in Android environments.&nbsp;For coding, Gemini uses AlphaCode 2, a code-generating system that shows the model&#8217;s proficiency in understanding and creating high-quality code in various languages.<\/p>\n\n\n\n<p>Central to the Gemini models is an architecture built upon enhanced Transformer decoders, specifically tailored for Google&#8217;s own Tensor Processing Units (TPUs). This coupling of hardware and software enables the models to achieve efficient training and inference processes, setting them apart in terms of speed and cost-effectiveness compared to previous iterations like PaLM.<\/p>\n\n\n\n<p>A key element of the Gemini suite is its multimodal nature &#8211; trained on a vast array of datasets including text, images, audio, and code. Gemini&#8217;s reportedly surpass OpenAI&#8217;s GPT-4 in various performance benchmarks, especially in multimodal understanding and Python code generation. The version just released, Gemini Pro, is a lighter variant of a more advanced model, Gemini Ultra, expected next year. Gemini Pro is now powering Bard, Google&#8217;s ChatGPT rival, and promises improved abilities in reasoning and understanding. <\/p>\n\n\n\n<p>Gemini Ultra is said to be &#8220;natively multimodal,&#8221; processing a diverse range of data including text, images, audio, and videos. This capability surpasses OpenAI&#8217;s GPT-4 in vision problem domains, but the improvements are marginal in many aspects. In some benchmarks, for example, Gemini Ultra only slightly outperforms GPT-4.<\/p>\n\n\n\n<p>A concerning aspect of Gemini is Google&#8217;s secrecy around the model&#8217;s training data. Questions about the data&#8217;s sources and creators&#8217; rights were not answered. This is critical, as increasingly the AI industry is facing lawsuits over using copyrighted content without compensation and\/or credit.<\/p>\n\n\n\n<p>Gemini is getting a&nbsp;<a href=\"https:\/\/techcrunch.com\/2023\/12\/07\/early-impressions-of-googles-gemini-arent-great\/\">mixed reception<\/a>&nbsp;after its big debut on Dec. 6, 2023, but users may have less confidence in the company\u2019s multimodal technology and\/or integrity after finding out that the most impressive demo of Gemini was pretty much faked. Parmy Olson at Bloomberg was the&nbsp;<a href=\"https:\/\/twitter.com\/parmy\/status\/1732811357068615969\" target=\"_blank\" rel=\"noreferrer noopener\">first to report<\/a>&nbsp;the discrepancy. <a href=\"https:\/\/techcrunch.com\/2023\/12\/07\/googles-best-gemini-demo-was-faked\/\" target=\"_blank\" rel=\"noreferrer noopener\">TechCrunch<\/a> does a great job itemizing the issues with the video below. <\/p>\n\n\n\n<figure class=\"wp-block-embed is-type-video is-provider-youtube wp-block-embed-youtube wp-embed-aspect-16-9 wp-has-aspect-ratio\"><div class=\"wp-block-embed__wrapper\">\n<iframe loading=\"lazy\"  id=\"_ytid_17149\"  width=\"480\" height=\"270\"  data-origwidth=\"480\" data-origheight=\"270\" src=\"https:\/\/www.youtube.com\/embed\/UIZAiXYceBI?enablejsapi=1&#038;autoplay=0&#038;cc_load_policy=0&#038;cc_lang_pref=&#038;iv_load_policy=1&#038;loop=0&#038;modestbranding=0&#038;rel=1&#038;fs=1&#038;playsinline=0&#038;autohide=2&#038;theme=dark&#038;color=red&#038;controls=1&#038;\" class=\"__youtube_prefs__  epyt-is-override  no-lazyload\" title=\"YouTube player\"  allow=\"accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture\" allowfullscreen data-no-lazy=\"1\" data-skipgform_ajax_framebjll=\"\"><\/iframe>\n<\/div><figcaption class=\"wp-element-caption\">Source: Google &#8211; Hands-on with Gemini: Interacting with multimodal AI<\/figcaption><\/figure>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full is-resized\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" style=\"aspect-ratio:7.608695652173913;width:700px;height:auto\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n\n<p>[12\/8\/2023] <a href=\"https:\/\/arstechnica.com\/ai\/2023\/12\/the-real-research-behind-the-wild-rumors-about-openais-q-project\/\" target=\"_blank\" rel=\"noreferrer noopener\">OpenAI&#8217;s Q* model<\/a> can reportedly perform math on the level of grade-school students. OpenAI hasn&#8217;t said what Q* is, but it has revealed plenty of clues. While Q* might not be the crucial breakthrough that will lead to AGI, it could be a strategic step towards an AI with general reasoning abilities.<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n\n<p>[12\/8\/2023] Databricks <a href=\"https:\/\/www.databricks.com\/blog\/improve-your-rag-application-response-quality-real-time-structured-data\" target=\"_blank\" rel=\"noreferrer noopener\">unveiled<\/a> new retrieval augmented generation (RAG) tooling to help build high-quality large language model applications. Key features include vector search to integrate unstructured data, low latency feature serving for structured data, and monitoring systems to scan model responses. By combining relevant contextual data sources, these capabilities aim to simplify productionizing accurate and reliable RAG apps across various business use cases.<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n\n<p>[12\/7\/2023] <strong>Chain of Code: Reasoning with a Language Model-Augmented Code Emulator<\/strong>: Chain of Code (CoT), as described in a <a href=\"https:\/\/arxiv.org\/pdf\/2312.04474.pdf\" target=\"_blank\" rel=\"noreferrer noopener\">paper<\/a> by researchers from Google DeepMind, Stanford and U.C. Berkeley, significantly enhances language models&#8217; (LMs) reasoning capabilities by integrating code emulation, achieving a notable 12% improvement in performance over previous methods. Traditional LMs face challenges in accurately processing complex logic and linguistic tasks, especially when these tasks require understanding and manipulating code-like structures. CoT addresses this by allowing LMs to format tasks as pseudocode, which is then interpreted by a specialized emulator. The so-called &#8220;LMulator&#8221; effectively simulates code execution, providing a more robust reasoning framework for LMs. CoT&#8217;s effectiveness is demonstrated through its performance on the BIG-Bench Hard benchmark, where it achieves an 84% success rate, outperforming the previous Chain of Thought method by 12%. This showcases its ability to broaden the range of reasoning tasks LMs can handle.<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n<div class=\"wp-block-image\">\n<figure class=\"alignleft size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"400\" height=\"264\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/AMD_Instinct.png\" alt=\"\" class=\"wp-image-34150\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/AMD_Instinct.png 400w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/AMD_Instinct-300x198.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/AMD_Instinct-150x99.png 150w\" sizes=\"(max-width: 400px) 100vw, 400px\" \/><\/figure><\/div>\n\n\n<p>[12\/7\/2023] <a href=\"http:\/\/www.amd.com\/\" target=\"_blank\" rel=\"noreferrer noopener\">AMD<\/a>&nbsp;(NASDAQ: AMD) announced the availability of the AMD Instinct\u2122 MI300X accelerators \u2013 with industry leading memory bandwidth for generative AI&nbsp;and leadership performance for large language model (LLM) training and inferencing \u2013 as well as the AMD Instinct\u2122 MI300A accelerated processing unit (APU) \u2013 combining the latest AMD CDNA\u2122 3 architecture and \u201cZen 4\u201d CPUs to deliver breakthrough performance for HPC and AI workloads.<\/p>\n\n\n\n<p><em>\u201cAMD Instinct MI300 Series accelerators are designed with our most advanced technologies, delivering leadership performance, and will be in large scale cloud and enterprise deployments,\u201d said Victor Peng, president, AMD. \u201cBy leveraging our leadership hardware, software and open ecosystem approach, cloud providers, OEMs and ODMs are bringing to market technologies that empower enterprises to adopt and deploy AI-powered solutions.\u201d<\/em><\/p>\n\n\n\n<p><strong>AMD Instinct MI300X<\/strong><\/p>\n\n\n\n<p>AMD Instinct MI300X accelerators are powered by the new AMD CDNA 3 architecture. When compared to previous generation AMD Instinct MI250X accelerators, MI300X delivers nearly 40% more compute units<sup>2<\/sup>, 1.5x more memory capacity, 1.7x more peak theoretical memory bandwidth<sup>3<\/sup>&nbsp;as well as support for new math formats such as FP8 and sparsity; all geared towards AI and HPC workloads.<\/p>\n\n\n\n<p>Today\u2019s LLMs continue to increase in size and complexity, requiring massive amounts of memory and compute. AMD Instinct MI300X accelerators feature a best-in-class 192 GB of HBM3 memory capacity as well as 5.3 TB\/s peak memory bandwidth<sup>2<\/sup>&nbsp;to deliver the performance needed for increasingly demanding AI workloads. The AMD Instinct Platform is a leadership generative AI platform built on an industry standard OCP design with eight MI300X accelerators to offer an industry leading 1.5TB of HBM3 memory capacity. The AMD Instinct Platform\u2019s industry standard design allows OEM partners to design-in MI300X accelerators into existing AI offerings and simplify deployment and accelerate adoption of AMD Instinct accelerator-based servers.<\/p>\n\n\n\n<p>Compared to the Nvidia H100 HGX, the AMD Instinct Platform can offer a throughput increase of up to 1.6x when running inference on LLMs like BLOOM 176B<sup>4<\/sup>&nbsp;and is the only option on the market capable of running inference for a 70B parameter model, like Llama2, on a single MI300X accelerator; simplifying enterprise-class LLM deployments and enabling outstanding TCO.<\/p>\n\n\n\n<p><strong>AMD Instinct MI300A<\/strong><\/p>\n\n\n\n<p>The AMD Instinct MI300A APUs, the world\u2019s first data center APU for HPC and AI, leverage 3D packaging and the 4<sup>th<\/sup>&nbsp;Gen AMD Infinity Architecture to deliver leadership performance on critical workloads sitting at the convergence of HPC and AI. MI300A APUs combine high-performance AMD CDNA 3 GPU cores, the latest AMD \u201cZen 4\u201d x86-based CPU cores and 128GB of next-generation HBM3 memory, to deliver ~1.9x the performance-per-watt on FP32 HPC and AI workloads, compared to previous gen AMD Instinct MI250X<sup>5<\/sup>.<\/p>\n\n\n\n<p>Energy efficiency is of utmost importance for the HPC and AI communities, however these workloads are extremely data- and resource-intensive. AMD Instinct MI300A APUs benefit from integrating CPU and GPU cores on a single package delivering a highly efficient platform while also providing the compute performance to accelerate training the latest AI models. AMD is setting the pace of innovation in energy efficiency with the company\u2019s&nbsp;30&#215;25&nbsp;goal, aiming to deliver a 30x energy efficiency improvement in server processors and accelerators for AI-training and HPC from 2020-2025<sup>6<\/sup>.<\/p>\n\n\n\n<p>The APU advantage means that AMD Instinct MI300A APUs feature unified memory and cache resources giving customers an easily programmable GPU platform, highly performant compute, fast AI training and impressive energy efficiency to power the most demanding HPC and AI workloads.<\/p>\n\n\n\n<figure class=\"wp-block-embed is-type-video is-provider-youtube wp-block-embed-youtube wp-embed-aspect-16-9 wp-has-aspect-ratio\"><div class=\"wp-block-embed__wrapper\">\n<iframe loading=\"lazy\"  id=\"_ytid_90473\"  width=\"480\" height=\"270\"  data-origwidth=\"480\" data-origheight=\"270\" src=\"https:\/\/www.youtube.com\/embed\/pVl25BbczLI?enablejsapi=1&#038;autoplay=0&#038;cc_load_policy=0&#038;cc_lang_pref=&#038;iv_load_policy=1&#038;loop=0&#038;modestbranding=0&#038;rel=1&#038;fs=1&#038;playsinline=0&#038;autohide=2&#038;theme=dark&#038;color=red&#038;controls=1&#038;\" class=\"__youtube_prefs__  epyt-is-override  no-lazyload\" title=\"YouTube player\"  allow=\"accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture\" allowfullscreen data-no-lazy=\"1\" data-skipgform_ajax_framebjll=\"\"><\/iframe>\n<\/div><\/figure>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n\n<p>[12\/7\/2023] Stability released <a href=\"https:\/\/stability.ai\/news\/stablelm-zephyr-3b-stability-llm\" target=\"_blank\" rel=\"noreferrer noopener\">Stable LM Zephyr 3B<\/a>, a compact LLM: The 3 billion parameter LLM is 60% smaller than typical 7B models, efficiently runs on edge devices, and is fine-tuned on datasets like UltraChat and MetaMathQA, excelling in Q&amp;A tasks, benchmarked on MT Bench and AlpacaEval.<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n\n<p>[12\/6\/2023] An new MIT-spinout, <a href=\"https:\/\/www.liquid.ai\/\" target=\"_blank\" rel=\"noreferrer noopener\">Liquid AI<\/a>, emerged from stealth with<strong>&nbsp;$37.5M&nbsp;<\/strong>in seed funds at $303M post-money valuation to build a small-scale &#8220;liquid neural network&#8221; backed by Samsung Next, Bold Capital Partners, and ISAI Cap Venture. <\/p>\n\n\n\n<p>A research paper titled \u201c<a href=\"https:\/\/arxiv.org\/abs\/2006.04439\" target=\"_blank\" rel=\"noreferrer noopener\">Liquid Time-constant Networks<\/a>,\u201d published at the tail end of 2020 by Hasani, Rus, Lechner, Amini and others, put liquid neural networks on the map following several years of fits and starts; liquid neural networks as a concept have been around since 2018.<\/p>\n\n\n\n<p>Liquid neural networks consist of \u201cneurons\u201d governed by equations that predict each individual neuron\u2019s behavior over time, like most other modern model architectures. The \u201cliquid\u201d in the term \u201cliquid neural networks\u201d refers to the architecture\u2019s flexibility; inspired by the \u201cbrains\u201d of roundworms, not only are liquid neural networks much smaller than traditional AI models, but they require far less compute power to run.<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"92\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png\" alt=\"\" class=\"wp-image-33946\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-300x39.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/11\/AINewsBriefs_divider-150x20.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n\n<p>[12\/1\/2\/2023] Deci announced the release of a new object detection model, YOLO-NAS-POSE &#8211; a derivative of&nbsp;<a href=\"https:\/\/github.com\/Deci-AI\/super-gradients\/blob\/master\/YOLONAS.md\" target=\"_blank\" rel=\"noreferrer noopener\">YOLO-NAS<\/a>, pose estimation architecture, providing superior real-time object detection capabilities and production-ready performance. Deci&#8217;s mission is to provide AI teams with tools to remove development barriers and attain efficient inference performance more quickly.<\/p>\n\n\n\n<p>YOLO-NAS Pose offers superior accuracy-latency balance compared to YOLOv8 Pose with 38% lower latency and higher precision.&nbsp;YOLO-NAS-Pose performs simultaneous person detection and pose prediction in a single-pass image process along with simplified post-processing, enabling high speed and ease of deployment.&nbsp;With its one-line export to ONNX and NVIDIA TensorRT, conversion into production frameworks is swift and smooth. The YOLO-NAS Pose architecture is available under an open-source license. Its pre-trained weights are available for non-commercial use on SuperGradients, Deci&#8217;s PyTorch-based, open-source, computer vision training library.&nbsp;<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"432\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/Deci_POSE_model.png\" alt=\"\" class=\"wp-image-34194\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/Deci_POSE_model.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/Deci_POSE_model-300x185.png 300w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/12\/Deci_POSE_model-150x93.png 150w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n\n<p><em>Sign up for the free insideBIGDATA&nbsp;<a href=\"http:\/\/inside-bigdata.com\/newsletter\/\" target=\"_blank\" rel=\"noreferrer noopener\">newsletter<\/a>.<\/em><\/p>\n\n\n\n<p><em>Join us on Twitter:&nbsp;<a href=\"https:\/\/twitter.com\/InsideBigData1\" target=\"_blank\" rel=\"noreferrer noopener\">https:\/\/twitter.com\/InsideBigData1<\/a><\/em><\/p>\n\n\n\n<p><em>Join us on LinkedIn:&nbsp;<a href=\"https:\/\/www.linkedin.com\/company\/insidebigdata\/\" target=\"_blank\" rel=\"noreferrer noopener\">https:\/\/www.linkedin.com\/company\/insidebigdata\/<\/a><\/em><\/p>\n\n\n\n<p><em>Join us on Facebook:&nbsp;<a href=\"https:\/\/www.facebook.com\/insideBIGDATANOW\" target=\"_blank\" rel=\"noreferrer noopener\">https:\/\/www.facebook.com\/insideBIGDATANOW<\/a><\/em><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Welcome insideBIGDATA AI News Briefs Bulletin Board, our timely new feature bringing you the latest industry insights and perspectives surrounding the field of AI including deep learning, large language models, generative AI, and transformers. We\u2019re working tirelessly to dig up the most timely and curious tidbits underlying the day\u2019s most popular technologies. We know this field is advancing rapidly and we want to bring you a regular resource to keep you informed and state-of-the-art. The news bites are constantly being added in reverse date order (most recent on top). With our bulletin board you can check back often to see what&#8217;s happening in our rapidly accelerating industry. <\/p>\n","protected":false},"author":37,"featured_media":32924,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"jetpack_post_was_ever_published":false,"footnotes":""},"categories":[526,1369,65,115,62,63,64,66,182,1054,180,67,268,56,1],"tags":[280,133,1245,1248,277,96],"acf":[],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v20.6 - https:\/\/yoast.com\/wordpress\/plugins\/seo\/ -->\n<title>insideBIGDATA AI News Briefs Bulletin Board - insideBIGDATA<\/title>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/insidebigdata.com\/2023\/12\/19\/insidebigdata-ai-news-briefs-bulletin-board\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"insideBIGDATA AI News Briefs Bulletin Board - insideBIGDATA\" \/>\n<meta property=\"og:description\" content=\"Welcome insideBIGDATA AI News Briefs Bulletin Board, our timely new feature bringing you the latest industry insights and perspectives surrounding the field of AI including deep learning, large language models, generative AI, and transformers. We\u2019re working tirelessly to dig up the most timely and curious tidbits underlying the day\u2019s most popular technologies. We know this field is advancing rapidly and we want to bring you a regular resource to keep you informed and state-of-the-art. The news bites are constantly being added in reverse date order (most recent on top). With our bulletin board you can check back often to see what&#039;s happening in our rapidly accelerating industry.\" \/>\n<meta property=\"og:url\" content=\"https:\/\/insidebigdata.com\/2023\/12\/19\/insidebigdata-ai-news-briefs-bulletin-board\/\" \/>\n<meta property=\"og:site_name\" content=\"insideBIGDATA\" \/>\n<meta property=\"article:publisher\" content=\"http:\/\/www.facebook.com\/insidebigdata\" \/>\n<meta property=\"article:published_time\" content=\"2023-12-19T10:59:00+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2023-12-21T04:28:10+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/07\/AI-News-Briefs-column-banner.png\" \/>\n\t<meta property=\"og:image:width\" content=\"1100\" \/>\n\t<meta property=\"og:image:height\" content=\"550\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/png\" \/>\n<meta name=\"author\" content=\"Daniel Gutierrez\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:creator\" content=\"@AMULETAnalytics\" \/>\n<meta name=\"twitter:site\" content=\"@insideBigData\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"Daniel Gutierrez\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"24 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"WebPage\",\"@id\":\"https:\/\/insidebigdata.com\/2023\/12\/19\/insidebigdata-ai-news-briefs-bulletin-board\/\",\"url\":\"https:\/\/insidebigdata.com\/2023\/12\/19\/insidebigdata-ai-news-briefs-bulletin-board\/\",\"name\":\"insideBIGDATA AI News Briefs Bulletin Board - insideBIGDATA\",\"isPartOf\":{\"@id\":\"https:\/\/insidebigdata.com\/#website\"},\"datePublished\":\"2023-12-19T10:59:00+00:00\",\"dateModified\":\"2023-12-21T04:28:10+00:00\",\"author\":{\"@id\":\"https:\/\/insidebigdata.com\/#\/schema\/person\/2540da209c83a68f4f5922848f7376ed\"},\"breadcrumb\":{\"@id\":\"https:\/\/insidebigdata.com\/2023\/12\/19\/insidebigdata-ai-news-briefs-bulletin-board\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/insidebigdata.com\/2023\/12\/19\/insidebigdata-ai-news-briefs-bulletin-board\/\"]}]},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/insidebigdata.com\/2023\/12\/19\/insidebigdata-ai-news-briefs-bulletin-board\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\/\/insidebigdata.com\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"insideBIGDATA AI News Briefs Bulletin Board\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/insidebigdata.com\/#website\",\"url\":\"https:\/\/insidebigdata.com\/\",\"name\":\"insideBIGDATA\",\"description\":\"Your Source for AI, Data Science, Deep Learning &amp; Machine Learning Strategies\",\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/insidebigdata.com\/?s={search_term_string}\"},\"query-input\":\"required name=search_term_string\"}],\"inLanguage\":\"en-US\"},{\"@type\":\"Person\",\"@id\":\"https:\/\/insidebigdata.com\/#\/schema\/person\/2540da209c83a68f4f5922848f7376ed\",\"name\":\"Daniel Gutierrez\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/insidebigdata.com\/#\/schema\/person\/image\/\",\"url\":\"https:\/\/secure.gravatar.com\/avatar\/5780282e7e567e2a502233e948464542?s=96&d=mm&r=g\",\"contentUrl\":\"https:\/\/secure.gravatar.com\/avatar\/5780282e7e567e2a502233e948464542?s=96&d=mm&r=g\",\"caption\":\"Daniel Gutierrez\"},\"description\":\"Daniel D. Gutierrez is a Data Scientist with Los Angeles-based AMULET Analytics, a service division of AMULET Development Corp. He's been involved with data science and Big Data long before it came in vogue, so imagine his delight when the Harvard Business Review recently deemed \\\"data scientist\\\" as the sexiest profession for the 21st century. Previously, he taught computer science and database classes at UCLA Extension for over 15 years, and authored three computer industry books on database technology. He also served as technical editor, columnist and writer at a major computer industry monthly publication for 7 years. Follow his data science musings at @AMULETAnalytics.\",\"sameAs\":[\"http:\/\/www.insidebigdata.com\",\"https:\/\/twitter.com\/@AMULETAnalytics\"],\"url\":\"https:\/\/insidebigdata.com\/author\/dangutierrez\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"insideBIGDATA AI News Briefs Bulletin Board - insideBIGDATA","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/insidebigdata.com\/2023\/12\/19\/insidebigdata-ai-news-briefs-bulletin-board\/","og_locale":"en_US","og_type":"article","og_title":"insideBIGDATA AI News Briefs Bulletin Board - insideBIGDATA","og_description":"Welcome insideBIGDATA AI News Briefs Bulletin Board, our timely new feature bringing you the latest industry insights and perspectives surrounding the field of AI including deep learning, large language models, generative AI, and transformers. We\u2019re working tirelessly to dig up the most timely and curious tidbits underlying the day\u2019s most popular technologies. We know this field is advancing rapidly and we want to bring you a regular resource to keep you informed and state-of-the-art. The news bites are constantly being added in reverse date order (most recent on top). With our bulletin board you can check back often to see what's happening in our rapidly accelerating industry.","og_url":"https:\/\/insidebigdata.com\/2023\/12\/19\/insidebigdata-ai-news-briefs-bulletin-board\/","og_site_name":"insideBIGDATA","article_publisher":"http:\/\/www.facebook.com\/insidebigdata","article_published_time":"2023-12-19T10:59:00+00:00","article_modified_time":"2023-12-21T04:28:10+00:00","og_image":[{"width":1100,"height":550,"url":"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/07\/AI-News-Briefs-column-banner.png","type":"image\/png"}],"author":"Daniel Gutierrez","twitter_card":"summary_large_image","twitter_creator":"@AMULETAnalytics","twitter_site":"@insideBigData","twitter_misc":{"Written by":"Daniel Gutierrez","Est. reading time":"24 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"WebPage","@id":"https:\/\/insidebigdata.com\/2023\/12\/19\/insidebigdata-ai-news-briefs-bulletin-board\/","url":"https:\/\/insidebigdata.com\/2023\/12\/19\/insidebigdata-ai-news-briefs-bulletin-board\/","name":"insideBIGDATA AI News Briefs Bulletin Board - insideBIGDATA","isPartOf":{"@id":"https:\/\/insidebigdata.com\/#website"},"datePublished":"2023-12-19T10:59:00+00:00","dateModified":"2023-12-21T04:28:10+00:00","author":{"@id":"https:\/\/insidebigdata.com\/#\/schema\/person\/2540da209c83a68f4f5922848f7376ed"},"breadcrumb":{"@id":"https:\/\/insidebigdata.com\/2023\/12\/19\/insidebigdata-ai-news-briefs-bulletin-board\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/insidebigdata.com\/2023\/12\/19\/insidebigdata-ai-news-briefs-bulletin-board\/"]}]},{"@type":"BreadcrumbList","@id":"https:\/\/insidebigdata.com\/2023\/12\/19\/insidebigdata-ai-news-briefs-bulletin-board\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/insidebigdata.com\/"},{"@type":"ListItem","position":2,"name":"insideBIGDATA AI News Briefs Bulletin Board"}]},{"@type":"WebSite","@id":"https:\/\/insidebigdata.com\/#website","url":"https:\/\/insidebigdata.com\/","name":"insideBIGDATA","description":"Your Source for AI, Data Science, Deep Learning &amp; Machine Learning Strategies","potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/insidebigdata.com\/?s={search_term_string}"},"query-input":"required name=search_term_string"}],"inLanguage":"en-US"},{"@type":"Person","@id":"https:\/\/insidebigdata.com\/#\/schema\/person\/2540da209c83a68f4f5922848f7376ed","name":"Daniel Gutierrez","image":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/insidebigdata.com\/#\/schema\/person\/image\/","url":"https:\/\/secure.gravatar.com\/avatar\/5780282e7e567e2a502233e948464542?s=96&d=mm&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/5780282e7e567e2a502233e948464542?s=96&d=mm&r=g","caption":"Daniel Gutierrez"},"description":"Daniel D. Gutierrez is a Data Scientist with Los Angeles-based AMULET Analytics, a service division of AMULET Development Corp. He's been involved with data science and Big Data long before it came in vogue, so imagine his delight when the Harvard Business Review recently deemed \"data scientist\" as the sexiest profession for the 21st century. Previously, he taught computer science and database classes at UCLA Extension for over 15 years, and authored three computer industry books on database technology. He also served as technical editor, columnist and writer at a major computer industry monthly publication for 7 years. Follow his data science musings at @AMULETAnalytics.","sameAs":["http:\/\/www.insidebigdata.com","https:\/\/twitter.com\/@AMULETAnalytics"],"url":"https:\/\/insidebigdata.com\/author\/dangutierrez\/"}]}},"jetpack_featured_media_url":"https:\/\/insidebigdata.com\/wp-content\/uploads\/2023\/07\/AI-News-Briefs-column-banner.png","jetpack_shortlink":"https:\/\/wp.me\/p9eA3j-8Sx","jetpack-related-posts":[{"id":32923,"url":"https:\/\/insidebigdata.com\/2023\/07\/27\/insidebigdata-ai-news-briefs-7-27-2023\/","url_meta":{"origin":34133,"position":0},"title":"insideBIGDATA AI News Briefs &#8211; 7\/27\/2023","date":"July 27, 2023","format":false,"excerpt":"Welcome insideBIGDATA AI News Briefs, our podcast channel bringing you the latest industry insights and perspectives surrounding the field of AI including deep learning, large language models, generative AI, and transformers. We're working tirelessly to dig up the most timely and curious tidbits underlying the day's most popular technologies. We\u2026","rel":"","context":"In &quot;AI Deep Learning&quot;","img":{"alt_text":"","src":"https:\/\/i0.wp.com\/insidebigdata.com\/wp-content\/uploads\/2023\/07\/AI-News-Briefs-column-banner.png?resize=350%2C200&ssl=1","width":350,"height":200},"classes":[]},{"id":33242,"url":"https:\/\/insidebigdata.com\/2023\/08\/31\/insidebigdata-ai-news-briefs-8-31-2023\/","url_meta":{"origin":34133,"position":1},"title":"insideBIGDATA AI News Briefs \u2013 8\/31\/2023","date":"August 31, 2023","format":false,"excerpt":"Welcome insideBIGDATA AI News Briefs, our timely new feature bringing you the latest industry insights and perspectives surrounding the field of AI including deep learning, large language models, generative AI, and transformers. We\u2019re working tirelessly to dig up the most timely and curious tidbits underlying the day\u2019s most popular technologies.\u2026","rel":"","context":"In &quot;AI Deep Learning&quot;","img":{"alt_text":"","src":"https:\/\/i0.wp.com\/insidebigdata.com\/wp-content\/uploads\/2023\/07\/AI-News-Briefs-column-banner.png?resize=350%2C200&ssl=1","width":350,"height":200},"classes":[]},{"id":32718,"url":"https:\/\/insidebigdata.com\/2023\/06\/30\/power-to-the-data-report-podcast-generative-ai-economic-forecast\/","url_meta":{"origin":34133,"position":2},"title":"Power to the Data Report Podcast: Generative AI Economic Forecast","date":"June 30, 2023","format":false,"excerpt":"Welcome to the \u201cPower-to-the-Data Report\u201d podcast where we cover timely topics of the day from throughout the Big Data ecosystem. I am your host Daniel Gutierrez from insideBIGDATA where I serve as Editor-in-Chief & Resident Data Scientist. Today\u2019s topic is an economic forecast for the accelerating rise of generative AI\u2026","rel":"","context":"In &quot;AI Deep Learning&quot;","img":{"alt_text":"","src":"https:\/\/i0.wp.com\/insidebigdata.com\/wp-content\/uploads\/2023\/06\/Power-Data-column-banner_special.jpg?resize=350%2C200&ssl=1","width":350,"height":200},"classes":[]},{"id":33298,"url":"https:\/\/insidebigdata.com\/2023\/09\/06\/insidebigdata-ai-news-briefs-9-8-2023\/","url_meta":{"origin":34133,"position":3},"title":"insideBIGDATA AI News Briefs \u2013 9\/8\/2023","date":"September 6, 2023","format":false,"excerpt":"Welcome insideBIGDATA AI News Briefs, our timely new feature bringing you the latest industry insights and perspectives surrounding the field of AI including deep learning, large language models, generative AI, and transformers. We\u2019re working tirelessly to dig up the most timely and curious tidbits underlying the day\u2019s most popular technologies.\u2026","rel":"","context":"In &quot;AI Deep Learning&quot;","img":{"alt_text":"","src":"https:\/\/i0.wp.com\/insidebigdata.com\/wp-content\/uploads\/2023\/07\/AI-News-Briefs-column-banner.png?resize=350%2C200&ssl=1","width":350,"height":200},"classes":[]},{"id":32878,"url":"https:\/\/insidebigdata.com\/2023\/07\/25\/video-highlights-generative-ai-with-large-language-models\/","url_meta":{"origin":34133,"position":4},"title":"Video Highlights: Generative AI with Large Language Models","date":"July 25, 2023","format":false,"excerpt":"At an unprecedented pace, Large Language Models like GPT-4 are transforming the world in general and the field of data science in particular. This two-hour training video presentation by Jon Krohn, Co-Founder and Chief Data Scientist at the machine learning company Nebula, introduces deep learning transformer architectures including LLMs.","rel":"","context":"In &quot;AI Deep Learning&quot;","img":{"alt_text":"","src":"https:\/\/i0.wp.com\/insidebigdata.com\/wp-content\/uploads\/2023\/06\/GenerativeAI_shutterstock_2313909647_special.jpg?resize=350%2C200&ssl=1","width":350,"height":200},"classes":[]},{"id":33978,"url":"https:\/\/insidebigdata.com\/2023\/11\/22\/insidebigdata-ai-news-briefs-11-22-2023\/","url_meta":{"origin":34133,"position":5},"title":"insideBIGDATA AI News Briefs \u2013 11\/22\/2023","date":"November 22, 2023","format":false,"excerpt":"Welcome insideBIGDATA AI News Briefs, our timely new feature bringing you the latest industry insights and perspectives surrounding the field of AI including deep learning, large language models, generative AI, and transformers. We\u2019re working tirelessly to dig up the most timely and curious tidbits underlying the day\u2019s most popular technologies.\u2026","rel":"","context":"In &quot;AI Deep Learning&quot;","img":{"alt_text":"","src":"https:\/\/i0.wp.com\/insidebigdata.com\/wp-content\/uploads\/2023\/07\/AI-News-Briefs-column-banner.png?resize=350%2C200&ssl=1","width":350,"height":200},"classes":[]}],"_links":{"self":[{"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/posts\/34133"}],"collection":[{"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/users\/37"}],"replies":[{"embeddable":true,"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/comments?post=34133"}],"version-history":[{"count":0,"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/posts\/34133\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/media\/32924"}],"wp:attachment":[{"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/media?parent=34133"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/categories?post=34133"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/tags?post=34133"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}