{"id":250,"date":"2023-09-08T20:58:20","date_gmt":"2023-09-08T20:58:20","guid":{"rendered":"https:\/\/techwyns.com\/tech\/?p=250"},"modified":"2023-09-08T21:01:34","modified_gmt":"2023-09-08T21:01:34","slug":"artificial-intelligence-opensource-alternatives-to-chatgpt","status":"publish","type":"post","link":"https:\/\/techwyns.com\/tech\/blog\/2023\/09\/08\/artificial-intelligence-opensource-alternatives-to-chatgpt\/","title":{"rendered":"Artificial intelligence &#8211;  OpenSource alternatives to ChatGPT"},"content":{"rendered":"<p>&nbsp;<\/p>\n<div class=\"elementor-element elementor-element-4853044 e-con-boxed e-con\" data-id=\"4853044\" data-element_type=\"container\" data-settings=\"{&quot;content_width&quot;:&quot;boxed&quot;}\">\n<div class=\"e-con-inner\">\n<div class=\"elementor-element elementor-element-35eab5d elementor-widget elementor-widget-heading\" data-id=\"35eab5d\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n<div class=\"elementor-widget-container\">\n<h2 class=\"elementor-heading-title elementor-size-default\">Top 9 Free GPT-3 Alternative AI models<\/h2>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<section class=\"elementor-section elementor-top-section elementor-element elementor-element-28dbbeb elementor-section-boxed elementor-section-height-default elementor-section-height-default\" data-id=\"28dbbeb\" data-element_type=\"section\">\n<div class=\"elementor-container elementor-column-gap-default\">\n<div class=\"elementor-column elementor-col-100 elementor-top-column elementor-element elementor-element-e19eba9\" data-id=\"e19eba9\" data-element_type=\"column\">\n<div class=\"elementor-widget-wrap elementor-element-populated\">\n<div class=\"elementor-element elementor-element-9620ceb elementor-widget elementor-widget-text-editor\" data-id=\"9620ceb\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n<div class=\"elementor-widget-container\">\n<p>Now that you\u2019ve got an idea of what is the technology we are talking about, let\u2019s move on to the\u00a0OpenAI GPT-3\u00a0competitors.<\/p>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/section>\n<div class=\"elementor-element elementor-element-064fcc9 e-con-boxed e-con\" data-id=\"064fcc9\" data-element_type=\"container\" data-settings=\"{&quot;content_width&quot;:&quot;boxed&quot;}\">\n<div class=\"e-con-inner\">\n<div class=\"elementor-element elementor-element-c906dae elementor-widget elementor-widget-heading\" data-id=\"c906dae\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n<div class=\"elementor-widget-container\">\n<h3 class=\"elementor-heading-title elementor-size-default\">OPT by Meta<\/h3>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<section class=\"elementor-section elementor-top-section elementor-element elementor-element-427e0ee elementor-section-boxed elementor-section-height-default elementor-section-height-default\" data-id=\"427e0ee\" data-element_type=\"section\">\n<div class=\"elementor-container elementor-column-gap-default\">\n<div class=\"elementor-column elementor-col-100 elementor-top-column elementor-element elementor-element-a3909fe\" data-id=\"a3909fe\" data-element_type=\"column\">\n<div class=\"elementor-widget-wrap elementor-element-populated\">\n<div class=\"elementor-element elementor-element-6842211 elementor-widget elementor-widget-text-editor\" data-id=\"6842211\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n<div class=\"elementor-widget-container\">\n<p>Another solid\u00a0GPT-3 open-source\u00a0alternative was released by Meta in May 2022.\u00a0Open Pretrained Transformer language\u00a0model (OPT for short) contains 175B parameters. OPT was trained on multiple public datasets, including The Pile and BookCorpus.<\/p>\n<p>Its main distinctive feature is that OPT combines both pretrained models and the source code for using or training them.<\/p>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/section>\n<p>&nbsp;<\/p>\n<div class=\"elementor-element elementor-element-9e206b9 e-con-boxed e-con\" data-id=\"9e206b9\" data-element_type=\"container\" data-settings=\"{&quot;content_width&quot;:&quot;boxed&quot;}\">\n<div class=\"e-con-inner\">\n<div class=\"elementor-element elementor-element-a4c3f5e elementor-widget elementor-widget-heading\" data-id=\"a4c3f5e\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n<div class=\"elementor-widget-container\">\n<h3 class=\"elementor-heading-title elementor-size-default\">AlexaTM by Amazon<\/h3>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<section class=\"elementor-section elementor-top-section elementor-element elementor-element-f634f4c elementor-section-boxed elementor-section-height-default elementor-section-height-default\" data-id=\"f634f4c\" data-element_type=\"section\">\n<div class=\"elementor-container elementor-column-gap-default\">\n<div class=\"elementor-column elementor-col-100 elementor-top-column elementor-element elementor-element-82e8671\" data-id=\"82e8671\" data-element_type=\"column\">\n<div class=\"elementor-widget-wrap elementor-element-populated\">\n<div class=\"elementor-element elementor-element-276d8aa elementor-widget elementor-widget-text-editor\" data-id=\"276d8aa\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n<div class=\"elementor-widget-container\">\n<p>On November 18, 2022, Amazon publicly released\u00a0AlexaTM\u00a020B, a large-scale multilingual sequence2sequence model. What\u2019s so special about it? It utilizes an\u00a0encoder-decoder architecture\u00a0and was trained on a combo of causal-language-modeling (CLM) and denoising tasks.<\/p>\n<p>Thanks to that, AlexaTM is a better few-shot learner than decoder-only models. As a result, it\u00a0performs better in 1-shot summarization and machine translation\u00a0tasks than Google\u2019s PaLM 540B. Also, in zero-shot testing, the model tops GPT-3 on SuperGlue and SQuADv2 datasets.<\/p>\n<p>As for less technical things, AlexaTM supports multiple languages (as its type implies), including English, Spanish, Arabic, German, Hindi, French, Japanese, Italian, Portuguese, and others.<\/p>\n<p>Altogether, this makes AlexaTM quite a notable opponent to all other LLMs, whether free or not.<\/p>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/section>\n<div class=\"elementor-element elementor-element-55c2854 e-con-boxed e-con\" data-id=\"55c2854\" data-element_type=\"container\" data-settings=\"{&quot;content_width&quot;:&quot;boxed&quot;}\">\n<div class=\"e-con-inner\">\n<div class=\"elementor-element elementor-element-b353a49 elementor-widget elementor-widget-heading\" data-id=\"b353a49\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n<div class=\"elementor-widget-container\">\n<h3 class=\"elementor-heading-title elementor-size-default\">GPT-J and GPT-NeoX by EleutherAI<\/h3>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<section class=\"elementor-section elementor-top-section elementor-element elementor-element-ebe6e1c elementor-section-boxed elementor-section-height-default elementor-section-height-default\" data-id=\"ebe6e1c\" data-element_type=\"section\">\n<div class=\"elementor-container elementor-column-gap-default\">\n<div class=\"elementor-column elementor-col-100 elementor-top-column elementor-element elementor-element-fb18794\" data-id=\"fb18794\" data-element_type=\"column\">\n<div class=\"elementor-widget-wrap elementor-element-populated\">\n<div class=\"elementor-element elementor-element-1dd4fe1 elementor-widget elementor-widget-text-editor\" data-id=\"1dd4fe1\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n<div class=\"elementor-widget-container\">\n<p>GPT-J\u00a0is a small 6B-parameter autoregressive model for text generation, completely free to use. It was trained on The Pile, a dataset with 22 subsets of more than 800 GB of English texts.<\/p>\n<p>Despite its small size, the model performs nearly the same as GPT-3 6.7B-param and is better than its predecessor, GPT-Neo. The latter had 2 versions, 1.3 and 2.7 billion, and in February 2022 grew into\u00a0GPT-NeoX, containing 20B parameters.<\/p>\n<p>Here\u2019s a quick overview of how GPT-J and GPT-NeoX perform compared to\u00a0OpenAI GPT-3\u00a0versions.<\/p>\n<\/div>\n<\/div>\n<div class=\"elementor-element elementor-element-3c3bae4 elementor-widget elementor-widget-image\" data-id=\"3c3bae4\" data-element_type=\"widget\" data-widget_type=\"image.default\">\n<div class=\"elementor-widget-container\"><picture><source srcset=\"https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Accuracy-standard-language-modeling-tasks-.png.webp 1920w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Accuracy-standard-language-modeling-tasks--300x169.png.webp 300w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Accuracy-standard-language-modeling-tasks--1024x575.png.webp 1024w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Accuracy-standard-language-modeling-tasks--768x432.png.webp 768w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Accuracy-standard-language-modeling-tasks--1536x863.png.webp 1536w\" type=\"image\/webp\" sizes=\"795px\" data-srcset=\"https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Accuracy-standard-language-modeling-tasks-.png.webp 1920w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Accuracy-standard-language-modeling-tasks--300x169.png.webp 300w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Accuracy-standard-language-modeling-tasks--1024x575.png.webp 1024w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Accuracy-standard-language-modeling-tasks--768x432.png.webp 768w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Accuracy-standard-language-modeling-tasks--1536x863.png.webp 1536w\" \/><img loading=\"lazy\" decoding=\"async\" class=\"attachment-full size-full wp-image-40681 lazyautosizes lazyloaded\" src=\"https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Accuracy-standard-language-modeling-tasks-.png\" sizes=\"auto, 795px\" srcset=\"https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Accuracy-standard-language-modeling-tasks-.png 1920w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Accuracy-standard-language-modeling-tasks--300x169.png 300w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Accuracy-standard-language-modeling-tasks--1024x575.png 1024w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Accuracy-standard-language-modeling-tasks--768x432.png 768w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Accuracy-standard-language-modeling-tasks--1536x863.png 1536w\" alt=\"\" width=\"1920\" height=\"1079\" data-eio=\"p\" data-src=\"https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Accuracy-standard-language-modeling-tasks-.png\" data-srcset=\"https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Accuracy-standard-language-modeling-tasks-.png 1920w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Accuracy-standard-language-modeling-tasks--300x169.png 300w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Accuracy-standard-language-modeling-tasks--1024x575.png 1024w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Accuracy-standard-language-modeling-tasks--768x432.png 768w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Accuracy-standard-language-modeling-tasks--1536x863.png 1536w\" data-sizes=\"auto\" \/><\/picture><\/div>\n<\/div>\n<div class=\"elementor-element elementor-element-ba47d49 elementor-widget elementor-widget-text-editor\" data-id=\"ba47d49\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n<div class=\"elementor-widget-container\">\n<p>As you might see, there\u2019s little to no difference in performance between open-source GPT-J and GPT-NeoX and paid GPT-3 models.<\/p>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/section>\n<div class=\"elementor-element elementor-element-2e83406 e-con-boxed e-con\" data-id=\"2e83406\" data-element_type=\"container\" data-settings=\"{&quot;content_width&quot;:&quot;boxed&quot;}\">\n<div class=\"e-con-inner\">\n<div class=\"elementor-element elementor-element-487d074 elementor-widget elementor-widget-heading\" data-id=\"487d074\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n<div class=\"elementor-widget-container\">\n<h3 class=\"elementor-heading-title elementor-size-default\">Jurassic-1 language model by AI21 labs<\/h3>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<section class=\"elementor-section elementor-top-section elementor-element elementor-element-f07d2ef elementor-section-boxed elementor-section-height-default elementor-section-height-default\" data-id=\"f07d2ef\" data-element_type=\"section\">\n<div class=\"elementor-container elementor-column-gap-default\">\n<div class=\"elementor-column elementor-col-100 elementor-top-column elementor-element elementor-element-7d5bb9e\" data-id=\"7d5bb9e\" data-element_type=\"column\">\n<div class=\"elementor-widget-wrap elementor-element-populated\">\n<div class=\"elementor-element elementor-element-9c3ea82 elementor-widget elementor-widget-text-editor\" data-id=\"9c3ea82\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n<div class=\"elementor-widget-container\">\n<p>Jurassic-1 is an autoregressive natural language processing (NLP) model, available in open beta for developers and researchers.<\/p>\n<p>Yet, it\u2019s not fully open-source, but upon registration, you get $90 credits for free. You can use those credits in the playground with the pre-designed templates for rephrasing, summarization, writing, chatting, drafting outlines, tweeting, coding, and more. What\u2019s more, you can create and train your custom models.<\/p>\n<p>Jurassic-1\u00a0might become quite a serious rival to GPT-3, as it consists of 2 parts: J1-Jumbo, trained on over 178B parameters, and J1-Large, consisting of 7B parameters. This already makes it 3B parameters more advanced than GPT-3 language model.<\/p>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/section>\n<div class=\"elementor-element elementor-element-6ea767a e-con-boxed e-con\" data-id=\"6ea767a\" data-element_type=\"container\" data-settings=\"{&quot;content_width&quot;:&quot;boxed&quot;}\">\n<div class=\"e-con-inner\">\n<div class=\"elementor-element elementor-element-b50a9b1 elementor-widget elementor-widget-heading\" data-id=\"b50a9b1\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n<div class=\"elementor-widget-container\">\n<h3 class=\"elementor-heading-title elementor-size-default\">CodeGen by Salesforce<\/h3>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<section class=\"elementor-section elementor-top-section elementor-element elementor-element-77afdd2 elementor-section-boxed elementor-section-height-default elementor-section-height-default\" data-id=\"77afdd2\" data-element_type=\"section\">\n<div class=\"elementor-container elementor-column-gap-default\">\n<div class=\"elementor-column elementor-col-100 elementor-top-column elementor-element elementor-element-b1d95a4\" data-id=\"b1d95a4\" data-element_type=\"column\">\n<div class=\"elementor-widget-wrap elementor-element-populated\">\n<div class=\"elementor-element elementor-element-e664c84 elementor-widget elementor-widget-text-editor\" data-id=\"e664c84\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n<div class=\"elementor-widget-container\">\n<p>One more open-source\u00a0GPT-3 alternative\u00a0you couldn\u2019t miss. As you might have already guessed from its name,\u00a0<a href=\"https:\/\/blog.salesforceairesearch.com\/codegen\/\">CodeGen<\/a>\u00a0is a large-scale language model that can write programs, based on plain textual prompts. The model relies on the concept of\u00a0<i>conversational AI<\/i>, which aims to unify human creative input with nearly unlimited capabilities of AI coding.<\/p>\n<\/div>\n<\/div>\n<div class=\"elementor-element elementor-element-0ef2647 elementor-blockquote--skin-border elementor-blockquote--button-color-official elementor-widget elementor-widget-blockquote\" data-id=\"0ef2647\" data-element_type=\"widget\" data-widget_type=\"blockquote.default\">\n<div class=\"elementor-widget-container\">\n<blockquote class=\"elementor-blockquote\">\n<p class=\"elementor-blockquote__content\">The future of coding is the intersection of human and computer languages \u2014 and conversational AI is the perfect bridge to connect the two<\/p>\n<footer><cite class=\"elementor-blockquote__author\">Silvio Savarese, EVP &amp; Chief Scientist, AI Research, Salesforce<\/cite><\/footer>\n<\/blockquote>\n<\/div>\n<\/div>\n<div class=\"elementor-element elementor-element-6292cf4 elementor-widget elementor-widget-text-editor\" data-id=\"6292cf4\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n<div class=\"elementor-widget-container\">\n<p>CodeGen release comes in three model types (NL, multi, and mono) of different sizes (350M, 2B, 6B, and 16B). Each model type is trained on diverse datasets:<\/p>\n<ul>\n<li aria-level=\"1\">NL models use The Pile.<\/li>\n<li aria-level=\"1\">Multi models are based on NL models and use a corpus with code in various programming languages.<\/li>\n<li aria-level=\"1\">Mono models are based on multi models and use a corpus with Python code.<\/li>\n<\/ul>\n<\/div>\n<\/div>\n<div class=\"elementor-element elementor-element-5927ce3 elementor-widget elementor-widget-text-editor\" data-id=\"5927ce3\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n<div class=\"elementor-widget-container\">\n<p>The most fascinating thing about CodeGen is that even people without any tech background can use it. Still, programming knowledge will help to achieve better and more elegant solutions, as AI isn\u2019t perfect yet.<\/p>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/section>\n<div class=\"elementor-element elementor-element-20cab97 e-con-boxed e-con\" data-id=\"20cab97\" data-element_type=\"container\" data-settings=\"{&quot;content_width&quot;:&quot;boxed&quot;}\">\n<div class=\"e-con-inner\">\n<div class=\"elementor-element elementor-element-a589b98 elementor-widget elementor-widget-heading\" data-id=\"a589b98\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n<div class=\"elementor-widget-container\">\n<h3 class=\"elementor-heading-title elementor-size-default\">Megatron-Turing NLG by NVIDIA and Microsoft<\/h3>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<section class=\"elementor-section elementor-top-section elementor-element elementor-element-986e752 elementor-section-boxed elementor-section-height-default elementor-section-height-default\" data-id=\"986e752\" data-element_type=\"section\">\n<div class=\"elementor-container elementor-column-gap-default\">\n<div class=\"elementor-column elementor-col-100 elementor-top-column elementor-element elementor-element-8279db3\" data-id=\"8279db3\" data-element_type=\"column\">\n<div class=\"elementor-widget-wrap elementor-element-populated\">\n<div class=\"elementor-element elementor-element-f4e9fdd elementor-widget elementor-widget-text-editor\" data-id=\"f4e9fdd\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n<div class=\"elementor-widget-container\">\n<p>This LLM is among the largest ones, as it has over 530B parameters.\u00a0Megatron-Turing NLG\u00a0(Natural Language Generation) is a result of collaboration between Microsoft and NVIDIA. To train the model, they used The Pile dataset and NVIDIA DGX SuperPOD-based Selene supercomputer.<\/p>\n<p>The\u00a0research\u00a0that was released in October 2021 found that Megatron-Turing NLG is especially good at PiQA dev set tasks and LAMBADA test set tasks. The model also predicts on average over 50% in zero-shot tests and improves those numbers in one- and four-shot tests.<\/p>\n<p>At the moment, Microsoft and NVIDIA offer early access to Megatron-Turing NGL and invite\u00a0other companies to\u00a0join them for research. Their main goal is to develop policies of responsible AI usage and eliminate wrong responses, toxicity, and bias in large language models.<\/p>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/section>\n<div class=\"elementor-element elementor-element-3135c1a e-con-boxed e-con\" data-id=\"3135c1a\" data-element_type=\"container\" data-settings=\"{&quot;content_width&quot;:&quot;boxed&quot;}\">\n<div class=\"e-con-inner\">\n<div class=\"elementor-element elementor-element-c95dac5 elementor-widget elementor-widget-heading\" data-id=\"c95dac5\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n<div class=\"elementor-widget-container\">\n<h3 class=\"elementor-heading-title elementor-size-default\">LaMDA by Google<\/h3>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<section class=\"elementor-section elementor-top-section elementor-element elementor-element-98e1bc7 elementor-section-boxed elementor-section-height-default elementor-section-height-default\" data-id=\"98e1bc7\" data-element_type=\"section\">\n<div class=\"elementor-container elementor-column-gap-default\">\n<div class=\"elementor-column elementor-col-100 elementor-top-column elementor-element elementor-element-dc3dd99\" data-id=\"dc3dd99\" data-element_type=\"column\">\n<div class=\"elementor-widget-wrap elementor-element-populated\">\n<div class=\"elementor-element elementor-element-3d8596f elementor-widget elementor-widget-text-editor\" data-id=\"3d8596f\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n<div class=\"elementor-widget-container\">\n<p>LaMDA is an autoregressive Language Model for Dialog Applications, with a decoder-only architecture. Except for chit-chatting on different topics, the model can also create lists and can be trained to talk on some domain-specific topics.<\/p>\n<p>Dialog models can easily scale and can cope with long-term dependencies. This means that they can take into account the previous context, not just the current input. Also, they support domain grounding.<\/p>\n<p>For instance, Google researchers preconditioned\u00a0LaMDA\u00a0on several rounds of role-specific dialogs so that it could recommend music. Here\u2019s\u00a0one of the results:<\/p>\n<\/div>\n<\/div>\n<div class=\"elementor-element elementor-element-8b4b58b elementor-widget elementor-widget-image\" data-id=\"8b4b58b\" data-element_type=\"widget\" data-widget_type=\"image.default\">\n<div class=\"elementor-widget-container\">\n<figure class=\"wp-caption\"><picture><source srcset=\"https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/lamda.jpg.webp 1552w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/lamda-300x195.jpg.webp 300w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/lamda-1024x666.jpg.webp 1024w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/lamda-768x500.jpg.webp 768w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/lamda-1536x1000.jpg.webp 1536w\" type=\"image\/webp\" sizes=\"795px\" data-srcset=\"https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/lamda.jpg.webp 1552w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/lamda-300x195.jpg.webp 300w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/lamda-1024x666.jpg.webp 1024w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/lamda-768x500.jpg.webp 768w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/lamda-1536x1000.jpg.webp 1536w\" \/><img loading=\"lazy\" decoding=\"async\" class=\"attachment-full size-full wp-image-40775 lazyautosizes lazyloaded\" src=\"https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/lamda.jpg\" sizes=\"auto, 795px\" srcset=\"https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/lamda.jpg 1552w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/lamda-300x195.jpg 300w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/lamda-1024x666.jpg 1024w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/lamda-768x500.jpg 768w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/lamda-1536x1000.jpg 1536w\" alt=\"\" width=\"1552\" height=\"1010\" data-eio=\"p\" data-src=\"https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/lamda.jpg\" data-srcset=\"https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/lamda.jpg 1552w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/lamda-300x195.jpg 300w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/lamda-1024x666.jpg 1024w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/lamda-768x500.jpg 768w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/lamda-1536x1000.jpg 1536w\" data-sizes=\"auto\" \/><\/picture><figcaption class=\"widget-image-caption wp-caption-text\">LaMDA music recommendations example<\/figcaption><\/figure>\n<\/div>\n<\/div>\n<div class=\"elementor-element elementor-element-7e3da93 elementor-widget elementor-widget-text-editor\" data-id=\"7e3da93\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n<div class=\"elementor-widget-container\">\n<p>LaMDA was trained on a 1.56T words dataset, which contained not only public dialog data but also other public texts. The biggest model version has 137B parameters.<\/p>\n<p>Google is making it public, but to access the model, you need to join the waitlist.<\/p>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/section>\n<div class=\"elementor-element elementor-element-172d682 e-con-boxed e-con\" data-id=\"172d682\" data-element_type=\"container\" data-settings=\"{&quot;content_width&quot;:&quot;boxed&quot;}\">\n<div class=\"e-con-inner\">\n<div class=\"elementor-element elementor-element-330accd elementor-widget elementor-widget-heading\" data-id=\"330accd\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n<div class=\"elementor-widget-container\">\n<h3 class=\"elementor-heading-title elementor-size-default\">BLOOM<\/h3>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<section class=\"elementor-section elementor-top-section elementor-element elementor-element-e57f08b elementor-section-boxed elementor-section-height-default elementor-section-height-default\" data-id=\"e57f08b\" data-element_type=\"section\">\n<div class=\"elementor-container elementor-column-gap-default\">\n<div class=\"elementor-column elementor-col-100 elementor-top-column elementor-element elementor-element-317af9b\" data-id=\"317af9b\" data-element_type=\"column\">\n<div class=\"elementor-widget-wrap elementor-element-populated\">\n<div class=\"elementor-element elementor-element-a8c97ef elementor-widget elementor-widget-text-editor\" data-id=\"a8c97ef\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n<div class=\"elementor-widget-container\">\n<p>The BLOOM autoregressive LLM was developed by multiple contributors through the\u00a0BigScience Workshop\u00a0as the\u00a0GPT-3 open-source\u00a0alternative. More than 1000 AI researchers joined the project, including specialists from Microsoft, NVIDIA, PyTorch, and others. The\u00a0BLOOM\u00a0is available to any individual or team of researchers who want to study the performance and behavior of the large language models and agree with the\u00a0model\u2019s licensing terms.<\/p>\n<p>The model was trained on 176B parameters from March to July 2022 and can cope with 46 languages and 13 programming languages. Also, it has smaller versions that contain fewer parameters.<\/p>\n<p>The BLOOM has a decoder-only architecture, as it was created based on Megatron-LM, the 8.3B-parameter predecessor of Megatron-Turing NLG.<\/p>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/section>\n<div class=\"elementor-element elementor-element-b2c67a0 e-con-boxed e-con\" data-id=\"b2c67a0\" data-element_type=\"container\" data-settings=\"{&quot;content_width&quot;:&quot;boxed&quot;}\">\n<div class=\"e-con-inner\">\n<div class=\"elementor-element elementor-element-77ab26e elementor-widget elementor-widget-heading\" data-id=\"77ab26e\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n<div class=\"elementor-widget-container\">\n<h3 class=\"elementor-heading-title elementor-size-default\">BERT by Google<\/h3>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<section class=\"elementor-section elementor-top-section elementor-element elementor-element-8447405 elementor-section-boxed elementor-section-height-default elementor-section-height-default\" data-id=\"8447405\" data-element_type=\"section\">\n<div class=\"elementor-container elementor-column-gap-default\">\n<div class=\"elementor-column elementor-col-100 elementor-top-column elementor-element elementor-element-038004d\" data-id=\"038004d\" data-element_type=\"column\">\n<div class=\"elementor-widget-wrap elementor-element-populated\">\n<div class=\"elementor-element elementor-element-301bf96 elementor-widget elementor-widget-text-editor\" data-id=\"301bf96\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n<div class=\"elementor-widget-container\">\n<p>BERT\u00a0(Bidirectional Encoder Representations from Transformers) is one of the oldest transformer language models, open-sourced in 2018 and pretrained on texts from Wikipedia. Since 2019, Google has been using it to better understand search intent and offer more relevant queries prediction.<\/p>\n<p>By nature, BERT is a bidirectional, unsupervised language representation. This means to continue the sentence, the model takes into account the previous context and the conditions that will follow it.<\/p>\n<p>When first introduced, BERT was\u00a0tested against other models\u00a0and showed quite superior results. For example, here\u2019s the performance of the model on the GLUE Test:<\/p>\n<\/div>\n<\/div>\n<div class=\"elementor-element elementor-element-336aeed elementor-widget elementor-widget-image\" data-id=\"336aeed\" data-element_type=\"widget\" data-widget_type=\"image.default\">\n<div class=\"elementor-widget-container\">\n<figure class=\"wp-caption\"><picture><source srcset=\"https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/bert-model.png.webp 1408w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/bert-model-300x67.png.webp 300w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/bert-model-1024x228.png.webp 1024w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/bert-model-768x171.png.webp 768w\" type=\"image\/webp\" sizes=\"795px\" data-srcset=\"https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/bert-model.png.webp 1408w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/bert-model-300x67.png.webp 300w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/bert-model-1024x228.png.webp 1024w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/bert-model-768x171.png.webp 768w\" \/><img loading=\"lazy\" decoding=\"async\" class=\"attachment-full size-full wp-image-40686 lazyautosizes lazyloaded\" src=\"https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/bert-model.png\" sizes=\"auto, 795px\" srcset=\"https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/bert-model.png 1408w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/bert-model-300x67.png 300w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/bert-model-1024x228.png 1024w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/bert-model-768x171.png 768w\" alt=\"\" width=\"1408\" height=\"314\" data-eio=\"p\" data-src=\"https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/bert-model.png\" data-srcset=\"https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/bert-model.png 1408w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/bert-model-300x67.png 300w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/bert-model-1024x228.png 1024w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/bert-model-768x171.png 768w\" data-sizes=\"auto\" \/><\/picture><figcaption class=\"widget-image-caption wp-caption-text\">GLUE Test results for Pre-OpenAI SOTA, BiLSTM+ELMo+Attn, and OpenAI GPT, BERTBASE, and BERTLARGE (numbers under task names mean the number of training examples)<\/figcaption><\/figure>\n<\/div>\n<\/div>\n<div class=\"elementor-element elementor-element-ff91c39 elementor-widget elementor-widget-text-editor\" data-id=\"ff91c39\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n<div class=\"elementor-widget-container\">\n<p>BERT can be used as a technique for training diverse state-of-the-art (SOTA) NLP models, like question-answer systems, etc.<\/p>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/section>\n<div class=\"elementor-element elementor-element-7768891 e-con-boxed e-con\" data-id=\"7768891\" data-element_type=\"container\" data-settings=\"{&quot;content_width&quot;:&quot;boxed&quot;}\">\n<div class=\"e-con-inner\">\n<div class=\"elementor-element elementor-element-88bb7e4 elementor-widget elementor-widget-heading\" data-id=\"88bb7e4\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n<div class=\"elementor-widget-container\">\n<h3 class=\"elementor-heading-title elementor-size-default\">[Bonus] 3 Additional GPT-3 Alternative Models Worth Attention<\/h3>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<section class=\"elementor-section elementor-top-section elementor-element elementor-element-b74c0ba elementor-section-boxed elementor-section-height-default elementor-section-height-default\" data-id=\"b74c0ba\" data-element_type=\"section\">\n<div class=\"elementor-container elementor-column-gap-default\">\n<div class=\"elementor-column elementor-col-100 elementor-top-column elementor-element elementor-element-cbd2e79\" data-id=\"cbd2e79\" data-element_type=\"column\">\n<div class=\"elementor-widget-wrap elementor-element-populated\">\n<div class=\"elementor-element elementor-element-bb14038 elementor-widget elementor-widget-text-editor\" data-id=\"bb14038\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n<div class=\"elementor-widget-container\">\n<p>These models are not available to the public yet, but look quite promising. Let\u2019s overview what makes them stand out among the competitors.<\/p>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/section>\n<div class=\"elementor-element elementor-element-8de2381 e-con-boxed e-con\" data-id=\"8de2381\" data-element_type=\"container\" data-settings=\"{&quot;content_width&quot;:&quot;boxed&quot;}\">\n<div class=\"e-con-inner\">\n<div class=\"elementor-element elementor-element-fa0e864 elementor-widget elementor-widget-heading\" data-id=\"fa0e864\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n<div class=\"elementor-widget-container\">\n<h3 class=\"elementor-heading-title elementor-size-default\">GLaM by Google<\/h3>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<section class=\"elementor-section elementor-top-section elementor-element elementor-element-f4e879f elementor-section-boxed elementor-section-height-default elementor-section-height-default\" data-id=\"f4e879f\" data-element_type=\"section\">\n<div class=\"elementor-container elementor-column-gap-default\">\n<div class=\"elementor-column elementor-col-100 elementor-top-column elementor-element elementor-element-95a86d9\" data-id=\"95a86d9\" data-element_type=\"column\">\n<div class=\"elementor-widget-wrap elementor-element-populated\">\n<div class=\"elementor-element elementor-element-1bfd04f elementor-widget elementor-widget-text-editor\" data-id=\"1bfd04f\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n<div class=\"elementor-widget-container\">\n<p>GLaM\u00a0is the Generalist Language Model, developed by Google. It was introduced in December 2021 and has 1.2T of parameters, which makes it one of the largest existing models. Though Google hasn\u2019t provided public access to its source code, the model itself is noteworthy.<\/p>\n<p>Its key peculiarity is that it is a mixture of experts model (MoE). It consists of multiple layers or submodels (which are called experts), specializing in different domains. Depending on the input data, a gating network picks the most relevant experts (normally, two for each word or its part). Yet, this means the model doesn\u2019t use its capacity to the fullest; it usually activates around 97B of parameters during inference.<\/p>\n<p>Zero-shot and one-shot testing against 29 public NLP benchmarks in natural language processing (NLG) and natural language understanding (NLU) has shown that GLaM prevails over GPT-3.<\/p>\n<p>The tests included natural language interference (NLI), Winograd-style tasks, in-context reading comprehension, commonsense reasoning, open-domain questions answering, and others.<\/p>\n<p>Here are the results of the evaluation:<\/p>\n<\/div>\n<\/div>\n<div class=\"elementor-element elementor-element-a5c11a8 elementor-widget elementor-widget-image\" data-id=\"a5c11a8\" data-element_type=\"widget\" data-widget_type=\"image.default\">\n<div class=\"elementor-widget-container\">\n<figure class=\"wp-caption\"><picture><source srcset=\"https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Average-score-for-GLaM-GPT-3-on-NLG-and-NLU-task--1024x683.png.webp 1024w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Average-score-for-GLaM-GPT-3-on-NLG-and-NLU-task--300x200.png.webp 300w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Average-score-for-GLaM-GPT-3-on-NLG-and-NLU-task--768x512.png.webp 768w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Average-score-for-GLaM-GPT-3-on-NLG-and-NLU-task--1536x1024.png.webp 1536w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Average-score-for-GLaM-GPT-3-on-NLG-and-NLU-task-.png.webp 1920w\" type=\"image\/webp\" sizes=\"795px\" data-srcset=\"https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Average-score-for-GLaM-GPT-3-on-NLG-and-NLU-task--1024x683.png.webp 1024w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Average-score-for-GLaM-GPT-3-on-NLG-and-NLU-task--300x200.png.webp 300w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Average-score-for-GLaM-GPT-3-on-NLG-and-NLU-task--768x512.png.webp 768w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Average-score-for-GLaM-GPT-3-on-NLG-and-NLU-task--1536x1024.png.webp 1536w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Average-score-for-GLaM-GPT-3-on-NLG-and-NLU-task-.png.webp 1920w\" \/><img loading=\"lazy\" decoding=\"async\" class=\"attachment-large size-large wp-image-40679 lazyautosizes lazyloaded\" src=\"https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Average-score-for-GLaM-GPT-3-on-NLG-and-NLU-task--1024x683.png\" sizes=\"auto, 795px\" srcset=\"https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Average-score-for-GLaM-GPT-3-on-NLG-and-NLU-task--1024x683.png 1024w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Average-score-for-GLaM-GPT-3-on-NLG-and-NLU-task--300x200.png 300w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Average-score-for-GLaM-GPT-3-on-NLG-and-NLU-task--768x512.png 768w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Average-score-for-GLaM-GPT-3-on-NLG-and-NLU-task--1536x1024.png 1536w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Average-score-for-GLaM-GPT-3-on-NLG-and-NLU-task-.png 1920w\" alt=\"\" width=\"800\" height=\"534\" data-eio=\"p\" data-src=\"https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Average-score-for-GLaM-GPT-3-on-NLG-and-NLU-task--1024x683.png\" data-srcset=\"https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Average-score-for-GLaM-GPT-3-on-NLG-and-NLU-task--1024x683.png 1024w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Average-score-for-GLaM-GPT-3-on-NLG-and-NLU-task--300x200.png 300w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Average-score-for-GLaM-GPT-3-on-NLG-and-NLU-task--768x512.png 768w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Average-score-for-GLaM-GPT-3-on-NLG-and-NLU-task--1536x1024.png 1536w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Average-score-for-GLaM-GPT-3-on-NLG-and-NLU-task-.png 1920w\" data-sizes=\"auto\" \/><\/picture><figcaption class=\"widget-image-caption wp-caption-text\">Results of GLaM and GPT-3 completing NGL and NLU tasks<\/figcaption><\/figure>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/section>\n<div class=\"elementor-element elementor-element-cf215f9 e-con-boxed e-con\" data-id=\"cf215f9\" data-element_type=\"container\" data-settings=\"{&quot;content_width&quot;:&quot;boxed&quot;}\">\n<div class=\"e-con-inner\">\n<div class=\"elementor-element elementor-element-98b840b elementor-widget elementor-widget-heading\" data-id=\"98b840b\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n<div class=\"elementor-widget-container\">\n<h3 class=\"elementor-heading-title elementor-size-default\">Wu Dao 2.0<\/h3>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<section class=\"elementor-section elementor-top-section elementor-element elementor-element-629f28f elementor-section-boxed elementor-section-height-default elementor-section-height-default\" data-id=\"629f28f\" data-element_type=\"section\">\n<div class=\"elementor-container elementor-column-gap-default\">\n<div class=\"elementor-column elementor-col-100 elementor-top-column elementor-element elementor-element-c5c8aa0\" data-id=\"c5c8aa0\" data-element_type=\"column\">\n<div class=\"elementor-widget-wrap elementor-element-populated\">\n<div class=\"elementor-element elementor-element-bf36623 elementor-widget elementor-widget-text-editor\" data-id=\"bf36623\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n<div class=\"elementor-widget-container\">\n<p>Wu Dao\u00a0(which translates from Chinese as \u201croad to awareness\u201d) is a pretrained multimodal and multitasking deep learning model, developed by the Beijing Academy of Artificial Intelligence (BAAI). They claim it to be the world\u2019s largest transformer, with 1.75 trillion parameters. The first version was released in 2021, and the latest came out in May 2022.<\/p>\n<p>Wu Dao was trained in English, using The Pile, and in Chinese on a specifically designed dataset that contains around 3.7 terabytes of text and images. Thus, it can process language, generate texts, recognize and generate images, as well as create pictures based on textual prompts. The model has an MoE architecture, like Google GLaM.<\/p>\n<p>BAAI already partnered with such Chinese giants as Xiaomi Corporation and Kuaishou Technology (the owner of the short video social network).<\/p>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/section>\n<div class=\"elementor-element elementor-element-2b7cbf4 e-con-boxed e-con\" data-id=\"2b7cbf4\" data-element_type=\"container\" data-settings=\"{&quot;content_width&quot;:&quot;boxed&quot;}\">\n<div class=\"e-con-inner\">\n<div class=\"elementor-element elementor-element-f31ee21 elementor-widget elementor-widget-heading\" data-id=\"f31ee21\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n<div class=\"elementor-widget-container\">\n<h3 class=\"elementor-heading-title elementor-size-default\">Chinchilla by DeepMind<\/h3>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<section class=\"elementor-section elementor-top-section elementor-element elementor-element-0c62684 elementor-section-boxed elementor-section-height-default elementor-section-height-default\" data-id=\"0c62684\" data-element_type=\"section\">\n<div class=\"elementor-container elementor-column-gap-default\">\n<div class=\"elementor-column elementor-col-100 elementor-top-column elementor-element elementor-element-48ab068\" data-id=\"48ab068\" data-element_type=\"column\">\n<div class=\"elementor-widget-wrap elementor-element-populated\">\n<div class=\"elementor-element elementor-element-f19c4d6 elementor-widget elementor-widget-text-editor\" data-id=\"f19c4d6\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n<div class=\"elementor-widget-container\">\n<p>Chinchilla\u00a0is a recent compute-optimal language model, introduced in March 2022 by DeepMind, an AI lab, acquired by Google in 2014.<\/p>\n<p>The model itself is only 70 billion parameters in size, but it was trained on 1.4 trillion tokens (text data), which is 4x more compared to the most popular LLMs:<\/p>\n<\/div>\n<\/div>\n<div class=\"elementor-element elementor-element-b1307db elementor-widget elementor-widget-image\" data-id=\"b1307db\" data-element_type=\"widget\" data-widget_type=\"image.default\">\n<div class=\"elementor-widget-container\"><picture><source srcset=\"https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Chinchilla-DeepMind-1024x891.png.webp 1024w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Chinchilla-DeepMind-300x261.png.webp 300w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Chinchilla-DeepMind-768x668.png.webp 768w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Chinchilla-DeepMind-1536x1336.png.webp 1536w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Chinchilla-DeepMind.png.webp 1920w\" type=\"image\/webp\" sizes=\"795px\" data-srcset=\"https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Chinchilla-DeepMind-1024x891.png.webp 1024w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Chinchilla-DeepMind-300x261.png.webp 300w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Chinchilla-DeepMind-768x668.png.webp 768w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Chinchilla-DeepMind-1536x1336.png.webp 1536w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Chinchilla-DeepMind.png.webp 1920w\" \/><img loading=\"lazy\" decoding=\"async\" class=\"attachment-large size-large wp-image-40680 lazyautosizes lazyloaded\" src=\"https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Chinchilla-DeepMind-1024x891.png\" sizes=\"auto, 795px\" srcset=\"https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Chinchilla-DeepMind-1024x891.png 1024w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Chinchilla-DeepMind-300x261.png 300w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Chinchilla-DeepMind-768x668.png 768w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Chinchilla-DeepMind-1536x1336.png 1536w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Chinchilla-DeepMind.png 1920w\" alt=\"\" width=\"800\" height=\"696\" data-eio=\"p\" data-src=\"https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Chinchilla-DeepMind-1024x891.png\" data-srcset=\"https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Chinchilla-DeepMind-1024x891.png 1024w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Chinchilla-DeepMind-300x261.png 300w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Chinchilla-DeepMind-768x668.png 768w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Chinchilla-DeepMind-1536x1336.png 1536w, https:\/\/www.altamira.ai\/wp-content\/uploads\/2023\/01\/Chinchilla-DeepMind.png 1920w\" data-sizes=\"auto\" \/><\/picture><\/div>\n<\/div>\n<div class=\"elementor-element elementor-element-864c0cf elementor-widget elementor-widget-text-editor\" data-id=\"864c0cf\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n<div class=\"elementor-widget-container\">\n<p>Chinchilla\u00a0is a recent compute-optimal language model, introduced in March 2022 by DeepMind, an AI lab, acquired by Google in 2014.<\/p>\n<p>The model itself is only 70 billion parameters in size, but it was trained on 1.4 trillion tokens (text data), which is 4x more compared to the most popular LLMs:<\/p>\n<\/div>\n<\/div>\n<div class=\"elementor-element elementor-element-5394a94 elementor-widget elementor-widget-text-editor\" data-id=\"5394a94\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n<div class=\"elementor-widget-container\">\n<p><b>1.Language modeling.\u00a0<\/b>Chinchilla from 0.02 to 0.10 bits-per-byte improvement in different evaluation subsets of The Pile, compared to Gopher (another DeepMind\u2019s language model).<\/p>\n<\/div>\n<\/div>\n<div class=\"elementor-element elementor-element-9c00dd1 elementor-widget elementor-widget-text-editor\" data-id=\"9c00dd1\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n<div class=\"elementor-widget-container\">\n<p><b>2. MMLU (Massive Multitask Language Understanding).<\/b>\u00a0Chinchilla achieves 67.3% accuracy after 5 shots, while Gopher\u201460%, and GPT-3\u2014only 43.9%<\/p>\n<\/div>\n<\/div>\n<div class=\"elementor-element elementor-element-71f1477 elementor-widget elementor-widget-text-editor\" data-id=\"71f1477\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n<div class=\"elementor-widget-container\">\n<p><b>3.Reading comprehension.<\/b>\u00a0Chinchilla demonstrates an accuracy of 77.4% for predicting the final word of the sentence in the LAMBADA dataset, MT-NGL 530B\u201476.6%, and Gopher\u201474.5%.<\/p>\n<\/div>\n<\/div>\n<div class=\"elementor-element elementor-element-f871e24 elementor-widget elementor-widget-text-editor\" data-id=\"f871e24\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n<div class=\"elementor-widget-container\">\n<p>Chinchilla proves it\u2019s the number of training tokens, not the size of parameters, that defines high performance. This discovery potentially opens an opportunity for other models to scale through the amount of info they are trained on, rather than via the number of parameters.<\/p>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/section>\n<div class=\"elementor-element elementor-element-df58d54 e-con-boxed e-con\" data-id=\"df58d54\" data-element_type=\"container\" data-settings=\"{&quot;content_width&quot;:&quot;boxed&quot;}\">\n<div class=\"e-con-inner\">\n<div class=\"elementor-element elementor-element-99ac546 elementor-widget elementor-widget-heading\" data-id=\"99ac546\" data-element_type=\"widget\" data-widget_type=\"heading.default\">\n<div class=\"elementor-widget-container\">\n<h3 class=\"elementor-heading-title elementor-size-default\">To Sum It Up<\/h3>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<section class=\"elementor-section elementor-top-section elementor-element elementor-element-bf2781f elementor-section-boxed elementor-section-height-default elementor-section-height-default\" data-id=\"bf2781f\" data-element_type=\"section\">\n<div class=\"elementor-container elementor-column-gap-default\">\n<div class=\"elementor-column elementor-col-100 elementor-top-column elementor-element elementor-element-cce4660\" data-id=\"cce4660\" data-element_type=\"column\">\n<div class=\"elementor-widget-wrap elementor-element-populated\">\n<div class=\"elementor-element elementor-element-d13fc54 elementor-widget elementor-widget-text-editor\" data-id=\"d13fc54\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n<div class=\"elementor-widget-container\">\n<p>For now, we can observe a variety of\u00a0best AI tools\u00a0and a breakthrough they make in natural language processing, understanding, and generation. We\u2019ll definitely see even more models of different types coming up in the nearest future.<\/p>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/section>\n<p>&nbsp;<\/p>\n<p>&nbsp;<\/p>\n<p><strong>Here are the alternatives to ChatGPT<\/strong><\/p>\n<ul class=\"scroll-wrapper__list\">\n<li class=\"list__title\">Bing with ChatGPT<\/li>\n<li class=\"list__title\">Google Bard<\/li>\n<li class=\"list__title\">YouChat<\/li>\n<li class=\"list__title\">Auto-GPT<\/li>\n<li class=\"list__title active\">StableLM<\/li>\n<li class=\"list__title\">CatGPT<\/li>\n<li class=\"list__title\">Poe<\/li>\n<\/ul>\n<h2 id=\"section-auto-gpt\" class=\"article-body__section\">AUTO-GPT<\/h2>\n<figure class=\"van-image-figure inline-layout\" data-bordeaux-image-check=\"\">\n<div class=\"image-full-width-wrapper\">\n<div class=\"image-widthsetter\">\n<p class=\"vanilla-image-block\"><picture><source class=\" lazy-image-van\" srcset=\"https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-320-80.jpg.webp 320w, https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-480-80.jpg.webp 480w, https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-650-80.jpg.webp 650w, https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-970-80.jpg.webp 970w, https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-1024-80.jpg.webp 1024w, https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-1200-80.jpg.webp 1200w\" type=\"image\/webp\" sizes=\"(min-width: 1000px) 970px, calc(100vw - 40px)\" data-normal=\"https:\/\/vanilla.futurecdn.net\/tomsguide\/media\/img\/missing-image.svg\" data-srcset=\"https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-320-80.jpg.webp 320w, https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-480-80.jpg.webp 480w, https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-650-80.jpg.webp 650w, https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-970-80.jpg.webp 970w, https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-1024-80.jpg.webp 1024w, https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-1200-80.jpg.webp 1200w\" data-sizes=\"(min-width: 1000px) 970px, calc(100vw - 40px)\" data-original-mos=\"https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog.jpg\" data-pin-media=\"https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog.jpg\" \/><source class=\" lazy-image-van\" srcset=\"https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-320-80.jpg 320w, https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-480-80.jpg 480w, https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-650-80.jpg 650w, https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-970-80.jpg 970w, https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-1024-80.jpg 1024w, https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-1200-80.jpg 1200w\" type=\"image\/jpeg\" sizes=\"(min-width: 1000px) 970px, calc(100vw - 40px)\" data-normal=\"https:\/\/vanilla.futurecdn.net\/tomsguide\/media\/img\/missing-image.svg\" data-srcset=\"https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-320-80.jpg 320w, https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-480-80.jpg 480w, https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-650-80.jpg 650w, https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-970-80.jpg 970w, https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-1024-80.jpg 1024w, https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-1200-80.jpg 1200w\" data-sizes=\"(min-width: 1000px) 970px, calc(100vw - 40px)\" data-original-mos=\"https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog.jpg\" data-pin-media=\"https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog.jpg\" \/><img decoding=\"async\" class=\"lazy-image-van loaded\" src=\"https:\/\/vanilla.futurecdn.net\/tomsguide\/media\/img\/missing-image.svg\" sizes=\"(min-width: 1000px) 970px, calc(100vw - 40px)\" srcset=\"https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-320-80.jpg 320w, https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-480-80.jpg 480w, https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-650-80.jpg 650w, https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-970-80.jpg 970w, https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-1024-80.jpg 1024w, https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-1200-80.jpg 1200w\" alt=\"AgentGPT\" data-normal=\"https:\/\/vanilla.futurecdn.net\/tomsguide\/media\/img\/missing-image.svg\" data-srcset=\"https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-320-80.jpg 320w, https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-480-80.jpg 480w, https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-650-80.jpg 650w, https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-970-80.jpg 970w, https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-1024-80.jpg 1024w, https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog-1200-80.jpg 1200w\" data-sizes=\"(min-width: 1000px) 970px, calc(100vw - 40px)\" data-original-mos=\"https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog.jpg\" data-pin-media=\"https:\/\/cdn.mos.cms.futurecdn.net\/c4NTYfDZJHegnC4qK9ypog.jpg\" data-ll-status=\"loaded\" \/><\/picture><\/p>\n<\/div>\n<\/div><figcaption class=\" inline-layout\"><span class=\"credit\">(Image credit: AgentGPT)<\/span><\/figcaption><\/figure>\n<p><u>Auto-GPT<\/u>\u00a0is a really cool ChatGPT variant, but it takes a bit of coding skill to work. At first glance, it is very similar to ChatGPT, and in fact, it runs on OpenAI\u2019s GPT-4 LLM. But Auto-GPT is semi-autonomous, which is a game-changing feature.<\/p>\n<p>With traditional ChatGPT, you have to do the work of prompting the AI. Say you are trying to build a business plan for a restaurant \u2014 you ask ChatGPT the prompt, it gives you a response and then you ask follow-up prompts to fine-tune the plan. But with Auto-GPT, you just set the chatbot a goal of developing the business plan, and then the chatbot will handle setting all the tasks, asking and answering follow-up prompts, etc. It eliminates a significant amount of the work you need to do.<\/p>\n<p>The catch? You need to know how to code with Python. Auto-GPT relies on a Python environment to run. You also need to set up a\u00a0<u>ChatGPT API<\/u>\u00a0account with OpenAI to connect the GPT-4 LLM to the Python environment. So it\u2019s not a simple chatbot to use, unlike the previous three we\u2019ve discussed.<\/p>\n<p>However, there is a way to try out Auto-GPT if you\u2019re not a coding expert.\u00a0<a class=\"hawk-link-parsed\" href=\"https:\/\/agentgpt.reworkd.ai\/\" target=\"_blank\" rel=\"noopener\" data-url=\"https:\/\/agentgpt.reworkd.ai\/\" data-component-tracked=\"1\"><u>AgentGPT<\/u><\/a>\u00a0is a free beta that already has the Python environment set up and is connected to the GPT-4 LLM. You can\u2019t do a ton with it \u2014 the beta version limits you too how much time you can spend with Auto-GPT \u2014 but it still is a great way to try out Auto-GPT. And if you have a ChatGPT API key, you can connect it to AgentGPT and have it use your API key instead of the beta\u2019s API key. We haven\u2019t tried that feature yet though, so proceed at your own risk.<\/p>\n<h2 id=\"section-stablelm\" class=\"article-body__section\">STABLELM<\/h2>\n<figure class=\"van-image-figure inline-layout\" data-bordeaux-image-check=\"\">\n<div class=\"image-full-width-wrapper\">\n<div class=\"image-widthsetter\">\n<p class=\"vanilla-image-block\"><picture><source class=\" lazy-image-van\" srcset=\"https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-320-80.jpeg.webp 320w, https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-480-80.jpeg.webp 480w, https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-650-80.jpeg.webp 650w, https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-970-80.jpeg.webp 970w, https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-1024-80.jpeg.webp 1024w, https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-1200-80.jpeg.webp 1200w\" type=\"image\/webp\" sizes=\"(min-width: 1000px) 970px, calc(100vw - 40px)\" data-normal=\"https:\/\/vanilla.futurecdn.net\/tomsguide\/media\/img\/missing-image.svg\" data-srcset=\"https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-320-80.jpeg.webp 320w, https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-480-80.jpeg.webp 480w, https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-650-80.jpeg.webp 650w, https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-970-80.jpeg.webp 970w, https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-1024-80.jpeg.webp 1024w, https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-1200-80.jpeg.webp 1200w\" data-sizes=\"(min-width: 1000px) 970px, calc(100vw - 40px)\" data-original-mos=\"https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ.jpeg\" data-pin-media=\"https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ.jpeg\" \/><source class=\" lazy-image-van\" srcset=\"https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-320-80.jpeg 320w, https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-480-80.jpeg 480w, https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-650-80.jpeg 650w, https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-970-80.jpeg 970w, https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-1024-80.jpeg 1024w, https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-1200-80.jpeg 1200w\" type=\"image\/jpeg\" sizes=\"(min-width: 1000px) 970px, calc(100vw - 40px)\" data-normal=\"https:\/\/vanilla.futurecdn.net\/tomsguide\/media\/img\/missing-image.svg\" data-srcset=\"https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-320-80.jpeg 320w, https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-480-80.jpeg 480w, https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-650-80.jpeg 650w, https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-970-80.jpeg 970w, https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-1024-80.jpeg 1024w, https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-1200-80.jpeg 1200w\" data-sizes=\"(min-width: 1000px) 970px, calc(100vw - 40px)\" data-original-mos=\"https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ.jpeg\" data-pin-media=\"https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ.jpeg\" \/><img decoding=\"async\" class=\"lazy-image-van loaded\" src=\"https:\/\/vanilla.futurecdn.net\/tomsguide\/media\/img\/missing-image.svg\" sizes=\"(min-width: 1000px) 970px, calc(100vw - 40px)\" srcset=\"https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-320-80.jpeg 320w, https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-480-80.jpeg 480w, https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-650-80.jpeg 650w, https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-970-80.jpeg 970w, https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-1024-80.jpeg 1024w, https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-1200-80.jpeg 1200w\" alt=\"StableLM Alpha on Hugging Face\" data-normal=\"https:\/\/vanilla.futurecdn.net\/tomsguide\/media\/img\/missing-image.svg\" data-srcset=\"https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-320-80.jpeg 320w, https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-480-80.jpeg 480w, https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-650-80.jpeg 650w, https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-970-80.jpeg 970w, https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-1024-80.jpeg 1024w, https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ-1200-80.jpeg 1200w\" data-sizes=\"(min-width: 1000px) 970px, calc(100vw - 40px)\" data-original-mos=\"https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ.jpeg\" data-pin-media=\"https:\/\/cdn.mos.cms.futurecdn.net\/bcJ9RtRe89QffobAREJRBQ.jpeg\" data-ll-status=\"loaded\" \/><\/picture><\/p>\n<\/div>\n<\/div><figcaption class=\" inline-layout\"><span class=\"credit\">(Image credit: Hugging Face)<\/span><\/figcaption><\/figure>\n<p>Stablity AI is another player in the AI space that competes with Open AI. Its biggest success is the Stable Diffusion\u00a0<u>AI image generator<\/u>, but it has now launched an open source LLM called\u00a0<u>StableLM<\/u>. You can try it out over at\u00a0<u>Hugging Face<\/u>, an AI community site that has a free Alpha test of StableLM running that anyone can try.<\/p>\n<p><u>Google Bard<\/u>\u00a0is Google\u2019s response to ChatGPT, which appears to have caught the search giant totally off guard. Bard uses a combination of two LLMs \u2014 Language Model for Dialogue Applications (LaMDA) and Pathways Language Model (PaLM). PaLM in particular\u00a0<u>gives Bard a boost<\/u>, bringing improved math and logic capabilities to the AI chatbot.<\/p>\n<p>This chatbot is similar to Bing with ChatGPT and the ChatGPT you can access on Open AI\u2019s site, but its features are a bit more limited. However, Google is regularly updating Bard&#8217;s features through &#8220;Experiment updates&#8221; and has even upgraded Bard so that the AI chatbot\u00a0can now write code. It is also decent as a research tool that you can hold a conversation with, even if it\u00a0<u>sometimes gets things wrong<\/u>. We\u2019re hopeful that its feature set will continue to expand as time goes on.<\/p>\n<p>We even compared\u00a0<u>Bing with ChatGPT versus Google Bard<\/u>\u00a0and the results were surprisingly close. While Bing\u2019s chatbot was better at certain things and was more accurate, Bard held its own. We also asked it to answer\u00a0<u>five controversial sci-fi questions<\/u>\u00a0and again, it did a surprisingly good job.<\/p>\n<p>So if you want to try out Google Bard, make sure you get on the\u00a0<u>Bard waitlist<\/u>. And use our guide on\u00a0<u>how to use Google Bard<\/u>\u00a0to get the most out of the chatbot.<\/p>\n<h2 id=\"section-youchat\" class=\"article-body__section\">YOUCHAT<\/h2>\n<figure class=\"van-image-figure inline-layout\" data-bordeaux-image-check=\"\">\n<div class=\"image-full-width-wrapper\">\n<div class=\"image-widthsetter\">\n<p class=\"vanilla-image-block\"><picture><source class=\" lazy-image-van\" srcset=\"https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-320-80.png.webp 320w, https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-480-80.png.webp 480w, https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-650-80.png.webp 650w, https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-970-80.png.webp 970w, https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-1024-80.png.webp 1024w, https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-1200-80.png.webp 1200w\" type=\"image\/webp\" sizes=\"(min-width: 1000px) 970px, calc(100vw - 40px)\" data-normal=\"https:\/\/vanilla.futurecdn.net\/tomsguide\/media\/img\/missing-image.svg\" data-srcset=\"https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-320-80.png.webp 320w, https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-480-80.png.webp 480w, https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-650-80.png.webp 650w, https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-970-80.png.webp 970w, https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-1024-80.png.webp 1024w, https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-1200-80.png.webp 1200w\" data-sizes=\"(min-width: 1000px) 970px, calc(100vw - 40px)\" data-original-mos=\"https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7.png\" data-pin-media=\"https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7.png\" \/><source class=\" lazy-image-van\" srcset=\"https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-320-80.png 320w, https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-480-80.png 480w, https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-650-80.png 650w, https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-970-80.png 970w, https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-1024-80.png 1024w, https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-1200-80.png 1200w\" type=\"image\/png\" sizes=\"(min-width: 1000px) 970px, calc(100vw - 40px)\" data-normal=\"https:\/\/vanilla.futurecdn.net\/tomsguide\/media\/img\/missing-image.svg\" data-srcset=\"https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-320-80.png 320w, https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-480-80.png 480w, https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-650-80.png 650w, https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-970-80.png 970w, https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-1024-80.png 1024w, https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-1200-80.png 1200w\" data-sizes=\"(min-width: 1000px) 970px, calc(100vw - 40px)\" data-original-mos=\"https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7.png\" data-pin-media=\"https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7.png\" \/><img decoding=\"async\" class=\"lazy-image-van loaded\" src=\"https:\/\/vanilla.futurecdn.net\/tomsguide\/media\/img\/missing-image.svg\" sizes=\"(min-width: 1000px) 970px, calc(100vw - 40px)\" srcset=\"https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-320-80.png 320w, https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-480-80.png 480w, https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-650-80.png 650w, https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-970-80.png 970w, https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-1024-80.png 1024w, https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-1200-80.png 1200w\" alt=\"You.com featuring YouChat\" data-normal=\"https:\/\/vanilla.futurecdn.net\/tomsguide\/media\/img\/missing-image.svg\" data-srcset=\"https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-320-80.png 320w, https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-480-80.png 480w, https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-650-80.png 650w, https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-970-80.png 970w, https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-1024-80.png 1024w, https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7-1200-80.png 1200w\" data-sizes=\"(min-width: 1000px) 970px, calc(100vw - 40px)\" data-original-mos=\"https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7.png\" data-pin-media=\"https:\/\/cdn.mos.cms.futurecdn.net\/kuCKUCbFqHvGshDjoz54d7.png\" data-ll-status=\"loaded\" \/><\/picture><\/p>\n<\/div>\n<\/div><figcaption class=\" inline-layout\"><span class=\"credit\">(Image credit: You.com)<\/span><\/figcaption><\/figure>\n<p><u>You.com<\/u>\u00a0is a search engine that has actually had an AI chatbot longer than Microsoft\u2019s Bing. Created by former Salesforce employees, the search engine introduced YouChat in December 2022 \u2014 upgrading to YouChat 2.0 in February 2023.<\/p>\n<p>The big selling point of YouChat and its LLM called Conversation, Apps and Links (C-A-L) is that it can integrate You.com apps for sites such as Reddit and YouTube into its chatbot\u2019s responses.<\/p>\n<figure>\n<blockquote><p>YouChat showed some promise when we tested it, but the app integration is not fully there<\/p><\/blockquote>\n<p>&nbsp;<\/figure>\n","protected":false},"excerpt":{"rendered":"<p>&nbsp; Top 9 Free GPT-3 Alternative AI models Now that you\u2019ve got an idea of what is the technology we are talking about, let\u2019s move on to the\u00a0OpenAI GPT-3\u00a0competitors. OPT by Meta Another solid\u00a0GPT-3 open-source\u00a0alternative was released by Meta in May 2022.\u00a0Open Pretrained Transformer language\u00a0model (OPT for short) contains 175B parameters. OPT was trained on [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"pagelayer_contact_templates":[],"_pagelayer_content":"","footnotes":""},"categories":[1],"tags":[],"class_list":["post-250","post","type-post","status-publish","format-standard","hentry","category-uncategorized"],"_links":{"self":[{"href":"https:\/\/techwyns.com\/tech\/wp-json\/wp\/v2\/posts\/250","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/techwyns.com\/tech\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/techwyns.com\/tech\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/techwyns.com\/tech\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/techwyns.com\/tech\/wp-json\/wp\/v2\/comments?post=250"}],"version-history":[{"count":4,"href":"https:\/\/techwyns.com\/tech\/wp-json\/wp\/v2\/posts\/250\/revisions"}],"predecessor-version":[{"id":254,"href":"https:\/\/techwyns.com\/tech\/wp-json\/wp\/v2\/posts\/250\/revisions\/254"}],"wp:attachment":[{"href":"https:\/\/techwyns.com\/tech\/wp-json\/wp\/v2\/media?parent=250"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/techwyns.com\/tech\/wp-json\/wp\/v2\/categories?post=250"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/techwyns.com\/tech\/wp-json\/wp\/v2\/tags?post=250"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}