Skip to main content

The best ChatGPT alternatives (according to ChatGPT)

ChatGPT has quickly become the darling of generative AI, but it’s hardly the only player in the game. In addition to all the other AI tools out there that do things like image generation, there’s also a number of direct competitors with ChatGPT — or so I assumed.

Why not ask ChatGPT about it? That’s exactly what I did to get this list, hoping to find some options for those facing “at capacity” notices, or others who just want try something new. Not all of these are as accessible to the public as ChatGPT, but according to ChatGPT, these are the best alternatives.

Bing by Microsoft

Microsoft's redesigned Bing search engine.
Image used with permission by copyright holder

Before getting into the picks listed by the AI, the best alternative to ChatGPT is, well, ChatGPT. Microsoft recently added the AI into its Bing search engine, and it plans on rolling out the feature to the Edge browser soon.

It’s only in preview, but you can still try out the new AI chatbot at right now. Microsoft says it’s limiting the number of queries initially, but you can join the Bing ChatGPT waitlist to be notified when the full version is available.

BERT by Google

BERT (Bidirectional Encoder Representations from Transformers) is a machine-learning model developed by Google. Lots of ChatGPT’s results mentioned projects by Google, which you’ll see later on in this list.

BERT is known for its natural language-processing (NLP) abilities, such as question-answering and sentiment analysis. It uses BookCorpus and English Wikipedia as its models for pretraining references, having learned 800 million and 2.5 billion words respectively.

BERT was first announced as an open-source research project and academic paper in October 2018. The technology has since been implemented into Google Search. Early literature about BERT compareb it to OpenAI’s ChatGPT in November 2018, noting that Google’s technology is deep bidirectional, which helps with predicting incoming text. Meanwhile, OpenAI GPT is unidirectional and can only respond to complex queries.

Meena by Google

Meena is a chatbot that Google introduced in January 2020 with the ability to converse in a humanlike fashion. Examples of its functions include simple conversations that include interesting jokes and puns, such as Meena suggesting cows study “Bovine sciences” at Harvard.

Google Meena chatbot expample.

As a direct alternative to OpenAI’s GPT-2, Meena had the ability to process 8.5 times as much data as its competitor at the time. Its neural network comprises 2.6 parameters and it is trained on public domain social media conversations. Meena also received a metric score in Sensibleness and Specificity Average (SSA) of 79%, making it one of the most intelligent chatbots of its time.

The Meena code is available on GitHub.

RoBERTa by Facebook

RoBERTa (Robustly Optimized BERT Pretraining Approach) is another advanced version of the original BERT, which Facebook announced in July 2019.

Facebook created this NLP model with a larger source of data as its pretraining model. RoBERTa uses CommonCrawl (CC-News), which includes 63 million English news articles generated between September 2016 and February 2019, as its 76GB data set. In comparison, the original BERT uses 16GB of data between its English Wikipedia and BookCorpus data sets, according to Facebook.

Silimar to XLNet, RoBERTa beat BERT in a set of benchmark data sets, as per Facebook’s research. To get these results, the company not only used a larger data source but also pretrained its model for a longer period of time.

Facebook made RoBERTa open-source in September 2019, and its code is available on GitHub for community experimentation.

VentureBeat also mentioned GPT-2 among the emerging AI systems during that time.

XLNet by Google

XLNET is a transformer-based autoregressive language model developed by a team of Google Brain and Carnegie Mellon University researchers. The model is essentially a more advanced BERT and was first showcased in June 2019. The group found XLNet to be at least 16% more efficient than the original BERT, which was announced in 2018, with it able to beat BERT in a test of 20 NLP tasks.

XLNet: a new pretraining method for NLP that significantly improves upon BERT on 20 tasks (e.g., SQuAD, GLUE, RACE)


github (code + pretrained models):

with Zhilin Yang, @ZihangDai, Yiming Yang, Jaime Carbonell, @rsalakhu

— Quoc Le (@quocleix) June 20, 2019

With both XLNet and BERT using “masked” tokens to predict hidden text, XLNet improves efficiency by speeding up the predictive part of the process. For example, Amazon Alexa data scientist Aishwarya Srinivasan explained that XLNet is able to identify the word “New” as being associated with the term “is a city” before predicting the term “York” as also being associated with that term. Meanwhile, BERT needs to identify the words “New” and “York” separately and then associate them with the term “is a city,” for example.

Notably, GPT and GPT-2 are also mentioned in this explainer from 2019 as other examples of autoregressive language models.

XLNet code and pretrained models are available on GitHub. The model is well-known among the NLP research community.

DialoGPT by Microsoft Research

The DialoGPT (Dialogue Generative Pre-trained Transformer) is an autoregressive language model that was introduced in November 2019 by Microsoft Research. With similarities to GPT-2, the model was pretrained to generate humanlike conversation. However, its primary source of information was 147 million multi-turn dialogues scraped from Reddit threads.

DiabloGPT multi-turn generation examples.
Image used with permission by copyright holder

HumanFirst chief evangelist Cobus Greyling has noted his success at implementing DialoGPT into the Telegram messaging service to bring the model to life as a chatbot. He added that using Amazon Web Services and Amazon SageMaker can help with fine-tuning the code.

The DialoGPT code is available on GitHub.

ALBERT by Google

ALBERT (A Lite BERT) is a truncated version of the original BERT and was developed by Google in December 2019.

With ALBERT, Google limited the number of parameters allowed in the model by introducing parameters with “hidden layer embeddings.”

Machine performance on the RACE challenge (SAT-like reading comprehension) by Google
Image used with permission by copyright holder

This improved not only on the BERT model but also on XLNet and RoBERTa because ALBERT can be trained on the same larger data set of information used for the two newer models while adhering to smaller parameters. Essentially, ALBERT only works with the parameters necessary for its functions, which increased performance and accuracy. Google detailed that it found ALBERT to exceed BERT on 12 NLP benchmarks, including an SAT-like reading comprehension benchmark.

While not mentioned by name, GPT is included within the imaging for the ALBERT on Google’s Research blog.

Google released the ALBERT as open-source in January 2020, and it was implemented on top of Google’s TensorFlow. The code is available on GitHub.

T5 by Google

T5 (Text-to-Text Transfer Transformer) is a NLP model introduced by Google in 2019 that borrows from a host of prior models, including GPT, BERT, XLNet, RoBERTa, and ALBERT, among others. It adds a new and unique data set called Colossal Clean Crawled Corpus (C4), which allows the transformer to produce higher-quality and contextual results than other data sets in comparison to the Common Crawl web scrapes used for XLNet.
Google T5 Text-To-Text Transfer Transformer pre-training.
The T5 pretraining led to the creation of chatbot applications, including InferKit Talk To Transformer and the AI Dungeon game. The text generators resemble ChatGPT in that they allow you to generate realistic conversations based on what the AI generates after your initial prompts or queries.
The T5 code is available on GitHub.

CTRL by Salesforce

CTRL by Salesforce (Computational Trust and Reasoning Layer) was one of the largest publicly released language models when it was announced in September 2019 by Salesforce. The 1.6 billion-parameter language model can be used to analyze large bodies of text at once, such as those associated with webpages. Some potential practical uses include pairing with reviews, ratings, and attributions.
Salesforce CTRL source attribution example.
The CTRL language model can differentiate down to the punctuation the intent of a specific query.  Salesforce noted the model can pick up the difference between “Global warming is a lie.” as an unpopular opinion and “Global warming is a lie” as a conspiracy theory due to the difference of the period in the phrases and draft up corresponding Reddit threads for each.
CTRL references up to 140GB of data for its pretraining from sources, including Wikipedia, Project Gutenberg, Amazon reviews, and Reddit. It also references a number of international news, information, and trivia resources.
The CTRL code is available on GitHub.

GShard by Google

GShard is a giant language translation model that Google introduced in June 2020 for the purpose of neural network scaling. The model includes 600 billion parameters, which allows for large sets of data training at once. GShard is particularly adept at language translation and being trained to translate 100 languages into English in four days.

Blender by Facebook AI Research

Blender is an open-source chatbot that was introduced in April 2020 by Facebook AI Research. The chatbot has been noted to have improved conversational skills over competitor models, with the ability to provide engaging talking points, listen and show understanding of its’s partner’s input, and showcase empathy and personality.

Blender chatbot example.

Blender has been compared to Google’s Meena chatbot, which has in turn been compared to OpenAI’s GPT-2

The Blender code is available on

Pegasus by Google

Pegasus is a natural language processing model that was introduced by Google in December 2019. Pegasus can be trained to create summaries, and similar to other models like BERT, GPT-2, RoBERTa, XLNet, ALBERT, and T5, it can be fine-tuned to specific tasks. Pegasus has been tested on its efficiency in summarizing news, science, stories, instructions, emails, patents, and legislative bills in comparison to human subjects.

The PEGASUS NLP has been compared to a human in terms of summarizing quality.
Image used with permission by copyright holder

The Pegasus code is available on GitHub.

Editors' Recommendations

Fionna Agomuoh
Fionna Agomuoh is a technology journalist with over a decade of experience writing about various consumer electronics topics…
Most people distrust AI and want regulation, says new survey
A person's hand holding a smartphone. The smartphone is showing the website for the ChatGPT generative AI.

Most American adults do not trust artificial intelligence (AI) tools like ChatGPT and worry about their potential misuse, a new survey has found. It suggests that the frequent scandals surrounding AI-created malware and disinformation are taking their toll and that the public might be increasingly receptive to ideas of AI regulation.

The survey from the MITRE Corporation and the Harris Poll claims that just 39% of 2,063 U.S. adults polled believe that today’s AI tech is “safe and secure,” a drop of 9% from when the two firms conducted their last survey in November 2022.

Read more
Microsoft Copilot vs. Google Duet: battle of the next-gen AI smart assistants
Microsoft's AI Copilot being used in various Microsoft Office apps.

Microsoft Copilot and Google Duet are the two most prominent artificial intelligence assistants put out by the various tech giants since OpenAI debuted its ChatGPT chatbot in 2022. They're set to bring that powerful natural language assistance into the enterprise in ways that can enhance productivity, improve the digital fluency of workers, and leverage existing data in new and exciting ways.

But which one is best? Both offer comparable features at a comparable cost, but they aren't interchangeable and even getting access to either tool requires a bit of luck. Here's how these two awesome AI tools compare.

Read more
Meta is reportedly working on a GPT-4 rival, and it could have dire consequences
The Facebook app icon on an iPhone home screen, with other app icons surrounding it.

Facebook owner Meta is working on an artificial intelligence (AI) system that it hopes will be more powerful than GPT-4, the large language model developed by OpenAI that powers ChatGPT Plus. If successful, that could add much more competition to the world of generative AI chatbots -- and potentially bring a host of serious problems along with it.

According to The Wall Street Journal, Meta is aiming to launch its new AI model in 2024. The company reportedly wants the new model to be “several times more powerful” than Llama 2, the AI tool it launched as recently as July 2023.

Read more