Open AI’s GPT-4 Vs. GPT-3 (Chat GPT)

The emergence of the latest version of GPT models, namely GPT-4, has garnered significant attention towards the already well-known OpenAI language models. And why wouldn’t it? OpenAI has introduced GPT-4 as their most sophisticated system yet, capable of solving complex problems with greater precision, thanks to its wider general knowledge and enhanced problem-solving capabilities.

This article aims to provide a comprehensive comparison between GPT-3, GPT-3.5, and GPT-4 models. We will delve into the similarities and differences between these OpenAI models, and highlight their best use cases.

What is Open AI’s GPT-3?

OpenAI’s GPT-3 is a cutting-edge language model that has taken the tech world by storm since its release in June 2020. This advanced model has gained immense popularity for its exceptional language generation capabilities and has already proven to be a game-changer in the field of natural language processing.

GPT-3 comprises multiple base models with varying parameters and computational resources. The most prominent models are Ada, Babbage, Curie, and Davinci. These models range from smaller-scale models, which require less computational resources, to massive models that require significantly more computational power.

davinciMost capable GPT-3 model. Can do any task the other models can do, often with higher quality.
curieVery capable, but faster and lower cost than Davinci.
babbageCapable of straightforward tasks, very fast, and lower cost.
adaCapable of very simple tasks, usually the fastest model in the GPT-3 series, and lowest cost.
Source: Open AI

OpenAI’s release of the new version of GPT-3, known as “text-davinci-003,” on March 15, 2022, was a significant breakthrough in the field of natural language processing. This advanced model was touted as being more capable than its predecessors, with a broader range of capabilities and an enhanced ability to generate human-like text.

One of the most notable improvements of the “text-davinci-003” model was its updated training data. Unlike the previous versions of the model, which were trained on data up to October 2019, the “text-davinci-003” model was trained on data up to June 2021. This up-to-date training data allowed the model to offer more accurate and relevant language processing capabilities, making it a significant improvement over its predecessors.

Interestingly, eight months after its release, OpenAI began referring to this model as part of the “GPT-3.5” series. While the exact reasoning behind this classification is unclear, it likely indicates that the “text-davinci-003” model represents a significant step forward in the evolution of the GPT-3 family of language models.

What is GPT-3.5?

GPT-3.5 is a series of advanced language models that represent a significant evolution of the GPT-3 family. As of today, there are five different variants of the GPT-3.5 models, four of which are optimized for text completion tasks, while the fifth is optimized for code completion tasks.

These models build upon the capabilities of GPT-3, offering even greater accuracy, efficiency, and versatility.

Most capable GPT-3.5 model and optimized for chat at 1/10th the cost of text-davinci-003. Will be updated with our latest model iteration.
4,096 tokensUp to Sep 2021
gpt-3.5-turbo-0301Snapshot of gpt-3.5-turbo from
March 1st 2023. Unlike gpt-3.5-turbo, this model will not receive updates, and will only be supported for a three month period ending on June ist 2023.
4,096 tokensUp to Sep 2021
gpt-3.5-turbo-003Can do any language task with better quality, longer output, and consistent instruction-following than the curie, babbage, or ada models. Also supports inserting completions within text.4,097 tokensUp to Jun 2021
gpt-3.5-turbo-002Similar capabilities to text-davinci-003 but trained with supervised fine-tuning instead of reinforcement learning4,097 tokensUp to Jun 2021
code-davinci-002Optimized for code-completion tasks8,001 tokensUp to Jun 2021
Source: Open AI

The most recent addition to the GPT-3.5 series is the gpt-3.5-turbo, which was released on March 1, 2023. This latest model has already generated a great deal of interest due to its enhanced capabilities and improved performance.

As we anticipate the upcoming release of GPT-4, the gpt-3.5-turbo serves as a powerful example of the continued advancements in natural language processing and the potential for these models to revolutionize a wide range of industries and applications.

What is Open AI’s GPT-4?

GPT-4 is the most recent – and the most advanced – version of the OpenAI language models. Introduced on March 14, 2023, it is said to be the new milestone in deep learning development.

GPT-4 is said to be able to generate more factually accurate statements than GPT-3 and GPT-3.5, ensuring greater reliability and trustworthiness. It’s also multimodal, meaning it can accept images as inputs and generate captions, classifications, and analyses.

Last but not least, it has gained some creativity. As we can read in the official product update, “it can generate, edit, and iterate with users on creative and technical writing tasks, such as composing songs, writing screenplays, or learning a user’s writing style.” For now, in March 2023, the GPT-4 comes in two model variants:

  • gpt-4-8K
  • gpt-4-32K

which differ by the size of their size of the context window.

Even though GPT-4 is already used commercially, most users will need to wait some time until they get access to GPT-4 API and build their own GPT-4-powered applications and services. Is it worth waiting? Let’s see!

GPT-4 Vs. GPT-3 & GPT-3.5 (Chat GPT)

According to what Greg Brockman told TechCrunch, the president and co-founder of OpenAI, GPT-4 and GPT-3 are different from each other. Although GPT-4 still has a fair share of problems and errors, it shows an improvement in certain domains such as calculus or law. The latest research published by OpenAI on GPT-4 has provided some insights into the advancements of the model.

He quotes:

There’s still a lot of problems and mistakes that [the model] makes … but you can really see the jump in skill in things like calculus or law, where it went from being really bad at certain domains to actually quite good relative to humans.
Greg Brockman

Allow me to delve deeper into the topic at hand based on the recently published research on GPT-4 by OpenAI sheds light on a plethora of intricate details regarding the new models, which are nothing short of remarkable.

GPT 4 VS GPT 3: Capabilities

GPT-4, the successor to GPT-3.5, boasts a host of advanced features that surpass its predecessor. Its capabilities include enhanced reliability, creativity, collaboration, and nuanced instruction handling. OpenAI developers conducted various benchmark tests to illustrate the differences between the two models, including simulating human-designed exams.

The tests conducted by OpenAI developers using GPT-4 were based on the latest publicly available exams, or practice exams that were bought for the 2022-2023 editions. No specific training was conducted for these exams, and only a small fraction of the exam problems were seen by the model during training. Despite these limitations, the results obtained were impressive.

Compared to GPT-3, GPT-4 outperformed it on various simulated exams. For instance, GPT-4 scored 4 out of 5 in the AP Calculus BC exam, while GPT-3 only scored 1. In a simulated bar exam, GPT-4’s performance was in the top 10% of test takers, while GPT-3.5 performed poorly, ranking at the bottom 10%.

Furthermore, GPT-4’s language proficiency has been significantly improved, making it a true polyglot. While its English proficiency was already high in GPT-3 and GPT-3.5 versions, with shot accuracy at 70.1%, GPT-4’s accuracy has increased to over 85%.

This means that GPT-4 can now speak 25 languages better than its predecessor spoke English, including Mandarin, Polish, and Swahili. This achievement is remarkable, especially considering that most machine learning benchmarks are written in English.

In addition, GPT-4 can handle longer text at a single request than GPT-3, thanks to its increased context length. This means that GPT-4 can process more complex and detailed instructions and provide more accurate results.

GPT 4 VS GPT 3: Token Limits

Context length refers to the maximum number of tokens that can be included in a single API request. The original GPT-3 models, which were released in 2020, had a maximum request value of 2,049 tokens. However, the GPT-3.5 version increased this limit to 4,096 tokens, which is approximately equivalent to three pages of single-spaced English text.

OpenAI model’s versionGPT-3
(ada, babbage, curie, davinci)
(gpt-3.5-turbo, gpt-3.5-turbo-0301,text-davinci-003, text-davinci-002)
Context length (max request)2,0494,0968,19232,768
Number of English words~1,500~3,000~6,000~24,000
Number of single-spaced pages of English text361250

The upcoming GPT-4 model will be available in two variants. The first one, GPT-4-8K, has a context length of 8,192 tokens. The second one, GPT-4-32K, can handle as much as 32,768 tokens, which is equivalent to around 50 pages of text.

This expanded context length opens up exciting possibilities for the use of GPT-4 in various applications. With its ability to process 50 pages of text, the new OpenAI models can generate longer pieces of text, analyze and summarize larger documents or reports, and engage in conversations while maintaining context.

In a recent interview with Techcrunch, Greg Brockman highlighted the potential of GPT-4 in various use cases.

Moreover, GPT-4 is not limited to processing text inputs alone, but it can also interpret other input types.

Previously, the model didn’t have any knowledge of who you are, what you’re interested in and so on. Having that kind of history [with the larger context window] is definitely going to make it more able … It’ll turbocharge what people can do.
Greg Brockman

GPT 4 VS GPT 3: Input data support

The GPT-4 is an innovative language model that is a step ahead of its predecessors, GPT-3 and GPT-3.5. While these models were limited to processing text or code, the GPT-4 introduces a new type of input: images. With this feature, the GPT-4 can generate text outputs based on input that includes both text and images.

The capabilities of the GPT-4 are impressive, as it can perform various tasks depending on the input. It can generate captions, classify visible elements, or even analyze the image itself. In the GPT-4 research documentation, there are examples of the model analyzing graphs, explaining memes, and summarizing papers that contain both text and images.

Undeniably, the image understanding abilities of the GPT-4 are remarkable. Its newfound ability to process images, combined with the increased token limits, creates new possibilities for using the model in various fields such as academic research, personal training, or shopping assistants.

However, it is important to note that the GPT-4’s image input feature is still a research preview and not publicly available. Therefore, it may take some time before individuals can fully utilize this innovative capability.

GPT 4 VS GPT 3: Defining the Output context

One of the major differences between GPT-3 and GPT-4 is the ability to control the model’s tone, style, and behavior. With GPT-4, it is now possible to provide instructions to the model on an API level using “system” messages (described in detail in the OpenAI’s Usage policy). These messages determine the tone of the model’s responses and provide guidelines for how it should behave. For example, a system message could instruct the model to always ask questions to help a student learn, rather than giving them the answer directly.

In addition, these system messages establish boundaries for GPT-4’s interactions and act as “guardrails” to prevent the model from changing its behavior at the user’s request. This ensures that the model stays within its defined role and behaves consistently, regardless of the user’s input.

A similar ability was seen in the recently released GPT-3.5-Turbo, which allowed users to define the model’s role in a system prompt and receive a different response depending on the specified role. However, prior to the release of GPT-3.5-Turbo in March 2023, it was not possible to provide the model with system messages, and all context information had to be given within the prompt.

With the new GPT-4, users have greater control over the model’s behavior and can adjust it to conform with their brand communication guidelines or other external specifications. This feature opens up new possibilities for how the model can be used in various industries and contexts.

GPT 4 VS GPT 3: Cost of Operation

Undoubtedly, advanced technology comes with a price tag. GPT-3 models have a cost ranging from $0.0004 to $0.02 for every 1,000 tokens. Recently, the GPT-3.5-Turbo model has come 10 times cheaper than the most potent GPT davinci model at $0.002 per 1,000 tokens. However, if you aim to use the most sophisticated GPT-4 models, you should prepare to pay more.

The GPT-4 model with an 8K context window would cost $0.03 per 1,000 prompt tokens and $0.06 per 1,000 completion tokens. In contrast, the GPT-4 model with a 32K context window would cost $0.06 per 1,000 prompt tokens and $0.12 per 1,000 completion tokens.

For instance, processing 100,000 requests with an average length of 1,500 prompt tokens and 500 completion tokens would cost $4,000 with text-davinci-003 and $400 with gpt-3.5-turbo. However, with the GPT-4 model, the same task would cost $7,500 with the 8K context window and $15,000 with the 32K context window.

Not only is it expensive, but it is also complex to calculate. This is due to the cost of the prompt tokens being different from the cost of completion tokens. Estimating token usage is challenging because there is minimal correlation between input and output length. With the higher cost of output tokens, the cost of using GPT-4 models becomes even more unpredictable.

GPT 4 VS GPT 3: Errors and limitations

OpenAI’s CEO responded to various rumors surrounding the development of GPT-4, including one about the number of parameters it will use. He said,

The GPT-4 rumor mill is a ridiculous thing. I don’t know where it all comes from. People are begging to be disappointed, and they will be. (…) We don’t have an actual AGI, and that’s sort of what’s expected of us.
Greg Brockman

While it’s undeniable that GPT-4 showcases impressive creativity and capabilities, it’s crucial to acknowledge its limitations. As stated in the product research documentation, the changes made to the model are not substantial compared to its predecessors.

Similar to previous versions, GPT-4 lacks knowledge of events that have taken place after September 2021. Despite the advanced intelligence of ChatGPT, it is still not completely reliable, even with the incorporation of GPT-4. Although it has been reported to significantly reduce hallucinations when compared to previous models (achieving a 40% higher score than GPT-3.5 in internal evaluations), it still has the tendency to produce inaccurate information and make reasoning errors.

It is also capable of generating harmful advice, although it is less likely to do so, as well as producing buggy code. As a result, it is not recommended for use in applications where errors could result in high costs.

GPT-4 vs. GPT-3 – Bottom Line

GPT-4, OpenAI’s most sophisticated system to date, surpasses its predecessors in nearly every aspect of comparison. It boasts greater creativity and coherence than GPT-3, can handle longer text passages and even images and is more precise and less prone to fabricating “facts.” Thanks to its advanced capabilities, GPT-4 opens up numerous possibilities for generative AI.

However, this does not necessarily mean that GPT-4 will supplant GPT-3 and GPT-3.5 altogether. While the former is undoubtedly more powerful than its predecessors, it also comes with a significantly higher cost. In many scenarios where a model doesn’t need to process multi-page documents or recall lengthy conversations, the capabilities of GPT-3 and GPT-3.5 will suffice.

Also Read:

Rate this post
Posted by
Bhanu Pratap

Hi! I am Bhanu Pratap, co-founder of Yorker Media Group. A die-heart fan of tech and keeps track of every little happening of the same. When not writing, I usually keep myself busy on YouTube making and exploring new and awesome tech content.

Leave a Reply

Your email address will not be published. Required fields are marked *