In the artificial intelligence (AI) field, there has been a lot of talk about a significant statement made by OpenAI. The company recently released GPT-4, a sizable multimodal model that can handle both text and visual inputs. This new language model is an improvement on its predecessor, GPT-3, which was already revolutionary in and of itself.
GPT-4 and its features
- Large-scale multimodal model GPT-4 was developed by OpenAI.
- Text is just one component of multimodal models; GPT-4 also takes picture input. GPT-3 and GPT-3.5, on the other hand, only supported text as a mode of operation, which limited users to typing out queries.
- Moreover, GPT-4 “displays human-level performance on numerous academic and professional criteria.”
- The language model’s stronger general knowledge and problem-solving skills enable it to pass a mock bar exam with a score in the top 10% of test takers and to solve challenging questions more accurately.
- It may, for instance, “address tax-related queries, arrange a meeting for three busy individuals, or determine a user’s creative writing style.”
- A more comprehensive range of use cases, including lengthy discussions, document search and analysis, and long-form content production, is now possible because of GPT-4’s ability to handle texts longer than 25,000 words.
How is GPT-4 different from GPT-3?
Here are some of the major differences:
GPT-4 can ‘see’ images now
- The most obvious modification to GPT-4 is that it is multimodal, enabling it to comprehend input from several informational modalities.
- GPT-3 and ChatGPT’s GPT-3.5 could only read and write text, hence they were restricted to text input and output. GPT-4, however, may be instructed to produce data in response to pictures that are supplied to it.
- It makes sense if this makes you think of Google Lens. Lens, however, only looks for data that is relevant to a picture.
- GPT-4 is far more sophisticated in that it can comprehend and analyse images.
- An illustration of an outrageously huge iPhone connection with the language model explaining the humour was supplied by OpenAI. The main drawback is that picture inputs are currently at the research preview stage and are not accessible to the general public.
GPT-4 is harder to trick
- The tendency of generative models like ChatGPT and Bing to periodically go off course and provide suggestions that raise questions or, worse, outright scare users is one of their major shortcomings.
- They may also mess up the facts and spread false information.
- The company’s “best-ever results on factuality, steerability, and refusing to stray outside of guardrails” were achieved, according to OpenAI, after 6 months of training GPT-4 using lessons from its “adversarial testing programme” and ChatGPT.
GPT-4 can process a lot more information at a time
- Despite having been trained on trillions of parameters and infinite quantities of data, there are limitations to how much information Large Language Models (LLMs) can handle during a conversation.
- The GPT-3.5 model of ChatGPT was capable of handling 4,096 tokens, or around 8,000 words, while GPT-4 increases those capacities to 32,768 tokens or over 64,000 words.
- This improvement implies that, unlike ChatGPT, which could only process 8,000 words at a time before losing track of things, GPT-4 can continue to function properly for far longer talks.
- Moreover, it can handle longer documents and produce long-form material, which was much more restricted on GPT-3.5.
GPT-4 has an improved accuracy
- OpenAI acknowledges that GPT-4 still lacks complete reliability and commits reasoning gaffes, much as earlier iterations.
- Nonetheless, “GPT-4 dramatically lowers hallucinations compared to earlier models” and receives a factuality assessment score 40% higher than GPT-3.5.
- It will be far more difficult to persuade GPT-4 to generate undesired outputs like hate speech and false information.
GPT-4 is better at understanding languages that are not English
- Training LLMs in other languages might be difficult since machine learning data and most of the content on the internet nowadays are primarily in English.
- Yet, OpenAI has shown that it beats GPT-3.5 and other LLMs by correctly answering thousands of multiple-choice questions across 26 languages, whereas GPT-4 is more multilingual.
- With an accuracy rate of 85.5%, it clearly handles English the best, although Indian languages like Telugu aren’t far behind at 71.4%.
- This implies that consumers will be able to utilise chatbots built on GPT-4 to provide outputs in their local languages that are more accurate and clear.
Variety of risks that can arise out of GPT-4
- GPT-4 is still susceptible to manipulation by cyber hackers who want to create harmful programmes.
- It entails utilising the C++ programming language to create malware that can gather sensitive Portable Document Format (PDF) files and send them to distant servers through a covert file transfer mechanism.
- Additional risks that Check Point’s researchers may utilise include the “PHP Reverse Shell” technique, which hackers use to access a device and its data remotely, writing Java code to download malware remotely, and developing phishing draughts by pretending to be bank and employee emails.
- With advancements in technologies like GPT-4, people in outlying towns and cities may now launch more complex social engineering assaults, which can produce a significant amount of cyber threats.
- With one of the numerous generative AI tools, a significantly greater number of users who would not have been proficient at writing realistic phishing and spam letters can easily produce social engineering draughts, such as posing as an employee or a corporation, to target new customers.
Is GPT-4 available for the public right now?
- For various reasons, GPT-4 has already been included in services like Duolingo, Stripe, and Khan Academy.
- Even though it hasn’t yet been made freely accessible to everyone, a $20 per month ChatGPT Plus membership may get you to access right now. Although this is going on, GPT-3.5 continues to form the foundation of ChatGPT’s free tier.
- There is, however, an “unofficial” option to start utilising GPT-4 right away if you don’t want to pay.
- According to Microsoft, the new Bing search interface is now powered by GPT-4, and you can use it right now at bing.com/chat.