Cointime

Download App
iOS & Android

This is All That is Known About GPT-4 and What You Should Fear

Validated Individual Expert

A new beginning.

That’s what most of us felt when we tuned in yesterday to OpenAI’s official presentation of GPT-4, the newest generation of their world-famous pre-trained transformer.

And although there are plenty of impressive features to cover, there are also many open questions that were left unanswered, questions that need answering.

Can we answer for ourselves what is OpenAI trying to hide and what the future holds for all of us?

Your best work partner

During the 30-minute presentation, while more than one-hundred-thousand people looked in awe, OpenAI chose very carefully what they showed and how they showed it.

Understandable, but concerning in some aspects.

Truth be told, even though it was a very rehearsed presentation, they still managed to add some extra improvisation by allowing some viewers to test the model, in order to portray an image of going off the script.

But one way or another, they showed some mind-bending features.

Summarizing using the letter Q only

You read that title correctly, not only is GPT-4 capable of summarizing a long text into one sentence in a few seconds, it is capable of doing so using only words that begin with a ‘g’, or with ‘q’ (this last one at the request of a viewer).

Source: OpenAI

Albeit the stupidity of the request, it showcases the amazing linguistical capabilities of this model, far exceeding those of the GPT-3.5 turbo, the one running in ChatGPT’s API.

Also, it proved capable of cross-summarizing and finding a common theme among two different texts, thanks to the fact that GPT can now handle up to 32,000 tokens.

From OpenAI’s own documentation we know that 1,000 tokens equate to approximately 750 words. That means that GPT-4 is capable of receiving and understanding texts of up to 24,000 words, around 14 times the length of this article in one go, which is far more than the previous limit.

This is not trivial, considering that the longer the input is, the better context we can provide to it to generate better responses.

But this was just the beginning of what we were going to see.

A coding beast

Greg Brockman, OpenAI’s chairman, then moved into giving examples of how can GPT-4 improve the coding experience for developers.

And guys, I was speechless.

Firstly, it created a Discord bot from scratch by illustrating one of GPT-4’s most impressive features, instruction following.

By telling the model it now was an “AI programming assistant” and giving it some instructions, the model created the bot in a blink of an eye.

But that wasn’t the most impressive thing by a large margin.

As Greg acknowledged, GPT-4’s training cutoff is 2021, which means that GPT-4 has no knowledge from that point onwards. And the problem was that Discord had updated its API recently in 2022, something GPT-4 of course wasn’t aware of.

With the actual ChatGPT that would be the end of the story, but not with GPT-4.

Demonstrating great few-shot capabilities, the model was capable of using the new information that Greg gave it to update its knowledge and create the bot while meeting Discord’s 2022 requirements.

GPT-4 receiving instructions to create the Discord bot. Source: OpenAI

Again, language and semantic capabilities from another world that, in this case, match a human programmer’s performance (even exceeding if we consider the velocity at which GPT-4 created the bot).

However, the show was far from over, as the model also was capable of debugging its code when Greg sent it an error message.

At this point one is tempted to think, will human programmers still exist in a few years, or will simply serve as code reviewers?

Yet, as with everything, OpenAI had scheduled the best for last… GPT-4's new vision capabilities.

The Multimodal Era

Although Microsoft Germany’s CTO already leaked this, GPT-4 is going to be multimodal, which in simple terms means that the model is capable of receiving and understanding input from text, images, or even video.

For such purposes, they premiered GPT-4's perception capabilities, showing various images (some of them from the viewers) and then having the model describe them in accurate detail.

But probably the most impressive of all examples was when Greg drew a mockup of a website on pen and paper, sent it as a photograph to the model, and asked it to become a programmer again and create an HTML/CSS/JS website based on the mockup.

Et voilà. An interactive website in 10 seconds.

The actual mockup and the generated interactive website. Source: OpenAI

Truly incredible.

Even so, OpenAI was capable of squeezing one more feature into the presentation, one I personally wasn’t expecting.

Fear taxes no more

Taxes. Yes, taxes.

GPT-4 helped with a tax deduction after it was told that its new role was ‘TaxGPT’, and was also given a very long explanation of a tax deduction.

Again, not only GPT-4 managed to understand the whole text, remarkable considering how complex these documents are, but while doing impressive calculations and providing the correct answer.

Considering how poor ChatGPT’s performance has been regarding mathematics this was genuinely spectacular, mainly because it was clarified that no calculator APIs were connected to the model.

However, we must be cautious and see if these capabilities can be easily brought to shame with more complex mathematics.

Arriving to this point the presentation was closed, causing several questions that people were expecting to go unanswered.

And the issue is that some are very concerning.

We. need. answers.

As I said earlier, many questions remain regarding Generative AI in general.

Size, size, and more size

My main getaway from the presentation was that GPT-4 was considerably superior in its language skills to past versions.

Consequently, I’m led to believe that the biggest improvement we’ve seen is due to the fact that the model is simply much bigger.

That is, they are leveraging the fact that Large Language Models benefit from “scaling laws”, which is a fancy way of describing that these models get linguistically better as they get bigger.

From the multimodality standpoint, although we don’t have access to the model, by looking at the Kosmos-1 paper we can assume that they made important changes to the encoder part of the model, especially with regard to the visual transformers used to interpret images better.

But what about one of the most expected new features, video?

Video? Where?

With last week's news, we knew that GPT-4 would handle video. However, after yesterday’s presentation, we still don’t know in what sense.

As an input that it can then describe? Or will it be capable of actually generating videos?

Microsoft Germany’s CTO was unclear on this and so was OpenAI in the official launch. I guess we will have to wait on this matter.

But we still haven’t tackled the most concerning issue:

There’s an overarching issue with Generative AI in general, and that’s none other than its tendency to make stuff up.

Unreliability and business can’t be friends

Working in business environments, besides some specific cases like Programming or Marketing teams, no one will dare leverage Generative AI as a client-faced solution simply because it’s not reliable.

Sincerely, I was expecting that OpenAI would tackle this issue head-on in yesterday’s presentation, and the fact that they didn’t screams that the problem persists.

Yes, models are getting more reliable as they get bigger, and yes improvements on the Reinforcement Layer are allowing them to make better decisions, but the risk is undoubtedly still there.

From a business perspective, I still consider these solutions as great enhancers, solutions to make non-intelligent technologies like process automation or rule-based chatbots more intelligent, but never as standalone solutions.

To me, it still doesn’t seem like an option except for use cases with a high tolerance for error.

I mean, would you risk using a Chatbot that lies to your client? Of course not.

But what happens in those places where GPT-4 does fit?

Oh my.

Serious threat

In spite of Greg’s unwavering efforts to portray GPT-4 as a partner and not as a substitutor, which is something I agree with, the productivity enhancements will be immense.

Therefore, unless you’re different or unique, artists and writers are going to have a tough time with GPT-4. For instance, content writers seem completely substitutable, besides editors and others that will continue to play a role.

In a way, GPT-4 is going to elevate mediocre people into putting a fight against well-established artists and writers. Thus, I feel that only the greats in their fields will continue to thrive.

Having a unique perspective, a unique style, or a unique message will be key differentiators in this AI world that’s coming.

So, are you really special?

A final word

If you’ve read this article, you’re now ahead of 95% of society when it comes to AI.

But that still leaves a lot of people at that level.

So what if you’re capable of being above 99% of society?

That’s a totally different level, and if that’s where you want to be, I have news for you.

Read more: https://medium.com/@ignacio.de.gregorio.noblejas/gpt-4-released-81f8fc697def

AI
Comments

All Comments

Recommended for you

  • U.S. consumer confidence improves again in November, reaching a two-year high

    Dana M. Peterson, Chief Economist of the World Large Enterprises Federation, said, "US consumer confidence continued to improve in November, reaching the highest level in the past two years. The growth in November was mainly due to consumers' more positive assessment of the current situation, especially in the labor market. Compared with October, consumers' optimism about future employment opportunities has also greatly increased, reaching the highest level in nearly three years. At the same time, consumers' expectations for future business conditions have not changed, while their optimism about future income has slightly declined." Earlier, the US Conference Board Consumer Confidence Index for November recorded 111.7, a new high since July 2023.

  • Starknet: Phase 1 of STRK staking is now live on the mainnet

    Starknet announced that the first stage of STRK staking has officially launched on the mainnet.

  • CZ: Not trying to end the meme craze, just encouraging more builders

    CZ posted on X platform today, saying: "I am not against Meme coins, but Meme coins have become 'a little' strange now. Let's use blockchain technology to build practical applications." Some community users said that even Musk is a supporter of Meme coins, and it is very difficult to end this frenzy. CZ responded that "there is no attempt to end anything, everyone has the right to choose to invest or hold what they want. Just encourage more builders."

  • Talus Network Completes $6 Million Strategic Round of Financing with a Valuation of $150 Million

    decentralized AI protocol Talus Network raised $6 million in a strategic financing round led by Polychain Capital, valuing the company at $150 million. This funding will help further develop the Talus ecosystem, including the Protochain, Nexus framework, and "AI dating experience" application.

  • AXIOS: Trump is considering appointing a secretary of state for artificial intelligence

    according to AXIOS, Trump is considering appointing an AI minister to coordinate federal policies and government use of emerging technologies.

  • Coinbase International has launched COW perpetual contracts

     Coinbase International has launched COW perpetual contracts. COW-PERP market limit, market, stop loss, and stop loss limit orders are now all available.

  • Schuman Financial Completes $7.36 Million Seed Round, Led by RockawayX

    Schuman Financial has completed a $7.36 million seed round of financing, led by RockawayX, with participation from Lightspeed Faction, Kraken Ventures, Nexo Ventures, Gnosis VC, Delta Blockchain Fund and Bankless Ventures. In addition, Schuman Financial has launched a euro stablecoin, EURØP, which complies with the MiCA standard.

  • QCP: BTC's path to $100,000 has stalled, and ETH implied volatility has turned to put options

    QCP Capital has published an analysis indicating that the recent drop in the price of Bitcoin has resulted in long liquidations exceeding $430 million. This drop coincides with the end of five consecutive days of net inflows for spot ETFs, which recorded a outflow of $438 million on Monday, while MicroStrategy fell by 4.4%. With the US holiday approaching and no immediate catalyst to push prices higher, BTC's path towards $100,000 has stalled. In addition, the implied volatility of ETH has turned to bearish options rather than bullish options, and market concerns about downside risks may intensify, especially with the release of the FOMC meeting minutes and PCE data. However, in the long run, this market decline is not an excessive correction. Bitcoin has only retreated to last week's level. Since Trump's election, the market has become extremely overbought and leveraged, so a pause is inevitable.

  • Binance will delist GFT, IRIS, KEY, OAX, and REN

     Binance will delist the following trading pairs on December 10, 2024: GFT/USDT, IRIS/BTC, IRIS/USDT, KEY/USDT, OAX/BTC, OAX/USDT, REN/BTC, and REN/USDT. Additionally, Binance Futures will close all positions and automatically settle the KEYUSDT and RENUSDT USDⓈ-M perpetual contracts on December 3, 2024 at 09:00 (UTC). After the settlement is completed, the contracts will be delisted.

  • OpenTrade announces $4 million seed extension round led by AlbionVC

    OpenTrade has announced the completion of a $4 million seed extension financing round to build RWA-supported loan and stablecoin yield products. This round of financing was led by AlbionVC, with participation from a16z Crypto and CMCC Global. OpenTrade plans to use the funds to expand its operations and enhance its product capabilities.