The AI ​​race: How Google and OpenAI are following each other

Again, ChatGPT developer OpenAI has chosen a special time to release its innovation. On Monday — a day before Google’s I/O developer conference — the new AI model GPT-4o was presented to the public for the first time. The next day, Google also had a good piece of AI with them.

OÖN provides an overview of what the latest developments look like and why both companies use each other as models.

GPT-4o: You give “Omnimodel”.

According to OpenAI, the more than 100 million users working with ChatGPT have some innovations. Despite already having voice input, the software had to take a short break to process and answer questions. With the new “flagship model” GPT-4o, a “more natural and more simple” kind of interaction between humans and machines can now take place. From an AI perspective, this interaction should take place at the GPT-4 level, but much faster.

In recent years, OpenAI has focused on improving AI. Now, for the first time, a big step has been taken in terms of user-friendliness, said technology boss Meera Murati in an online presentation.

“The GPT-4o integrates logical thinking into speech, text and image recognition,” Murathi said. So OpenAI talks about an “omni model,” which explains the “o” in the name. Users have the option to upload photos and documents. In addition, the software can analyze the live image from a smartphone camera. GPT-4o takes and evaluates information from all these information sources.

Software can respond to different emotions when recording and outputting information. In a demonstration, ChatGPT created a bedtime story and read it aloud. You can interrupt the software and add more drama to your voice or ask it to speak like a robot. ChatGPT also sings the last sentence if requested. GPT-4o’s functions are available in 50 different languages ​​- even for free users. Paying customers are allowed to avail a wide range of offers.

See also  SPÖ leader Poplar categorically rejects an alliance with the FPÖ

OpenAI boss Sam Altman wrote after the presentation that it was the best way to use computing he had ever experienced. “It feels like AI you see in the movies. I’m still a little surprised that it’s real.” The presentation was preceded by rumors that OpenAI might compete with Google with an AI-powered search engine. There was no mention of that on Monday, but Muratti concluded the presentation with a reference to the “next big thing” that OpenAI wants to deliver “soon.”

More AI while Googling

After the GPT-4o announcements, the bar was set high for Google. After all, this software has the potential to become a better version of voice assistants like Siri, Alexa or Google Assistant. Google didn’t need to be asked twice when it came to AI at its I/O developer conference on Tuesday.

Google’s approach is called Gemini – an AI model that can generate queries not only in text, but also in image format. This is already possible on Google and Samsung smartphones. This functionality is now being integrated into other Google services under the motto “We Google for you”. The newest and fastest member of the Gemini family to date is called the Gemini 1.5 Flash and is intended to be fast and efficient.

A new feature of Google Search is called “Overviews”. This means that in the future, Google will remind you to interact with a chatbot. In a separate overview area, Google tries to answer the search query directly. Links to other websites known from previous Google searches are followed only below. A pilot phase shows that this type of search increases usage – and thereby user satisfaction. The new AI-powered search will initially launch in the US in English. But it should come to Europe “in the near future.”

See also  End winter training camp without Thorsten Schick

Like ChatGPT, Gemini can analyze uploaded files in the future. Google also has an update for paying customers titled Gemini Live. Users of the paid Gemini Advanced offer will be able to interact with the AI ​​assistant on their mobile devices. As with the GPT-4o, users can choose between several “natural-sounding” voices and interrupt Gemini’s responses to ask questions. The feature is expected to be available “in the coming months.”

Author of the book

Thomas Nigel

Online and technical teacher

Thomas Nigel

Thomas Nigel

loading

Information You can add a keyword to your titles by clicking the icon.

Information
Click the icon to open your “My Topics” page. They have 15 keywords saved and keywords to be removed.

Information You can remove a keyword from your titles by clicking the icon.

Add a title to your titles.

Leave a Reply

Your email address will not be published. Required fields are marked *