Latest Posts

GPT-4o ‘omni’ will be able to provide simultaneous translation in 50 languages

GPT-4o ‘omni’ will be able to provide simultaneous translation in 50 languages

OpenAI is launching GPT-4o ‘omni’, a new generative AI language model that powers ChatGPT. GPT-4o allows users to interact with ChatGPT more like an assistant. GPT-4o, which can provide simultaneous translation in 50 languages, see and perceive the environment with the phone camera, and accompany users in the solution process of a mathematical problem like a private teacher, will be available in the next few weeks.

At its spring update event today, OpenAI introduced GPT-4o, a new generative AI language model that powers ChatGPT.

“o” stands for “omni,” referring to the model’s ability to process text, speech, and video. GPT-4o will be available in the company’s developer and consumer products within the next few weeks.

GPT-4o, which can provide simultaneous translation in 50 languages, see and perceive the environment with the phone camera, and accompany users in the solution process of a mathematical problem like a private teacher, will be available in the next few weeks.

OpenAI Chief Technology Officer Mira Murati said ChatGPT will now also have memory capabilities, meaning it can learn from previous conversations with users and do real-time translation. Mira Murati said GPT-4o provides “GPT-4-level” intelligence but enhances GPT-4’s capabilities in a variety of methods and environments.

OpenAI’s previous “pioneering ‘state-of-the-art’ model, GPT-4 Turbo, was trained on a combination of images and text and could analyze images and text to perform tasks like extracting text from images and even identifying the content of those images. But GPT-4o adds speech to the mix .

What are the features of GPT-4o?

GPT-4o will enable ChatGPT to interact using text, voice and pseudo-image. This means it can view screenshots, photos, documents or graphics uploaded by users and say something about them. GPT-4o works faster in text, audio and video fields.

The model differentiates itself from ChatGPT’s voice mode with its capabilities when talking to users in real time. Within the scope of these capabilities, the model can distinguish the difference between voice intonations. Likewise, it can produce different kinds of sounds. It should also be noted that the model can change its tone by adding expressions to its voice.

At the event, it was also demonstrated that ChatGPT can see and detect the environment with the phone camera. The system, which can detect a given software code, provide information about the code, and chat by answering questions about the code, creates the feeling of actually talking to a human as it can be interrupted in voice communication.

GPT-4o can solve math problems by talking like a private teacher

GPT-4o also supports users in solving mathematical problems by speaking fluently. The model, which can view videos in real time, simultaneously accompanied the solution process of a mathematical problem, step by step, like a private teacher.

GPT-4o allows users to interact with ChatGPT more like an assistant.

GPT-4o will be able to provide simultaneous translation in 50 languages

GPT-4o also stood out with its translation capabilities. The model provided simultaneous translation between Italian and English at the event. By the way, the model supports 50 languages.
The model, which is stated to be smarter and 2 times faster, will surprisingly be available free of charge to users, with limits only in the free part.

Latest Posts

Don't Miss