From the course: AI Trends
Unlock this course with a free trial
Join today to access over 23,100 courses taught by industry experts.
GPT-4o, multimodal AI, and more
From the course: AI Trends
GPT-4o, multimodal AI, and more
- OpenAI had their spring update on May 13th, 2024 where they released their latest model named GPT-4o for omni and a bunch of other stuff. Here's the breakdown and what you need to know. The main update is the release of GPT-4o, the first model that integrates text, image and audio and can be used to combine all these three modalities in both the input and the output phase. In effect, this means what used to require multiple operations, turn speech into transcription then run the transcription through GPT then turn the response back into speech is now simplified and takes significantly less time. The new GPT-4o model is described as being two times faster than GPT-4 Turbo and the GPT-4o API is two times faster, 50% cheaper and has five times higher rate limits than GPT-4 Turbo. Long story short, GPT-4o replaces GPT-4 Turbo as the new benchmark model from OpenAI today. With this in mind, here are the four things you need to…
Contents
-
-
-
(Locked)
Claude Artifacts4m 2s
-
(Locked)
Microsoft Build 2024: New computers and developer tools6m 45s
-
(Locked)
NPUs vs. GPUs vs. CPUs2m 45s
-
(Locked)
New Google Gemini Models and Google I/O Announcements4m 44s
-
(Locked)
GPT-4o, multimodal AI, and more5m 4s
-
(Locked)
OpenAI Sora: Text-to-video1m 34s
-
(Locked)
AI regulations6m 48s
-
(Locked)
General artificial intelligence3m 43s
-
(Locked)
The LLM landscape2m 43s
-
(Locked)
-
-