– OpenAI released an update to GPT-4-Turbo with improved responses and analysis capabilities
– The model includes AI vision technology for analyzing video, image, and audio content
– The update aims to streamline workflows for developers and will eventually be added to consumer apps like ChatGPT
OpenAI has released an update to its GPT-4-Turbo artificial intelligence model, which includes AI vision technology for analyzing video, image, and audio content. Initially available only to developers, these features will eventually come to ChatGPT. This is the first time GPT-4-Turbo with vision technology has been accessible to third party developers, opening the door for new apps and services in areas like fashion, coding, and gaming. The new model includes an expanded knowledge cut-off date of December 2023.
The focus of GPT-4-Turbo is on improving developer workflows by streamlining processes and creating more efficient apps. The future plan is to expand the model’s vision analysis capabilities and integrate them into consumer apps like ChatGPT. This aligns with Google’s approach with Gemini Pro 1.5, which is currently limited to developer platforms. The update aims to enhance enterprise customer experience until GPT-5 is released.
While GPT-4 hasn’t performed well in recent benchmark tests against newer models like Claude 3 Opus and Google’s Gemini, the updates are expected to address this and introduce compelling features for enterprise users. The model continues to offer a 128,000 token context window, suitable for most use cases. OpenAI has previously focused on audio analysis, with the new update expanding video capabilities to a broader user base through ChatGPT. Users may soon be able to upload short video clips for the AI to summarize or identify key moments.