×
OpenAI unveils advanced visual AI models with superior image processing capabilities
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

OpenAI‘s latest AI models, o3 and o4-mini, represent a significant leap forward in capabilities for ChatGPT, offering enhanced multimodal understanding and greater autonomy when solving complex problems. These new “brains” for ChatGPT arrive as the company benefits from unprecedented computing resources, signaling a pivotal moment in AI development where processing power constraints no longer limit innovation potential.

The big picture: OpenAI has released two advanced AI models, o3 and o4-mini, exclusively for ChatGPT Plus, Pro, and Team users, while announcing the retirement of GPT-4 by month’s end.

  • These models feature significantly improved ability to interpret poor-quality images, including blurry photos and handwritten notes, enhancing their multimodal capabilities.
  • The update comes days after ChatGPT received major memory enhancements designed to make interactions more personalized and context-aware.

Key capabilities: The new models can independently utilize all of ChatGPT’s integrated tools without requiring explicit user instructions.

  • For the first time, the AI can autonomously decide when to deploy the browser, Python code interpreter, image generation, or image analysis tools based on the task at hand.
  • Users can now interact with images in real-time, performing operations like rotating or zooming during problem-solving sessions.

Developer focus: OpenAI is launching Codex CLI, a new open-source coding agent that runs locally in terminal windows and works with these new models.

  • The company is offering $1 million in API credits distributed in $25,000 increments to encourage developers to build with these tools.
  • A more advanced o3-pro model is scheduled for release in the coming weeks, while Pro plan users can continue using the existing o1-pro model.

Why this matters: OpenAI CEO Sam Altman recently revealed the company is no longer “compute constrained,” marking a significant shift in development capabilities.

  • This newfound processing capacity enables faster development cycles and the creation of models capable of handling increasingly complex tasks.
  • The strategic focus on autonomy and tool integration points toward OpenAI’s vision of AI systems that function less as simple assistants and more as collaborative partners for creative and analytical work.
OpenAI just launched its smartest AI yet that can think with images — here's how to try it

Recent News

AI firms Musk’s xAI, TWG Global and Palantir target finance industry

The partnership will integrate xAI's Grok language models with Palantir's analytics to enhance data-driven decision making in finance and insurance operations.

Suno 4.5 AI music creator launches with major upgrades

Suno's latest AI music tool brings significantly better vocals and genre handling while doubling maximum song length to eight minutes.

Oura Ring unveils major feature update for users

Oura Ring now integrates with Dexcom's continuous glucose monitors while introducing AI-powered meal tracking to help users better understand connections between sleep, activity, and metabolic health.