OpenAI has introduced its latest artificial intelligence models, designed to interpret and manipulate drawings and diagrams irrespective of the creator’s artistic skill. The new models, known as o3 and its smaller counterpart o4-mini, expand on the capabilities of the September release, o1, which concentrated on addressing complex problems through multi-step reasoning.

Users can now upload sketches ranging from whiteboard scribbles to casual napkin drawings, with the AI able to analyse these visual inputs effectively. The technology transcends simple recognition, enabling the AI models to perform operations such as rotating, zooming, and editing images autonomously.

Since the global surge in popularity following the release of ChatGPT in 2022, OpenAI has accelerated its development efforts beyond text-based models, venturing into multimedia applications. The company is positioning itself in a competitive landscape with rivals including Google, Anthropic, and Elon Musk’s xAI.

A significant feature of these new models is the integration of visual data directly into the AI’s reasoning processes, marking a notable shift from merely processing images to actively working with visual information in context.

Recent financial developments highlight investor confidence in OpenAI, with the company’s valuation reaching an estimated $300 billion as of last month. The same period saw their image generator gain viral popularity on social media for producing artwork inspired by Studio Ghibli.

Access to the o3 and o4-mini models has been granted immediately to subscribers of ChatGPT Plus, Pro, and Team plans following their announcement. However, OpenAI continues to face criticism over its naming conventions for AI products, with CEO Sam Altman acknowledging the issue and pledging to improve naming clarity by the summer.

Ongoing safety concerns accompany these technological advances. OpenAI’s updated policies now suggest possible relaxation of safety standards if competitors deploy high-risk AI systems without similar safeguards, a strategy that has prompted apprehension among industry observers. Transparency remains a contentious topic as well; the company recently removed safety testing mandates for some fine-tuned models and has refrained from releasing detailed documentation for GPT-4.1. Additionally, the launch of their Deep Research tool preceded the release of its safety documentation by several weeks, further intensifying scrutiny over OpenAI’s approach to responsible AI development.

Source: Noah Wire Services