← Back to all briefings
AI 5 min read Published Updated Credibility 40/100

OpenAI Announces GPT-4 Turbo, Assistants API, and Custom GPTs

At OpenAI DevDay, the company unveils GPT-4 Turbo with 128K context window, knowledge through April 2023, and 3x lower pricing. New Assistants API provides stateful persistent threads, code interpreter, and retrieval capabilities. Custom GPTs enable no-code AI agent creation. The releases accelerate AI application development and reduce costs for enterprise deployment.

Horizontal bar chart of credibility scores per cited source.
Credibility scores for every source cited in this briefing. Source data (JSON)

OpenAI hosted its first developer conference, DevDay, on November 6, 2023, announcing major platform updates including GPT-4 Turbo with 128K context window, Assistants API for building AI agents, and Custom GPTs for no-code AI creation. The company reduced API pricing by up to 3x while improving capabilities, addressing developer feedback about costs and technical limitations. The announcements position OpenAI to accelerate enterprise AI adoption and expand its lead in generative AI infrastructure.

GPT-4 Turbo Capabilities and Pricing

GPT-4 Turbo (gpt-4-1106-preview) extends context window from 8K/32K tokens to 128K tokens—equivalent to 300 pages of text or an entire novel. The model's knowledge cutoff updates from September 2021 to April 2023, providing more current information. GPT-4 Turbo improves instruction following, reduces 'laziness' in task completion, and better adheres to response format requirements including JSON mode for structured outputs.

OpenAI reduced GPT-4 Turbo pricing to $0.01 per 1K input tokens and $0.03 per 1K output tokens, representing 3x cost reduction compared to GPT-4. The gpt-3.5-turbo-1106 update (16K context) costs $0.001 input and $0.002 output per 1K tokens. Lower pricing makes AI integration economically viable for high-volume applications previously constrained by costs. Developers can process entire codebases, lengthy documents, and multi-turn conversations without context limitations.

Assistants API Architecture

The Assistants API introduces persistent threads storing conversation history, eliminating need for developers to manage state and context. Assistants combine base models with instructions (system prompts), tools (code interpreter, retrieval, function calling), and persistent thread storage. Developers create assistants once and instantiate multiple threads per user, simplifying application architecture.

Code Interpreter enables assistants to execute Python code in sandboxed environments, process uploaded files (CSVs, documents, images), generate data visualizations, and perform calculations. Retrieval augmented generation (RAG) integrates with assistants, automatically chunking and embedding documents, storing in vector database, and retrieving relevant context. Function calling allows assistants to invoke external APIs, query databases, and interact with business systems.

Custom GPTs and GPT Store

Custom GPTs enable non-technical users to create specialized AI agents through conversational interface without coding. Users define instructions, upload knowledge files, and configure capabilities (web browsing, DALL-E image generation, code execution). Example custom GPTs include Canva design assistant, Zapier automation connector, and academic paper analyzer.

OpenAI announced GPT Store launching later in November, creating marketplace for custom GPTs. Creators can publish GPTs publicly, share with organizations, or keep private. OpenAI plans revenue sharing based on GPT usage, creating economic model for AI agent development. Enterprise customers receive tools to create internal GPT libraries with organizational knowledge while maintaining data privacy.

Multimodal Capabilities and Vision API

OpenAI made GPT-4 with Vision (GPT-4V) generally available through API, enabling applications processing images alongside text. Developers send images via URL or base64 encoding, with models analyzing visual content, reading text in images, describing scenes, and answering questions about images. Pricing is input tokens plus image tokens calculated by image size and detail level.

Use cases include document processing extracting structured data from forms and invoices, visual Q&A for e-commerce and customer support, accessibility tools describing images for visually impaired users, and manufacturing quality control analyzing product images. DALL-E 3 API enables image generation directly from applications, priced at $0.040-$0.120 per image depending on resolution.

Text-to-Speech and Whisper V3

OpenAI released Text-to-Speech API with six natural-sounding voices optimized for real-time streaming. Priced at $0.015 per 1K characters for standard quality and $0.030 for HD quality, the API enables voice assistants, audiobook narration, content accessibility, and IVR systems. The API streams audio in chunks for low-latency applications requiring immediate audio feedback.

Whisper large-v3 model improves speech recognition accuracy across languages while maintaining real-time transcription speed. The API supports 57 languages with improved handling of accents, background noise, and technical terminology. Integration with Assistants API enables voice-driven AI agents combining speech recognition, language model reasoning, and text-to-speech responses.

Enterprise and Safety Features

OpenAI introduced copyright shield protecting customers from legal claims if AI-generated content infringes intellectual property rights. The company committed to defending enterprise customers and paying damages, addressing major concern deterring AI adoption. Custom models program enables enterprises to work with OpenAI researchers to fine-tune GPT-4 on proprietary data, with dedicated compute capacity.

The platform added fine-tuning support for GPT-4, enabling domain-specific model adaptation. Organizations upload training examples demonstrating desired outputs, with OpenAI managing infrastructure and optimization. Fine-tuning improves accuracy on specialized tasks, enables brand voice consistency, and reduces prompt engineering requirements. Pricing is compute-based during training plus per-token inference costs for custom models.

Competitive Landscape Impact

The announcements intensified pressure on competitors including Anthropic (Claude), Google (Bard/Gemini), and Meta (Llama). OpenAI's 3x price reduction forced competitors to lower pricing or differentiate on capabilities. Anthropic countered with Claude 2.1 featuring 200K context window, while Google prepared Gemini launch. The rapid capability improvements and cost reductions accelerated enterprise AI adoption timelines.

Developers previously considering alternatives due to GPT-4 costs reevaluated decisions given improved price-performance. The Assistants API reduced development complexity compared to building equivalent functionality with competitors' APIs. Custom GPTs democratized AI agent creation, expanding addressable market beyond technical developers to domain experts and business users.

Strategic Implications for CTIOs

CTIOs should reevaluate AI roadmaps considering cost reductions, longer context windows, and Assistants API. Applications previously too expensive become economically viable. The 128K context window enables new use cases including full codebase analysis, comprehensive document processing, and complex multi-step reasoning over extensive context.

Organizations should pilot Assistants API for customer support, technical documentation Q&A, and internal knowledge management. Custom GPTs enable rapid experimentation with AI agents across business functions. Technical teams should develop governance frameworks for GPT creation, establish review processes, and implement monitoring for organizational GPTs. CTIOs must balance innovation velocity against risk management, ensuring appropriate guardrails while enabling business units to experiment with AI capabilities.

Horizontal bar chart of credibility scores per cited source.
Credibility scores for every source cited in this briefing. Source data (JSON)

Continue in the AI pillar

Return to the hub for curated research and deep-dive guides.

Visit pillar hub

Latest guides

  • GPT-4 Turbo
  • Assistants API
  • custom GPTs
  • OpenAI
Back to curated briefings

Comments

Community

We publish only high-quality, respectful contributions. Every submission is reviewed for clarity, sourcing, and safety before it appears here.

    Share your perspective

    Submissions showing "Awaiting moderation" are in review. Spam, low-effort posts, or unverifiable claims will be rejected. We verify submissions with the email you provide, and we never publish or sell that address.

    Verification

    Complete the CAPTCHA to submit.