Audio AI by OpenAI
As discussed in my last blog on ‘Emerging AI Trends in 2025’, Audio AI looks to be the hottest trend right now. OpenAI has just now released OpenAI FM, which is a playground to test out OpenAI’s text-to-speech models, and they are just amazing.
https://medium.com/media/74912389276383a2791dc8811ae10ab9/href
What is OpenAI FM?
OpenAI FM (Foundation Models) refers to a collection of advanced AI models that facilitate a wide range of tasks, from text processing to speech recognition. These models are designed to be highly intelligent, customizable, and versatile, making them suitable for applications in customer support, content creation, and more.
Data Science in Your Pocket – No Rocket Science
OpenAI FM Playground Key Features
his UI from OpenAI.fm is an interactive demo for developers to test OpenAI’s new text-to-speech (TTS) model. Here are the key features available in this interface:

1. Voice Selection
Users can choose from a variety of voices, such as Alloy, Ash, Ballad, Coral, Echo, Fable, Onyx, Nova, Sage, Shimmer, and Verse.
Each voice likely has distinct characteristics, including pitch, tone, and style.
Some voices have a diamond icon (✦), possibly indicating premium or expressive voices.
A randomization button (shuffle icon) allows users to get a random voice selection.
2. Vibe Selection
- Users can select an emotional tone or style, like:
Medieval Knight
Calm
Emo Teenager
Serene
Patient Teacher
- These affect the speech’s intonation, rhythm, and emotion, making AI-generated voices more engaging and dynamic.
- A red dot indicates the currently selected vibe (Medieval Knight in this case).
3. Script Input
Users can type or paste text into the provided field to be converted into speech.
The example script in the screenshot mimics an old-fashioned, medieval style of speech.
4. Playback & Controls
Play Button: Starts the TTS conversion and plays the generated speech.
Download Button: Allows users to download the generated speech as an audio file.
Share Button: Provides an option to share the generated TTS output.
5. Developer Mode Toggle
A toggle switch (top right), possibly enabling access to developer settings, APIs, or raw JSON output for integrating TTS features into applications.
Which Models Are Used in OpenAI FM?
OpenAI FM includes several powerful models:
GPT-4o: The latest and most advanced language model, capable of superior text comprehension and generation.
GPT-4o-Transcribe & GPT-4o-Mini-Transcribe: Cutting-edge speech-to-text models designed for high accuracy, even in noisy environments.
GPT-4o-Mini-TTS: A text-to-speech model with enhanced customization options, allowing AI-generated speech to match specific tones and expressions.
How to Use OpenAI FM?
- Via OpenAI Playground: The easiest way to experiment with OpenAI FM is through the OpenAI Playground, which offers a simple interface for testing different models.
- Using the API: Developers can integrate OpenAI FM’s capabilities into their applications by leveraging OpenAI’s API.
- With Agents SDK: OpenAI has introduced an SDK to help developers create conversational agents with advanced speech and text functionalities.
Wrapping It Up — Audio AI Is Here to Stay!
If there’s one thing that’s clear, it’s that Audio AI is about to take center stage in 2025. With OpenAI FM, we’re seeing the next evolution in text-to-speech technology, making AI voices more natural, expressive, and customizable than ever before. Whether you’re a developer building voice-first applications, a content creator looking for high-quality narration, or just someone who wants to experiment with AI-generated voices, OpenAI FM offers a fun and powerful playground to explore.
So, what’s next? We can expect even more refined speech synthesis, multi-modal interactions, and AI-driven voice assistants that sound indistinguishable from humans. If you haven’t already, head over to OpenAI.fm and give it a spin — because the future of Audio AI isn’t coming, it’s already here!
OpenAI FM: OpenAI releases text-speech model playground was originally published in Data Science in your pocket on Medium, where people are continuing the conversation by highlighting and responding to this story.