View Source ExOpenAI.Components.AudioTranscription (ex_openai.ex v2.0.0-beta2)

Module for representing the OpenAI schema AudioTranscription.

Fields

  • :language - optional - String.t()
    The language of the input audio. Supplying the input language in ISO-639-1 (e.g. en) format will improve accuracy and latency.

  • :model - optional - String.t() | :"whisper-1" | :"gpt-4o-mini-transcribe" | :"gpt-4o-mini-transcribe-2025-12-15" | :"gpt-4o-transcribe" | :"gpt-4o-transcribe-diarize"
    The model to use for transcription. Current options are whisper-1, gpt-4o-mini-transcribe, gpt-4o-mini-transcribe-2025-12-15, gpt-4o-transcribe, and gpt-4o-transcribe-diarize. Use gpt-4o-transcribe-diarize when you need diarization with speaker labels.

  • :prompt - optional - String.t()
    An optional text to guide the model's style or continue a previous audio segment. For whisper-1, the prompt is a list of keywords. For gpt-4o-transcribe models (excluding gpt-4o-transcribe-diarize), the prompt is a free text string, for example "expect words related to technology".

Summary

Types

@type t() :: %ExOpenAI.Components.AudioTranscription{
  language: String.t() | nil,
  model:
    (String.t()
     | (((:"whisper-1" | :"gpt-4o-mini-transcribe")
         | :"gpt-4o-mini-transcribe-2025-12-15")
        | :"gpt-4o-transcribe")
     | :"gpt-4o-transcribe-diarize")
    | nil,
  prompt: String.t() | nil
}