Search Results for

    Show / Hide Table of Contents

    Class RealtimeEvent.Server.InputAudioBuffer

    Inheritance
    object
    RealtimeEvent.Server.InputAudioBuffer
    Inherited Members
    object.Equals(object)
    object.Equals(object, object)
    object.GetHashCode()
    object.GetType()
    object.MemberwiseClone()
    object.ReferenceEquals(object, object)
    object.ToString()
    Namespace: Glitch9.AIDevKit.OpenAI.Realtime
    Assembly: Glitch9.AIDevKit.Provider.OpenAI.dll
    Syntax
    public static class RealtimeEvent.Server.InputAudioBuffer

    Fields

    | Edit this page View Source

    Cleared

    Returned when the input audio buffer is cleared by the client.

    Declaration
    public const string Cleared = "input_audio_buffer.cleared"
    Field Value
    Type Description
    string
    | Edit this page View Source

    Committed

    Returned when an input audio buffer is committed, either by the client or automatically in server VAD mode.

    Declaration
    public const string Committed = "input_audio_buffer.committed"
    Field Value
    Type Description
    string
    | Edit this page View Source

    SpeechStarted

    Returned in server turn detection mode when speech is detected.

    Declaration
    public const string SpeechStarted = "input_audio_buffer.speech_started"
    Field Value
    Type Description
    string
    | Edit this page View Source

    SpeechStopped

    Returned in server turn detection mode when speech stops.

    Declaration
    public const string SpeechStopped = "input_audio_buffer.speech_stopped"
    Field Value
    Type Description
    string
    | Edit this page View Source

    TimeoutTriggered

    Added 2024-09-19
    Returned when the Server VAD timeout is triggered for the input audio buffer. This is configured with idle_timeout_ms in the turn_detection settings of the session, and it indicates that there hasn't been any speech detected for the configured duration.

    The audio_start_ms and audio_end_ms fields indicate the segment of audio after the last model response up to the triggering time, as an offset from the beginning of audio written to the input audio buffer. This means it demarcates the segment of audio that was silent and the difference between the start and end values will roughly match the configured timeout.

    The empty audio will be committed to the conversation as an input_audio item (there will be a input_audio_buffer.committed event) and a model response will be generated. There may be speech that didn't trigger VAD but is still detected by the model, so the model may respond with something relevant to the conversation or a prompt to continue speaking.

    Declaration
    public const string TimeoutTriggered = "input_audio_buffer.timeout_triggered"
    Field Value
    Type Description
    string
    • Edit this page
    • View Source
    In this article
    Back to top AI DevKit Documentation