Class: OpenAI::Models::Realtime::RealtimeTranslationInputAudioBufferAppendEvent

Inherits:
Internal::Type::BaseModel show all
Defined in:
lib/openai/models/realtime/realtime_translation_input_audio_buffer_append_event.rb

Instance Attribute Summary collapse

Instance Method Summary collapse

Methods inherited from Internal::Type::BaseModel

==, #==, #[], coerce, #deconstruct_keys, #deep_to_h, dump, fields, hash, #hash, inherited, inspect, #inspect, known_fields, optional, recursively_to_h, required, #to_h, #to_json, #to_s, to_sorbet_type, #to_yaml

Methods included from Internal::Type::Converter

#coerce, coerce, #dump, dump, #inspect, inspect, meta_info, new_coerce_state, type_info

Methods included from Internal::Util::SorbetRuntimeSupport

#const_missing, #define_sorbet_constant!, #sorbet_constant_defined?, #to_sorbet_type, to_sorbet_type

Constructor Details

#initialize(audio:, event_id: nil, type: :"session.input_audio_buffer.append") ⇒ Object

Send this event to append audio bytes to the translation session input audio buffer.

WebSocket translation sessions accept base64-encoded 24 kHz PCM16 mono little-endian raw audio bytes. Unsupported websocket audio formats return a validation error because lower-quality audio materially degrades translation quality.

Translation consumes 200 ms engine frames. For best realtime behavior, append audio in 200 ms chunks. If a chunk is shorter, the server buffers it until it has enough audio for one frame. If a chunk is longer, the server splits it into 200 ms frames and enqueues them back-to-back.

Keep appending silence while the session is active. If a client stops sending audio and later resumes, model time treats the resumed audio as contiguous with the previous audio rather than as a real-world pause.

Parameters:

  • audio (String)

    Base64-encoded 24 kHz PCM16 mono audio bytes.

  • event_id (String) (defaults to: nil)

    Optional client-generated ID used to identify this event.

  • type (Symbol, :"session.input_audio_buffer.append") (defaults to: :"session.input_audio_buffer.append")

    The event type, must be ‘session.input_audio_buffer.append`.



# File 'lib/openai/models/realtime/realtime_translation_input_audio_buffer_append_event.rb', line 25

Instance Attribute Details

#audioString

Base64-encoded 24 kHz PCM16 mono audio bytes.

Returns:

  • (String)


11
# File 'lib/openai/models/realtime/realtime_translation_input_audio_buffer_append_event.rb', line 11

required :audio, String

#event_idString?

Optional client-generated ID used to identify this event.

Returns:

  • (String, nil)


23
# File 'lib/openai/models/realtime/realtime_translation_input_audio_buffer_append_event.rb', line 23

optional :event_id, String

#typeSymbol, :"session.input_audio_buffer.append"

The event type, must be ‘session.input_audio_buffer.append`.

Returns:

  • (Symbol, :"session.input_audio_buffer.append")


17
# File 'lib/openai/models/realtime/realtime_translation_input_audio_buffer_append_event.rb', line 17

required :type, const: :"session.input_audio_buffer.append"