Class: OpenAI::Models::Realtime::RealtimeTranslationInputAudioBufferAppendEvent
- Inherits:
-
Internal::Type::BaseModel
- Object
- Internal::Type::BaseModel
- OpenAI::Models::Realtime::RealtimeTranslationInputAudioBufferAppendEvent
- Defined in:
- lib/openai/models/realtime/realtime_translation_input_audio_buffer_append_event.rb
Instance Attribute Summary collapse
-
#audio ⇒ String
Base64-encoded 24 kHz PCM16 mono audio bytes.
-
#event_id ⇒ String?
Optional client-generated ID used to identify this event.
-
#type ⇒ Symbol, :"session.input_audio_buffer.append"
The event type, must be ‘session.input_audio_buffer.append`.
Instance Method Summary collapse
-
#initialize(audio:, event_id: nil, type: :"session.input_audio_buffer.append") ⇒ Object
constructor
Send this event to append audio bytes to the translation session input audio buffer.
Methods inherited from Internal::Type::BaseModel
==, #==, #[], coerce, #deconstruct_keys, #deep_to_h, dump, fields, hash, #hash, inherited, inspect, #inspect, known_fields, optional, recursively_to_h, required, #to_h, #to_json, #to_s, to_sorbet_type, #to_yaml
Methods included from Internal::Type::Converter
#coerce, coerce, #dump, dump, #inspect, inspect, meta_info, new_coerce_state, type_info
Methods included from Internal::Util::SorbetRuntimeSupport
#const_missing, #define_sorbet_constant!, #sorbet_constant_defined?, #to_sorbet_type, to_sorbet_type
Constructor Details
#initialize(audio:, event_id: nil, type: :"session.input_audio_buffer.append") ⇒ Object
Send this event to append audio bytes to the translation session input audio buffer.
WebSocket translation sessions accept base64-encoded 24 kHz PCM16 mono little-endian raw audio bytes. Unsupported websocket audio formats return a validation error because lower-quality audio materially degrades translation quality.
Translation consumes 200 ms engine frames. For best realtime behavior, append audio in 200 ms chunks. If a chunk is shorter, the server buffers it until it has enough audio for one frame. If a chunk is longer, the server splits it into 200 ms frames and enqueues them back-to-back.
Keep appending silence while the session is active. If a client stops sending audio and later resumes, model time treats the resumed audio as contiguous with the previous audio rather than as a real-world pause.
|
|
# File 'lib/openai/models/realtime/realtime_translation_input_audio_buffer_append_event.rb', line 25
|
Instance Attribute Details
#audio ⇒ String
Base64-encoded 24 kHz PCM16 mono audio bytes.
11 |
# File 'lib/openai/models/realtime/realtime_translation_input_audio_buffer_append_event.rb', line 11 required :audio, String |
#event_id ⇒ String?
Optional client-generated ID used to identify this event.
23 |
# File 'lib/openai/models/realtime/realtime_translation_input_audio_buffer_append_event.rb', line 23 optional :event_id, String |
#type ⇒ Symbol, :"session.input_audio_buffer.append"
The event type, must be ‘session.input_audio_buffer.append`.
17 |
# File 'lib/openai/models/realtime/realtime_translation_input_audio_buffer_append_event.rb', line 17 required :type, const: :"session.input_audio_buffer.append" |