Method List
-
#access_token LlmGateway::Clients::ClaudeCode::TokenManager
-
#access_token LlmGateway::Clients::OpenAI::TokenManager
-
#account_id LlmGateway::Clients::OpenAI
-
#account_id LlmGateway::Clients::OpenAI::TokenManager
-
after_execute LlmGateway::Prompt
-
after_execute_callbacks LlmGateway::Prompt
-
#api_key LlmGateway::BaseClient
-
#base_endpoint LlmGateway::BaseClient
-
before_execute LlmGateway::Prompt
-
before_execute_callbacks LlmGateway::Prompt
-
blank? LlmGateway::Utils
-
#blocks StreamAccumulator
-
build_client LlmGateway::Client
-
build_provider LlmGateway
-
cache LlmGateway::Tool
-
chat LlmGateway::Client
-
#chat LlmGateway::Clients::Groq
-
#chat LlmGateway::Clients::OpenAI
-
#chat LlmGateway::Adapters::Adapter
-
#chat LlmGateway::Clients::Anthropic
-
#chat_codex LlmGateway::Clients::OpenAI
-
#client LlmGateway::Adapters::Adapter
-
#client_id LlmGateway::Clients::ClaudeCode::OAuthFlow
-
#client_id LlmGateway::Clients::OpenAI::OAuthFlow
-
#client_id LlmGateway::Clients::ClaudeCode::TokenManager
-
#client_id LlmGateway::Clients::OpenAI::TokenManager
-
#client_secret LlmGateway::Clients::ClaudeCode::TokenManager
-
#code LlmGateway::Errors::ClientError
-
configure LlmGateway
-
configured_clients LlmGateway
-
context_overflow_message? LlmGateway::Errors
-
deep_symbolize_keys LlmGateway::Utils
-
definition LlmGateway::Tool
-
description LlmGateway::Tool
-
#direction LlmGateway::Adapters::Anthropic::BidirectionalMessageMapper
-
#direction LlmGateway::Adapters::OpenAI::ChatCompletions::BidirectionalMessageMapper
-
download_file LlmGateway::Client
-
#download_file LlmGateway::Clients::OpenAI
-
#download_file LlmGateway::Adapters::Adapter
-
#download_file LlmGateway::Clients::Anthropic
-
#ensure_valid_token LlmGateway::Clients::ClaudeCode::TokenManager
-
#ensure_valid_token LlmGateway::Clients::OpenAI::TokenManager
-
#exchange_code LlmGateway::Clients::ClaudeCode::OAuthFlow
-
#exchange_code LlmGateway::Clients::OpenAI::OAuthFlow
-
#execute LlmGateway::Tool
-
#expires_at LlmGateway::Clients::ClaudeCode::TokenManager
-
#expires_at LlmGateway::Clients::OpenAI::TokenManager
-
extract_account_id_from_token LlmGateway::Clients::OpenAI::OAuthFlow
-
find_tool LlmGateway::Prompt
-
#generate_embeddings LlmGateway::Clients::OpenAI
-
#get LlmGateway::BaseClient
-
#get_oauth_access_token LlmGateway::Clients::OpenAI
-
#get_oauth_access_token LlmGateway::Clients::Anthropic
-
included LlmGateway::Adapters::OpenAI::PromptCacheOptionMapper
-
inherited LlmGateway::Prompt
-
#initialize LlmGateway::Tool
-
#initialize LlmGateway::Errors::ClientError
-
#initialize LlmGateway::Prompt
-
#initialize LlmGateway::BaseClient
-
#initialize LlmGateway::Clients::Groq
-
#initialize LlmGateway::Clients::OpenAI
-
#initialize LlmGateway::Adapters::Adapter
-
#initialize LlmGateway::Clients::Anthropic
-
#initialize StreamAccumulator
-
#initialize LlmGateway::Clients::ClaudeCode::OAuthFlow
-
#initialize LlmGateway::Clients::OpenAI::OAuthFlow
-
#initialize LlmGateway::Clients::ClaudeCode::TokenManager
-
#initialize LlmGateway::Clients::OpenAI::TokenManager
-
#initialize LlmGateway::Adapters::Anthropic::BidirectionalMessageMapper
-
#initialize LlmGateway::Adapters::OpenAI::ChatCompletions::BidirectionalMessageMapper
-
input_schema LlmGateway::Tool
-
#login LlmGateway::Clients::OpenAI::OAuthFlow
-
map LlmGateway::Adapters::OptionMapper
-
map LlmGateway::Adapters::Groq::OptionMapper
-
map LlmGateway::Adapters::Anthropic::InputMapper
-
map LlmGateway::Adapters::Anthropic::FileOutputMapper
-
map LlmGateway::Adapters::Anthropic::OutputMapper
-
#map LlmGateway::Adapters::Anthropic::StreamMapper
-
map LlmGateway::Adapters::AnthropicOptionMapper
-
map LlmGateway::Adapters::OpenAI::FileOutputMapper
-
map LlmGateway::Adapters::OpenAICodex::OptionMapper
-
map LlmGateway::Adapters::OpenAI::Responses::OptionMapper
-
map LlmGateway::Adapters::OpenAI::Responses::OutputMapper
-
#map LlmGateway::Adapters::OpenAI::Responses::StreamMapper
-
map LlmGateway::Adapters::OpenAI::ChatCompletions::InputMapper
-
map LlmGateway::Adapters::OpenAI::ChatCompletions::OptionMapper
-
map LlmGateway::Adapters::OpenAI::ChatCompletions::OutputMapper
-
#map LlmGateway::Adapters::OpenAI::ChatCompletions::StreamMapper
-
map_assistant_history_message LlmGateway::Adapters::OpenAI::Responses::InputMapper
-
#map_cache_key! LlmGateway::Adapters::OpenAI::PromptCacheOptionMapper
-
map_choices LlmGateway::Adapters::Anthropic::OutputMapper
-
map_choices LlmGateway::Adapters::OpenAI::Responses::OutputMapper
-
map_choices LlmGateway::Adapters::OpenAI::ChatCompletions::OutputMapper
-
#map_content LlmGateway::Adapters::Anthropic::BidirectionalMessageMapper
-
#map_content LlmGateway::Adapters::OpenAI::Responses::BidirectionalMessageMapper
-
#map_content LlmGateway::Adapters::OpenAI::ChatCompletions::BidirectionalMessageMapper
-
map_messages LlmGateway::Adapters::Anthropic::InputMapper
-
map_messages LlmGateway::Adapters::OpenAICodex::InputMapper
-
map_messages LlmGateway::Adapters::OpenAI::Responses::InputMapper
-
map_messages LlmGateway::Adapters::OpenAI::ChatCompletions::InputMapper
-
#map_prompt_cache_retention! LlmGateway::Adapters::OpenAI::PromptCacheOptionMapper
-
map_system LlmGateway::Adapters::Anthropic::InputMapper
-
map_system LlmGateway::Adapters::OpenAI::ChatCompletions::InputMapper
-
map_tools LlmGateway::Adapters::OpenAI::Responses::InputMapper
-
map_tools LlmGateway::Adapters::OpenAI::ChatCompletions::InputMapper
-
#message_hash StreamAccumulator
-
message_mapper LlmGateway::Adapters::OpenAI::Responses::InputMapper
-
#model LlmGateway::Prompt
-
#model_key LlmGateway::BaseClient
-
name LlmGateway::Tool
-
new AssistantMessage
-
normalize_output_config LlmGateway::Adapters::AnthropicOptionMapper
-
normalize_reasoning LlmGateway::Adapters::AnthropicOptionMapper
-
normalize_reasoning LlmGateway::Adapters::OpenAI::Responses::OptionMapper
-
normalize_reasoning_effort LlmGateway::Adapters::OpenAI::ChatCompletions::OptionMapper
-
normalize_response_format LlmGateway::Adapters::Groq::OptionMapper
-
#on_token_refresh LlmGateway::Clients::ClaudeCode::TokenManager
-
#on_token_refresh LlmGateway::Clients::OpenAI::TokenManager
-
#parse_callback LlmGateway::Clients::ClaudeCode::OAuthFlow
-
#parse_callback LlmGateway::Clients::OpenAI::OAuthFlow
-
#post LlmGateway::Prompt
-
#post LlmGateway::BaseClient
-
#post_file LlmGateway::BaseClient
-
#post_stream LlmGateway::BaseClient
-
present? LlmGateway::Utils
-
provider_from_model LlmGateway::Client
-
provider_id_from_client LlmGateway::Client
-
providers LlmGateway::ProviderRegistry
-
#push StreamAccumulator
-
#redirect_uri LlmGateway::Clients::ClaudeCode::OAuthFlow
-
#redirect_uri LlmGateway::Clients::OpenAI::OAuthFlow
-
refresh_access_token LlmGateway::Clients::OpenAI::OAuthFlow
-
#refresh_access_token LlmGateway::Clients::ClaudeCode::TokenManager
-
#refresh_access_token! LlmGateway::Clients::OpenAI::TokenManager
-
#refresh_token LlmGateway::Clients::ClaudeCode::TokenManager
-
#refresh_token LlmGateway::Clients::OpenAI::TokenManager
-
register LlmGateway::ProviderRegistry
-
registered? LlmGateway::ProviderRegistry
-
reset! LlmGateway::ProviderRegistry
-
reset_configuration! LlmGateway
-
resolve LlmGateway::ProviderRegistry
-
responses LlmGateway::Client
-
#responses LlmGateway::Clients::OpenAI
-
#result StreamAccumulator
-
#run LlmGateway::Prompt
-
sanitize LlmGateway::Adapters::InputMessageSanitizer
-
sanitize LlmGateway::Adapters::OpenAI::ChatCompletions::InputMessageSanitizer
-
#scope LlmGateway::Clients::OpenAI::OAuthFlow
-
#scopes LlmGateway::Clients::ClaudeCode::OAuthFlow
-
#start LlmGateway::Clients::ClaudeCode::OAuthFlow
-
#start LlmGateway::Clients::OpenAI::OAuthFlow
-
#stream LlmGateway::Clients::OpenAI
-
#stream LlmGateway::Adapters::Adapter
-
#stream LlmGateway::Clients::Anthropic
-
#stream_codex LlmGateway::Clients::OpenAI
-
#stream_responses LlmGateway::Clients::OpenAI
-
#system_prompt LlmGateway::Prompt
-
#to_h TextContent
-
#to_h ReasoningContent
-
#to_h ToolCall
-
#to_h AssistantMessage
-
#token_expired? LlmGateway::Clients::ClaudeCode::TokenManager
-
#token_expired? LlmGateway::Clients::OpenAI::TokenManager
-
tool_name LlmGateway::Tool
-
#tools LlmGateway::Prompt
-
upload_file LlmGateway::Client
-
#upload_file LlmGateway::Clients::OpenAI
-
#upload_file LlmGateway::Adapters::Adapter
-
#upload_file LlmGateway::Clients::Anthropic
-
#usage_hash StreamAccumulator