Method List
-
adapted_registry_tools Legion::LLM::Inference
-
#adjustment Legion::LLM::Router::HealthTracker
-
after_chat Legion::LLM::Hooks
-
agent Legion::LLM
-
#agent Legion::LLM::Inference::Request
-
#agent Legion::LLM::Inference::Response
-
agent_queue_name Legion::LLM::Fleet::ReplyDispatcher
-
all Legion::LLM::Skills::Registry
-
all_overrides Legion::LLM::Tools::Confidence
-
analyze_for_knowledge Legion::LLM::Hooks::Reflection
-
anthropic_api_key Legion::LLM::Call::ClaudeConfigLoader
-
api_defaults Legion::LLM::Settings
-
#app_id Legion::LLM::Transport::Message
-
append Legion::LLM::Inference::Conversation
-
apply_api_keys Legion::LLM::Call::ClaudeConfigLoader
-
#apply_cache_control Legion::LLM::Inference::Steps::PromptCache
-
apply_claude_config Legion::LLM::Call::ClaudeConfigLoader
-
apply_codex_config Legion::LLM::Call::CodexConfigLoader
-
#apply_conversation_breakpoint Legion::LLM::Inference::Steps::PromptCache
-
apply_credential_to_config Legion::LLM::Call::Providers
-
apply_credential_to_rubyllm Legion::LLM::Call::Providers
-
apply_model_preference Legion::LLM::Call::ClaudeConfigLoader
-
apply_provider_config Legion::LLM::Call::Providers
-
apply_response_guards Legion::LLM::Inference
-
arbitrage_defaults Legion::LLM::Settings
-
#arguments Legion::LLM::Types::ToolCall
-
ask Legion::LLM
-
ask Legion::LLM::Inference
-
ask_direct Legion::LLM::Inference
-
assert_external_allowed! Legion::LLM::Inference
-
assign Legion::LLM::Inference::Steps::TierAssigner
-
#at_least? Legion::LLM::Quality::Confidence::Score
-
attach_escalation_history Legion::LLM::Inference
-
attempt_provider_call Legion::LLM::Call::Providers
-
attributes_for Legion::LLM::Inference::Steps::SpanAnnotator
-
#audit Legion::LLM::Inference::Executor
-
#audit Legion::LLM::Inference::Response
-
auto_compact Legion::LLM::Context::Compressor
-
auto_configure_defaults Legion::LLM::Config
-
auto_enable_from_resolved_credentials Legion::LLM::Call::Providers
-
auto_register_providers Legion::LLM::Call::Providers
-
availability_method_for Legion::LLM::Fleet::Handler
-
available Legion::LLM::Call::Registry
-
#available? Legion::LLM::Call::Dispatch
-
available? Legion::LLM::Call::DaemonClient
-
available_memory_mb Legion::LLM::Discovery::System
-
#band Legion::LLM::Quality::Confidence::Score
-
batch_defaults Legion::LLM::Settings
-
#bedrock_bearer_token RubyLLM::Configuration
-
bedrock_bearer_token Legion::LLM::Call::ClaudeConfigLoader
-
before_chat Legion::LLM::Hooks
-
#billing Legion::LLM::Inference::Request
-
#billing Legion::LLM::Inference::Response
-
branch Legion::LLM::Inference::Conversation
-
broker_has_credential? Legion::LLM::Call::Providers
-
budget_defaults Legion::LLM::Settings
-
budget_exceeded_response Legion::LLM::Hooks::BudgetGuard
-
budget_setting Legion::LLM::Hooks::BudgetGuard
-
build Legion::LLM::Types::Message
-
build Legion::LLM::Types::ToolCall
-
build Legion::LLM::Inference::Request
-
build Legion::LLM::Inference::Response
-
build Legion::LLM::Skills::StepResult
-
build Legion::LLM::Skills::SkillRunResult
-
build Legion::LLM::Quality::Confidence::Score
-
build_attempt Legion::LLM::Inference
-
build_cache_key Legion::LLM::Inference
-
build_chain Legion::LLM::Inference::Conversation
-
build_event Legion::LLM::Inference::Steps::Metering
-
build_event Legion::LLM::Inference::AuditPublisher
-
build_md_skill_class Legion::LLM::Skills::DiskLoader
-
build_message_context Legion::LLM::Inference::AuditPublisher
-
build_model_list Legion::LLM::API::OpenAI::Models
-
build_openai_tool_classes Legion::LLM::API::OpenAI::ChatCompletions
-
#build_partner_context Legion::LLM::Inference::Steps::GaiaAdvisory
-
build_response Legion::LLM::Fleet::Handler
-
build_routing_key Legion::LLM::Fleet::Dispatcher
-
build_tool_calls Legion::LLM::API::Translators::OpenAIResponse
-
build_with Legion::LLM::Types::ContentBlock
-
by_trigger Legion::LLM::Skills::Registry
-
#cache Legion::LLM::Inference::Request
-
#cache Legion::LLM::Inference::Response
-
#cache_control Legion::LLM::Types::ContentBlock
-
#cache_read_tokens Legion::LLM::Call::NativeResponseAdapter
-
#cache_read_tokens Legion::LLM::Usage
-
#cache_write_tokens Legion::LLM::Call::NativeResponseAdapter
-
#cache_write_tokens Legion::LLM::Usage
-
cacheable? Legion::LLM::Inference
-
#call Legion::LLM::Inference::Executor
-
call_local_llm Legion::LLM::Fleet::Handler
-
#call_stream Legion::LLM::Inference::Executor
-
#caller Legion::LLM::Inference::Request
-
#caller Legion::LLM::Inference::Response
-
caller_descriptor Legion::LLM::Inference
-
can_embed? Legion::LLM
-
can_embed? Legion::LLM::Discovery
-
cancel_consumer Legion::LLM::Fleet::ReplyDispatcher
-
cancel_skill! Legion::LLM::Inference::Conversation
-
#cause Legion::LLM::Skills::StepError
-
chain_for Legion::LLM::Skills::Registry
-
chat Legion::LLM
-
chat Legion::LLM::Inference
-
chat Legion::LLM::Call::DaemonClient
-
chat Legion::LLM::Inference::GaiaCaller
-
chat_direct Legion::LLM
-
chat_direct Legion::LLM::Inference
-
chat_single Legion::LLM::Inference
-
chat_via_pipeline Legion::LLM::Inference
-
chat_with_escalation Legion::LLM::Inference
-
cheapest_for Legion::LLM::Router::Arbitrage
-
check Legion::LLM::Quality::Checker
-
check_budget Legion::LLM::Hooks::BudgetGuard
-
check_catalog_override Legion::LLM::Tools::Dispatcher
-
check_health Legion::LLM::Call::DaemonClient
-
check_override Legion::LLM::Tools::Dispatcher
-
check_rag_faithfulness Legion::LLM::Hooks::RagGuard
-
#check_rag_faithfulness Legion::LLM::Inference::Steps::RagGuard
-
check_settings_override Legion::LLM::Tools::Dispatcher
-
#circuit_state Legion::LLM::Router::HealthTracker
-
#classification Legion::LLM::Inference::Request
-
#classification Legion::LLM::Inference::Response
-
claude_auto_discover? Legion::LLM::Skills::ExternalDiscovery
-
claude_cli_defaults Legion::LLM::Settings
-
claude_config_path Legion::LLM::Call::ClaudeConfigLoader
-
claude_directories Legion::LLM::Skills::ExternalDiscovery
-
claude_settings_path Legion::LLM::Call::ClaudeConfigLoader
-
cleanup Legion::LLM::Cache::Response
-
clear Legion::LLM::Metering::Recorder
-
clear Legion::LLM::EscalationTracker
-
clear_cancel_flag Legion::LLM::Inference::Conversation
-
clear_history Legion::LLM::Quality::ShadowEval
-
clear_skill_state Legion::LLM::Inference::Conversation
-
#cloud? Legion::LLM::Router::Resolution
-
#code Legion::LLM::Types::ContentBlock
-
codex_auto_discover? Legion::LLM::Skills::ExternalDiscovery
-
codex_directories Legion::LLM::Skills::ExternalDiscovery
-
collect_credential_candidates Legion::LLM::Call::Providers
-
compact_enrichments Legion::LLM::Inference::AuditPublisher
-
compact_timeline Legion::LLM::Inference::AuditPublisher
-
compare Legion::LLM::Quality::ShadowEval
-
complete Legion::LLM::Cache::Response
-
#complete Legion::LLM::Skills::SkillRunResult
-
compliance_defaults Legion::LLM::Settings
-
compress Legion::LLM::Context::Compressor
-
#compress_level Legion::LLM::Router::Resolution
-
condition Legion::LLM::Skills::Base
-
#conditions Legion::LLM::Router::Rule
-
confidence_defaults Legion::LLM::Settings
-
#confidence_score Legion::LLM::Inference::Executor
-
configuration_requirements RubyLLM::Providers::Bedrock
-
configure_anthropic Legion::LLM::Call::Providers
-
configure_azure Legion::LLM::Call::Providers
-
configure_bedrock Legion::LLM::Call::Providers
-
configure_gemini Legion::LLM::Call::Providers
-
configure_ollama Legion::LLM::Call::Providers
-
configure_openai Legion::LLM::Call::Providers
-
configure_providers Legion::LLM::Call::Providers
-
const_missing Legion::LLM::Audit
-
const_missing Legion::LLM::Fleet
-
const_missing Legion::LLM
-
const_missing Legion::LLM::Metering
-
#constraint Legion::LLM::Router::Rule
-
content Legion::LLM::Skills::Base
-
#content Legion::LLM::Call::NativeResponseAdapter
-
#content Legion::LLM::Types::Message
-
#content Legion::LLM::Patches::ToolResultWrapper
-
#content? Legion::LLM::Types::Chunk
-
#content_block_index Legion::LLM::Types::Chunk
-
content_delta Legion::LLM::Types::Chunk
-
context_curation_defaults Legion::LLM::Settings
-
#context_strategy Legion::LLM::Inference::Request
-
conversation_context Legion::LLM::Hooks::Reflection
-
conversation_defaults Legion::LLM::Settings
-
conversation_exists? Legion::LLM::Inference::Conversation
-
#conversation_id Legion::LLM::Types::Chunk
-
#conversation_id Legion::LLM::Types::Message
-
#conversation_id Legion::LLM::Inference::Request
-
#conversation_id Legion::LLM::Inference::Response
-
#correlation_id Legion::LLM::Transport::Message
-
#cost Legion::LLM::Inference::Response
-
#cost_multiplier Legion::LLM::Router::Rule
-
cost_table Legion::LLM::Router::Arbitrage
-
create_conversation Legion::LLM::Inference::Conversation
-
credential_available_for? Legion::LLM::Call::Providers
-
#curate_turn Legion::LLM::Context::Curator
-
#curated_messages Legion::LLM::Context::Curator
-
daemon_ask Legion::LLM::Inference
-
daemon_defaults Legion::LLM::Settings
-
daemon_url Legion::LLM::Call::DaemonClient
-
#data Legion::LLM::Types::ContentBlock
-
debate_defaults Legion::LLM::Settings
-
#debate_enabled? Legion::LLM::Inference::Steps::Debate
-
decide Legion::LLM::Inference::Prompt
-
#dedup_similar Legion::LLM::Context::Curator
-
deduplicate_messages Legion::LLM::Context::Compressor
-
default Legion::LLM::Settings
-
default_model Legion::LLM::Call::Embeddings
-
#default_type Legion::LLM::Transport::Exchanges::Audit
-
#default_type Legion::LLM::Transport::Exchanges::Fleet
-
#default_type Legion::LLM::Transport::Exchanges::Metering
-
#default_type Legion::LLM::Transport::Exchanges::Escalation
-
#delta Legion::LLM::Types::Chunk
-
#deprecation Legion::LLM::Inference::Response
-
deregister Legion::LLM::Fleet::ReplyDispatcher
-
derive Legion::LLM::Inference::Profile
-
description Legion::LLM::Skills::Base
-
#description Legion::LLM::Tools::Adapter
-
#detail Legion::LLM::Types::ContentBlock
-
detect_embedding_capability Legion::LLM::Discovery
-
direct_chat_session? Legion::LLM::Inference
-
discover Legion::LLM::Skills::ExternalDiscovery
-
#discovered_tools Legion::LLM::Inference::Executor
-
discovery_defaults Legion::LLM::Settings
-
dispatch Legion::LLM::Fleet::Dispatcher
-
dispatch Legion::LLM::Inference::Prompt
-
dispatch Legion::LLM::Tools::Dispatcher
-
dispatch_builtin Legion::LLM::Tools::Dispatcher
-
dispatch_chat Legion::LLM::Inference
-
#dispatch_chat Legion::LLM::Call::Dispatch
-
#dispatch_count_tokens Legion::LLM::Call::Dispatch
-
#dispatch_embed Legion::LLM::Call::Dispatch
-
dispatch_extension Legion::LLM::Tools::Dispatcher
-
dispatch_mcp Legion::LLM::Tools::Dispatcher
-
#dispatch_stream Legion::LLM::Call::Dispatch
-
#distill_tool_result Legion::LLM::Context::Curator
-
done Legion::LLM::Types::Chunk
-
#done? Legion::LLM::Types::Chunk
-
#duration_ms Legion::LLM::Types::ToolCall
-
#each Legion::LLM::Router::EscalationChain
-
effective_tier_is_external? Legion::LLM::Inference
-
elapsed_ms_since Legion::LLM::Inference
-
embed Legion::LLM
-
embed Legion::LLM::Inference::GaiaCaller
-
embed_batch Legion::LLM
-
embed_direct Legion::LLM
-
embedding_defaults Legion::LLM::Settings
-
embedding_fallback_chain Legion::LLM
-
embedding_fallback_chain Legion::LLM::Discovery
-
embedding_model Legion::LLM
-
embedding_model Legion::LLM::Discovery
-
embedding_provider Legion::LLM
-
embedding_provider Legion::LLM::Discovery
-
emit Legion::LLM::Metering
-
emit_prompt Legion::LLM::Audit
-
emit_skill Legion::LLM::Audit
-
emit_tools Legion::LLM::Audit
-
#empty? Legion::LLM::Router::EscalationChain
-
enabled? Legion::LLM::Cache
-
enabled? Legion::LLM::Scheduling
-
enabled? Legion::LLM::Router::Arbitrage
-
enabled? Legion::LLM::Scheduling::Batch
-
enabled? Legion::LLM::Quality::ShadowEval
-
#encrypt? Legion::LLM::Transport::Messages::ToolEvent
-
#encrypt? Legion::LLM::Transport::Messages::FleetError
-
#encrypt? Legion::LLM::Transport::Messages::SkillEvent
-
#encrypt? Legion::LLM::Transport::Messages::PromptEvent
-
#encrypt? Legion::LLM::Transport::Messages::MeteringEvent
-
#end_index Legion::LLM::Types::ContentBlock
-
enforcing? Legion::LLM::Hooks::BudgetGuard
-
enqueue Legion::LLM::Scheduling::Batch
-
#enrichments Legion::LLM::Inference::Request
-
#enrichments Legion::LLM::Inference::Executor
-
#enrichments Legion::LLM::Inference::Response
-
ensure_consumer Legion::LLM::Fleet::ReplyDispatcher
-
enterprise_privacy? Legion::LLM::Inference
-
env_present? Legion::LLM::Call::Providers
-
error Legion::LLM::Cache::Response
-
#error Legion::LLM::Types::ToolCall
-
#error? Legion::LLM::Types::ToolCall
-
error_result Legion::LLM::Fleet::Dispatcher
-
#escalated? Legion::LLM::EscalationHistory
-
#escalation_chain Legion::LLM::Inference::Executor
-
#escalation_chain Legion::LLM::EscalationHistory
-
escalation_enabled? Legion::LLM::Inference
-
#escalation_history Legion::LLM::EscalationHistory
-
escalation_quality_threshold Legion::LLM::Inference
-
escalation_rate Legion::LLM::EscalationTracker
-
estimate Legion::LLM::Metering::Pricing
-
estimate_tokens Legion::LLM::Context::Compressor
-
estimated_cost Legion::LLM::Router::Arbitrage
-
evaluate Legion::LLM::Quality::ShadowEval
-
#events Legion::LLM::Inference::Timeline
-
#evict_superseded Legion::LLM::Context::Curator
-
#exchange Legion::LLM::Transport::Messages::ToolEvent
-
#exchange Legion::LLM::Transport::Messages::AuditEvent
-
#exchange Legion::LLM::Transport::Messages::SkillEvent
-
#exchange Legion::LLM::Transport::Messages::PromptEvent
-
#exchange Legion::LLM::Transport::Messages::FleetRequest
-
#exchange Legion::LLM::Transport::Messages::MeteringEvent
-
#exchange_id Legion::LLM::Types::Chunk
-
#exchange_id Legion::LLM::Types::ToolCall
-
exchange_id Legion::LLM::Inference::Tracing
-
#exchange_name Legion::LLM::Transport::Exchanges::Audit
-
#exchange_name Legion::LLM::Transport::Exchanges::Fleet
-
#exchange_name Legion::LLM::Transport::Exchanges::Metering
-
#exchange_name Legion::LLM::Transport::Exchanges::Escalation
-
#execute Legion::LLM::Tools::Adapter
-
execute_chat_request Legion::LLM::Fleet::Handler
-
#expiration Legion::LLM::Transport::Messages::ToolEvent
-
#expiration Legion::LLM::Transport::Messages::FleetError
-
#expiration Legion::LLM::Transport::Messages::SkillEvent
-
#expiration Legion::LLM::Transport::Messages::PromptEvent
-
#expiration Legion::LLM::Transport::Messages::FleetRequest
-
#expiration Legion::LLM::Transport::Messages::FleetResponse
-
#expiration Legion::LLM::Transport::Messages::MeteringEvent
-
#external? Legion::LLM::Router::Resolution
-
external_tier? Legion::LLM::Inference
-
#extra Legion::LLM::Inference::Request
-
extract Legion::LLM::Hooks::Reflection
-
extract Legion::LLM::Inference::Prompt
-
extract_async Legion::LLM::Hooks::Reflection
-
extract_content Legion::LLM::Hooks::Reflection
-
extract_content Legion::LLM::API::Translators::OpenAIRequest
-
extract_decisions Legion::LLM::Hooks::Reflection
-
extract_error Legion::LLM::Fleet::Handler
-
extract_facts Legion::LLM::Hooks::Reflection
-
extract_field Legion::LLM::Fleet::Handler
-
extract_identity Legion::LLM::Inference::AuditPublisher
-
extract_messages_and_system Legion::LLM::API::Translators::OpenAIRequest
-
extract_metering_data Legion::LLM::Hooks::Metering
-
extract_model Legion::LLM::Metering
-
extract_model Legion::LLM::Hooks::CostTracking
-
extract_model_id Legion::LLM::Hooks::Metering
-
extract_patterns Legion::LLM::Hooks::Reflection
-
extract_provider Legion::LLM::Metering
-
extract_provider Legion::LLM::Hooks::Metering
-
extract_provider Legion::LLM::Hooks::CostTracking
-
extract_success Legion::LLM::Fleet::Handler
-
extract_terminal_content Legion::LLM::Fleet::Handler
-
extract_token Legion::LLM::Fleet::Handler
-
extract_token_count Legion::LLM::API::Translators::OpenAIResponse
-
extract_usage Legion::LLM::Metering
-
extract_usage Legion::LLM::Hooks::Metering
-
extract_usage Legion::LLM::Hooks::CostTracking
-
fail_request Legion::LLM::Cache::Response
-
#failures Legion::LLM::Quality::Checker::QualityResult
-
#fallback Legion::LLM::Router::Rule
-
#features Legion::LLM::Inference::Response
-
file_change_trigger_patterns Legion::LLM::Skills::Base
-
file_change_triggers Legion::LLM::Skills::Base
-
#file_id Legion::LLM::Types::ContentBlock
-
file_trigger_skills Legion::LLM::Skills::Registry
-
#final_resolution Legion::LLM::EscalationHistory
-
find Legion::LLM::Skills::Registry
-
find_role_mapping Legion::LLM::Inference::Steps::TierAssigner
-
#finished_at Legion::LLM::Types::ToolCall
-
first_present Legion::LLM::Call::ClaudeConfigLoader
-
#fleet? Legion::LLM::Router::Resolution
-
fleet_available? Legion::LLM::Fleet::Dispatcher
-
fleet_enabled? Legion::LLM::Fleet::Dispatcher
-
flush Legion::LLM::Scheduling::Batch
-
flush_spool Legion::LLM::Metering
-
flush_spool Legion::LLM::Inference::Steps::Metering
-
#fold_resolved_exchanges Legion::LLM::Context::Curator
-
follows Legion::LLM::Skills::Base
-
follows_skill Legion::LLM::Skills::Base
-
for Legion::LLM::Call::Registry
-
#fork Legion::LLM::Inference::Request
-
format Legion::LLM::API::Translators::AnthropicResponse
-
format_chat_completion Legion::LLM::API::Translators::OpenAIResponse
-
format_chunk Legion::LLM::API::Translators::AnthropicResponse
-
format_embeddings Legion::LLM::API::Translators::OpenAIResponse
-
format_model_object Legion::LLM::API::Translators::OpenAIResponse
-
format_stream_chunk Legion::LLM::API::Translators::OpenAIResponse
-
from_chat_args Legion::LLM::Inference::Request
-
from_hash Legion::LLM::Router::Rule
-
from_hash Legion::LLM::Types::Message
-
from_hash Legion::LLM::Types::ToolCall
-
from_hash Legion::LLM::Types::ContentBlock
-
from_ruby_llm Legion::LLM::Inference::Response
-
#frontier? Legion::LLM::Router::Resolution
-
fulfill_nack Legion::LLM::Fleet::ReplyDispatcher
-
fulfill_return Legion::LLM::Fleet::ReplyDispatcher
-
fuzzy_match Legion::LLM::Metering::Pricing
-
gaia_caller Legion::LLM::Inference::GaiaCaller
-
#gaia_debate_trigger? Legion::LLM::Inference::Steps::Debate
-
gaia_tracing Legion::LLM::Inference::GaiaCaller
-
#gate Legion::LLM::Skills::StepResult
-
#gate Legion::LLM::Skills::SkillRunResult
-
#gated Legion::LLM::Skills::SkillRunResult
-
gateway_defaults Legion::LLM::Settings
-
gateway_enabled? Legion::LLM::Router::GatewayInterceptor
-
gateway_headers Legion::LLM::Router::GatewayInterceptor
-
gateway_metering? Legion::LLM::Hooks::Metering
-
gateway_settings Legion::LLM::Router::GatewayInterceptor
-
generate Legion::LLM::Call::Embeddings
-
generate Legion::LLM::Call::StructuredOutput
-
generate_batch Legion::LLM::Call::Embeddings
-
#generation Legion::LLM::Inference::Request
-
get Legion::LLM::Cache
-
guard_response Legion::LLM::Hooks::ResponseGuard
-
handle_delivery Legion::LLM::Fleet::ReplyDispatcher
-
handle_fleet_request Legion::LLM::Fleet::Handler
-
#handle_tool_calls Legion::LLM::Patches::RubyLLMParallelTools
-
hash_response_details Legion::LLM::Inference
-
hash_token Legion::LLM::Fleet::Handler
-
#headers Legion::LLM::Transport::Message
-
#headers Legion::LLM::Transport::Messages::ToolEvent
-
#headers Legion::LLM::Transport::Messages::FleetError
-
#headers Legion::LLM::Transport::Messages::SkillEvent
-
#headers Legion::LLM::Transport::Messages::PromptEvent
-
#headers Legion::LLM::Transport::Messages::FleetResponse
-
#headers Legion::LLM::Transport::Messages::MeteringEvent
-
health_tracker Legion::LLM::Router
-
history Legion::LLM::Quality::ShadowEval
-
history Legion::LLM::EscalationTracker
-
#hooks Legion::LLM::Inference::Request
-
http_get Legion::LLM::Call::DaemonClient
-
http_post Legion::LLM::Call::DaemonClient
-
hydrate_from_apollo Legion::LLM::Tools::Confidence
-
hydrate_from_l2 Legion::LLM::Tools::Confidence
-
#id Legion::LLM::Types::Message
-
#id Legion::LLM::Types::ToolCall
-
#id Legion::LLM::Inference::Request
-
#id Legion::LLM::Inference::Response
-
#id Legion::LLM::Types::ContentBlock
-
#id Legion::LLM::Patches::ToolResultWrapper
-
#idempotency_key Legion::LLM::Inference::Request
-
identity_fields Legion::LLM::Inference::Steps::Metering
-
image Legion::LLM::Types::ContentBlock
-
in_memory? Legion::LLM::Inference::Conversation
-
#index Legion::LLM::Types::Chunk
-
inference Legion::LLM::Call::DaemonClient
-
inference_input_payload Legion::LLM::Inference
-
inference_response_details Legion::LLM::Inference
-
inference_text_length Legion::LLM::Inference
-
inference_token_value Legion::LLM::Inference
-
init Legion::LLM::Inference::Tracing
-
init_request Legion::LLM::Cache::Response
-
#initialize Legion::LLM::RateLimitError
-
#initialize Legion::LLM::PipelineError
-
#initialize Legion::LLM::Router::Rule
-
#initialize Legion::LLM::Call::NativeResponseAdapter
-
#initialize Legion::LLM::Skills::InvalidSkill
-
#initialize Legion::LLM::Skills::StepError
-
#initialize Legion::LLM::Tools::Adapter
-
#initialize Legion::LLM::Usage
-
#initialize Legion::LLM::Context::Curator
-
#initialize Legion::LLM::Router::Resolution
-
#initialize Legion::LLM::Inference::Executor
-
#initialize Legion::LLM::Inference::Timeline
-
#initialize Legion::LLM::Router::HealthTracker
-
#initialize Legion::LLM::Router::EscalationChain
-
#inject Legion::LLM::Skills::StepResult
-
#inject Legion::LLM::Skills::SkillRunResult
-
inject Legion::LLM::Inference::EnrichmentInjector
-
inject_anthropic_cache_control! Legion::LLM::Call::Providers
-
#input Legion::LLM::Types::ContentBlock
-
#input_tokens Legion::LLM::Call::NativeResponseAdapter
-
#input_tokens Legion::LLM::Types::Message
-
#input_tokens Legion::LLM::Usage
-
install Legion::LLM::Hooks::Metering
-
install Legion::LLM::Hooks::Reflection
-
install Legion::LLM::Hooks::Reciprocity
-
install Legion::LLM::Hooks::BudgetGuard
-
install Legion::LLM::Hooks::CostTracking
-
install_defaults Legion::LLM::Hooks
-
install_hook Legion::LLM::Metering
-
intercept Legion::LLM::Tools::Interceptor
-
intercept Legion::LLM::Router::GatewayInterceptor
-
interpret_response Legion::LLM::Call::DaemonClient
-
#is_a? Legion::LLM::Patches::ToolResultWrapper
-
#is_error Legion::LLM::Types::ContentBlock
-
key Legion::LLM::Cache
-
#llm_ask Legion::LLM::Helper
-
llm_available_for? Legion::LLM::Fleet::Handler
-
#llm_budget_remaining Legion::LLM::Helper
-
#llm_can_embed? Legion::LLM::Helper
-
#llm_chat Legion::LLM::Helper
-
#llm_connected? Legion::LLM::Helper
-
#llm_cost_estimate Legion::LLM::Helper
-
#llm_cost_summary Legion::LLM::Helper
-
#llm_default_intent Legion::LLM::Helper
-
#llm_default_model Legion::LLM::Helper
-
#llm_default_provider Legion::LLM::Helper
-
#llm_distill_tool_result Legion::LLM::Context::Curator
-
#llm_embed Legion::LLM::Helper
-
#llm_embed_batch Legion::LLM::Helper
-
#llm_routing_enabled? Legion::LLM::Helper
-
#llm_session Legion::LLM::Helper
-
#llm_structured Legion::LLM::Helper
-
load Legion::LLM::Call::CodexConfigLoader
-
load Legion::LLM::Call::ClaudeConfigLoader
-
load_all Legion::LLM::Transport
-
load_defaults Legion::LLM::Tools::Interceptor
-
load_directory Legion::LLM::Skills::DiskLoader
-
load_from_directories Legion::LLM::Skills::DiskLoader
-
load_md_skill Legion::LLM::Skills::DiskLoader
-
load_transport Legion::LLM::Audit
-
load_transport Legion::LLM::Fleet
-
load_transport Legion::LLM::Metering
-
#local? Legion::LLM::Router::Resolution
-
log_inference_error Legion::LLM::Inference
-
log_inference_request Legion::LLM::Inference
-
log_inference_response Legion::LLM::Inference
-
lookup Legion::LLM::Tools::Confidence
-
map_finish_reason Legion::LLM::API::Translators::OpenAIResponse
-
mark_unhealthy Legion::LLM::Call::DaemonClient
-
match? Legion::LLM::Tools::Interceptors::PythonVenv
-
#matches_intent? Legion::LLM::Router::Rule
-
#max_attempts Legion::LLM::Router::EscalationChain
-
maybe_shadow_evaluate Legion::LLM::Inference
-
#media_type Legion::LLM::Types::ContentBlock
-
memory_pressure? Legion::LLM::Discovery::System
-
merged_config Legion::LLM::Call::ClaudeConfigLoader
-
#message Legion::LLM::Transport::Message
-
#message Legion::LLM::Inference::Response
-
#message Legion::LLM::Types::ContentBlock
-
message_content Legion::LLM::Fleet::Handler
-
#message_context Legion::LLM::Transport::Message
-
#message_id Legion::LLM::Transport::Message
-
#messages Legion::LLM::Inference::Request
-
messages Legion::LLM::Inference::Conversation
-
meta Legion::LLM::Cache::Response
-
#metadata Legion::LLM::Inference::Request
-
#metadata Legion::LLM::Router::Resolution
-
#metadata Legion::LLM::Skills::StepResult
-
metering_available? Legion::LLM::Hooks::Metering
-
migrate_parent_links! Legion::LLM::Inference::Conversation
-
#modality Legion::LLM::Inference::Request
-
#model Legion::LLM::Types::Message
-
#model Legion::LLM::Router::Resolution
-
model_allowed? Legion::LLM::Router::GatewayInterceptor
-
model_available? Legion::LLM::Discovery::Ollama
-
model_names Legion::LLM::Discovery::Ollama
-
model_size Legion::LLM::Discovery::Ollama
-
models Legion::LLM::Discovery::Ollama
-
models_from_discovery Legion::LLM::API::OpenAI::Models
-
models_from_providers Legion::LLM::API::OpenAI::Models
-
#name Legion::LLM::Router::Rule
-
#name Legion::LLM::Tools::Adapter
-
#name Legion::LLM::Types::Message
-
#name Legion::LLM::Types::ToolCall
-
#name Legion::LLM::Types::ContentBlock
-
namespace Legion::LLM::Skills::Base
-
next_off_peak Legion::LLM::Scheduling
-
next_off_peak Legion::LLM::Scheduling::OffPeak
-
normalize Legion::LLM::API::Translators::OpenAIRequest
-
normalize Legion::LLM::API::Translators::AnthropicRequest
-
normalize_ask_direct_hash Legion::LLM::Inference
-
normalize_error Legion::LLM::Fleet::ReplyDispatcher
-
normalize_messages Legion::LLM::Fleet::Handler
-
normalize_secret Legion::LLM::Call::ClaudeConfigLoader
-
normalize_tools Legion::LLM::API::Translators::OpenAIRequest
-
#note Legion::LLM::Router::Rule
-
oauth_account_available? Legion::LLM::Call::ClaudeConfigLoader
-
object_response_details Legion::LLM::Inference
-
ollama_running? Legion::LLM::Call::Providers
-
openai_api_key Legion::LLM::Call::ClaudeConfigLoader
-
#openai_compat? Legion::LLM::Router::Resolution
-
#output_tokens Legion::LLM::Call::NativeResponseAdapter
-
#output_tokens Legion::LLM::Types::Message
-
#output_tokens Legion::LLM::Usage
-
#params_schema Legion::LLM::Tools::Adapter
-
#parent_id Legion::LLM::Types::Message
-
parse_frontmatter Legion::LLM::Skills::DiskLoader
-
parse_payload Legion::LLM::Fleet::ReplyDispatcher
-
#participants Legion::LLM::Inference::Response
-
#participants Legion::LLM::Inference::Timeline
-
#passed Legion::LLM::Quality::Checker::QualityResult
-
peak_hour? Legion::LLM::Scheduling::OffPeak
-
peak_hours? Legion::LLM::Scheduling
-
pending_count Legion::LLM::Fleet::ReplyDispatcher
-
pipeline_enabled? Legion::LLM::Inference
-
pipeline_response_details Legion::LLM::Inference
-
platform Legion::LLM::Discovery::System
-
poll Legion::LLM::Cache::Response
-
#predictions Legion::LLM::Inference::Request
-
#predictions Legion::LLM::Inference::Response
-
pricing_for Legion::LLM::Metering::Recorder
-
#primary Legion::LLM::Router::EscalationChain
-
#priority Legion::LLM::Router::Rule
-
#priority Legion::LLM::Inference::Request
-
#priority Legion::LLM::Transport::Messages::ToolEvent
-
#priority Legion::LLM::Transport::Messages::FleetError
-
#priority Legion::LLM::Transport::Messages::SkillEvent
-
#priority Legion::LLM::Transport::Messages::PromptEvent
-
#priority Legion::LLM::Transport::Messages::FleetRequest
-
#priority Legion::LLM::Transport::Messages::FleetResponse
-
#priority Legion::LLM::Transport::Messages::MeteringEvent
-
probe_provider_credentials Legion::LLM::Call::Providers
-
probe_via_chat Legion::LLM::Call::Providers
-
probe_via_model_list Legion::LLM::Call::Providers
-
#profile Legion::LLM::Inference::Executor
-
prompt_caching_defaults Legion::LLM::Settings
-
#provider Legion::LLM::Types::Message
-
#provider Legion::LLM::Router::Resolution
-
provider_layer_defaults Legion::LLM::Settings
-
providers Legion::LLM::Settings
-
publish Legion::LLM::Inference::AuditPublisher
-
#publish Legion::LLM::Transport::Messages::FleetError
-
#publish Legion::LLM::Transport::Messages::FleetResponse
-
publish_entry Legion::LLM::Hooks::Reflection
-
publish_escalation_event Legion::LLM::Inference
-
publish_event Legion::LLM::Inference::Steps::Metering
-
publish_metering Legion::LLM::Hooks::Metering
-
publish_or_spool Legion::LLM::Inference::Steps::Metering
-
publish_reply Legion::LLM::Fleet::Handler
-
publish_request Legion::LLM::Fleet::Dispatcher
-
#quality Legion::LLM::Inference::Response
-
queue_size Legion::LLM::Scheduling::Batch
-
rag_defaults Legion::LLM::Settings
-
#rate_limit Legion::LLM::Inference::Response
-
read_config Legion::LLM::Call::CodexConfigLoader
-
read_json Legion::LLM::Call::CodexConfigLoader
-
read_json Legion::LLM::Call::ClaudeConfigLoader
-
read_metadata Legion::LLM::Inference::Conversation
-
read_openai_api_key Legion::LLM::Call::CodexConfigLoader
-
read_openai_credential Legion::LLM::Call::CodexConfigLoader
-
read_sticky_state Legion::LLM::Inference::Conversation
-
read_token Legion::LLM::Call::CodexConfigLoader
-
record Legion::LLM::Hooks::Metering
-
record Legion::LLM::Metering::Tokens
-
record Legion::LLM::Metering::Recorder
-
record Legion::LLM::Tools::Confidence
-
#record Legion::LLM::Inference::Timeline
-
record Legion::LLM::EscalationTracker
-
#record_escalation_attempt Legion::LLM::EscalationHistory
-
record_failure Legion::LLM::Tools::Confidence
-
record_reciprocity Legion::LLM::Hooks::Reciprocity
-
record_success Legion::LLM::Tools::Confidence
-
recover_openai_with_codex Legion::LLM::Call::Providers
-
recover_with_alternative_credentials Legion::LLM::Call::Providers
-
refresh! Legion::LLM::Discovery::Ollama
-
refresh! Legion::LLM::Discovery::System
-
register Legion::LLM::Call::Registry
-
register Legion::LLM::Skills::Registry
-
register Legion::LLM::Tools::Interceptor
-
register Legion::LLM::Fleet::ReplyDispatcher
-
register! Legion::LLM::Tools::Interceptors::PythonVenv
-
#register_handler Legion::LLM::Router::HealthTracker
-
register_routes Legion::LLM::API
-
registered Legion::LLM::API
-
registered Legion::LLM::API::Auth
-
registered Legion::LLM::API::Native::Chat
-
registered Legion::LLM::API::OpenAI::Models
-
registered Legion::LLM::Tools::Interceptor
-
registered Legion::LLM::API::Native::Helpers
-
registered Legion::LLM::API::Native::Inference
-
registered Legion::LLM::API::Native::Providers
-
registered Legion::LLM::API::OpenAI::Embeddings
-
registered Legion::LLM::API::Anthropic::Messages
-
registered Legion::LLM::API::OpenAI::ChatCompletions
-
registered? Legion::LLM::Call::Registry
-
remaining Legion::LLM::Hooks::BudgetGuard
-
replace Legion::LLM::Inference::Conversation
-
#reply_to Legion::LLM::Transport::Messages::FleetRequest
-
#report Legion::LLM::Router::HealthTracker
-
report_health Legion::LLM::Inference
-
request Legion::LLM::Inference::Prompt
-
#request Legion::LLM::Inference::Executor
-
#request_id Legion::LLM::Types::Chunk
-
#request_id Legion::LLM::Inference::Response
-
require_auth? Legion::LLM::Fleet::Handler
-
#reset Legion::LLM::Router::HealthTracker
-
reset! Legion::LLM::Hooks
-
reset! Legion::LLM::Router
-
reset! Legion::LLM::Discovery
-
reset! Legion::LLM::Call::Registry
-
reset! Legion::LLM::Metering::Tokens
-
reset! Legion::LLM::Skills::Registry
-
reset! Legion::LLM::Discovery::Ollama
-
reset! Legion::LLM::Discovery::System
-
reset! Legion::LLM::Hooks::Reflection
-
reset! Legion::LLM::Scheduling::Batch
-
reset! Legion::LLM::Tools::Confidence
-
reset! Legion::LLM::Tools::Interceptor
-
reset! Legion::LLM::Call::DaemonClient
-
reset! Legion::LLM::Fleet::ReplyDispatcher
-
reset! Legion::LLM::Inference::Conversation
-
#reset_all Legion::LLM::Router::HealthTracker
-
resolve Legion::LLM::Router
-
resolve_ask_direct_response Legion::LLM::Inference
-
resolve_baseline Legion::LLM::Inference::EnrichmentInjector
-
resolve_broker_aws_credentials Legion::LLM::Call::Providers
-
resolve_broker_credential Legion::LLM::Call::Providers
-
resolve_chain Legion::LLM::Router
-
resolve_env_api_key Legion::LLM::Call::CodexConfigLoader
-
resolve_llm_secrets Legion::LLM::Call::Providers
-
resolve_price Legion::LLM::Metering::Pricing
-
resolve_setting_reference Legion::LLM::Call::ClaudeConfigLoader
-
resolve_timeout Legion::LLM::Fleet::Dispatcher
-
response Legion::LLM::Cache::Response
-
#response_format Legion::LLM::Inference::Request
-
response_guards_enabled? Legion::LLM::Inference
-
#result Legion::LLM::Types::ToolCall
-
#result Legion::LLM::Patches::ToolResultWrapper
-
#resume_at Legion::LLM::Skills::SkillRunResult
-
#retry Legion::LLM::Inference::Response
-
#retry_after Legion::LLM::RateLimitError
-
#retryable? Legion::LLM::LLMError
-
#retryable? Legion::LLM::RateLimitError
-
#retryable? Legion::LLM::ContextOverflow
-
#retryable? Legion::LLM::ProviderError
-
rewrite Legion::LLM::Tools::Interceptors::PythonVenv
-
rewrite_command Legion::LLM::Tools::Interceptors::PythonVenv
-
#role Legion::LLM::Types::Message
-
#routing Legion::LLM::Inference::Request
-
#routing Legion::LLM::Inference::Response
-
routing_defaults Legion::LLM::Settings
-
routing_enabled? Legion::LLM::Router
-
#routing_key Legion::LLM::Transport::Messages::ToolEvent
-
#routing_key Legion::LLM::Transport::Messages::AuditEvent
-
#routing_key Legion::LLM::Transport::Messages::FleetError
-
#routing_key Legion::LLM::Transport::Messages::SkillEvent
-
#routing_key Legion::LLM::Transport::Messages::PromptEvent
-
#routing_key Legion::LLM::Transport::Messages::FleetRequest
-
#routing_key Legion::LLM::Transport::Messages::FleetResponse
-
#routing_key Legion::LLM::Transport::Messages::MeteringEvent
-
#routing_key Legion::LLM::Transport::Messages::EscalationEvent
-
#rule Legion::LLM::Router::Resolution
-
run Legion::LLM::Discovery
-
#run Legion::LLM::Skills::Base
-
run_after Legion::LLM::Hooks
-
run_before Legion::LLM::Hooks
-
#run_debate Legion::LLM::Inference::Steps::Debate
-
run_shadow Legion::LLM::Tools::Dispatcher
-
safe_inference_value Legion::LLM::Inference
-
#safety Legion::LLM::Inference::Response
-
sanitize_model Legion::LLM::Fleet::Dispatcher
-
#schedule Legion::LLM::Router::Rule
-
scheduling_defaults Legion::LLM::Settings
-
#schema_version Legion::LLM::Inference::Request
-
#schema_version Legion::LLM::Inference::Response
-
#score Legion::LLM::Quality::Confidence::Score
-
score Legion::LLM::Quality::Confidence::Scorer
-
#seq Legion::LLM::Types::Message
-
serialize_tokens Legion::LLM::Inference::AuditPublisher
-
session_budget Legion::LLM::Hooks::BudgetGuard
-
session_exceeded? Legion::LLM::Metering::Tokens
-
session_warning? Legion::LLM::Metering::Tokens
-
set Legion::LLM::Cache
-
set_defaults Legion::LLM::Config
-
set_skill_state Legion::LLM::Inference::Conversation
-
setting_has_usable_credential? Legion::LLM::Call::ClaudeConfigLoader
-
settings Legion::LLM
-
settings_directories Legion::LLM::Skills
-
setup Legion::LLM::Call::Providers
-
should_defer? Legion::LLM::Scheduling
-
should_defer? Legion::LLM::Scheduling::OffPeak
-
should_extract? Legion::LLM::Hooks::Reflection
-
should_override? Legion::LLM::Tools::Confidence
-
should_sample? Legion::LLM::Quality::ShadowEval
-
should_shadow? Legion::LLM::Tools::Confidence
-
shutdown Legion::LLM
-
sidechain_messages Legion::LLM::Inference::Conversation
-
#signals Legion::LLM::Quality::Confidence::Score
-
#size Legion::LLM::Router::EscalationChain
-
skill_cancelled? Legion::LLM::Inference::Conversation
-
skill_name Legion::LLM::Skills::Base
-
skill_state Legion::LLM::Inference::Conversation
-
skills_defaults Legion::LLM::Settings
-
skip? Legion::LLM::Inference::Profile
-
social_runner Legion::LLM::Hooks::Reciprocity
-
#sort_tools_deterministically Legion::LLM::Inference::Steps::PromptCache
-
#source Legion::LLM::Types::ToolCall
-
#source Legion::LLM::Types::ContentBlock
-
#source Legion::LLM::Quality::Confidence::Score
-
#source_type Legion::LLM::Types::ContentBlock
-
spool_available? Legion::LLM::Metering
-
spool_event Legion::LLM::Metering
-
stale? Legion::LLM::Discovery::Ollama
-
stale? Legion::LLM::Discovery::System
-
start Legion::LLM
-
start Legion::LLM::Skills
-
#start_index Legion::LLM::Types::ContentBlock
-
started? Legion::LLM
-
#started_at Legion::LLM::Types::ToolCall
-
status Legion::LLM::Scheduling
-
#status Legion::LLM::Types::Message
-
status Legion::LLM::Cache::Response
-
#status Legion::LLM::Types::ToolCall
-
status Legion::LLM::Scheduling::Batch
-
status Legion::LLM::Hooks::BudgetGuard
-
#step Legion::LLM::PipelineError
-
#step_billing Legion::LLM::Inference::Steps::Billing
-
#step_classification Legion::LLM::Inference::Steps::Classification
-
#step_confidence_scoring Legion::LLM::Inference::Steps::ConfidenceScoring
-
#step_debate Legion::LLM::Inference::Steps::Debate
-
#step_gaia_advisory Legion::LLM::Inference::Steps::GaiaAdvisory
-
#step_knowledge_capture Legion::LLM::Inference::Steps::KnowledgeCapture
-
#step_mcp_discovery Legion::LLM::Inference::Steps::McpDiscovery
-
#step_post_response Legion::LLM::Inference::Steps::PostResponse
-
#step_rag_context Legion::LLM::Inference::Steps::RagContext
-
#step_rbac Legion::LLM::Inference::Steps::Rbac
-
#step_skill_injector Legion::LLM::Inference::Steps::SkillInjector
-
#step_sticky_persist Legion::LLM::Inference::Steps::StickyPersist
-
#step_sticky_runners Legion::LLM::Inference::Steps::StickyRunners
-
#step_token_budget Legion::LLM::Inference::Steps::TokenBudget
-
#step_tool_calls Legion::LLM::Inference::Steps::ToolCalls
-
#step_tool_discovery Legion::LLM::Inference::Steps::ToolDiscovery
-
#step_tool_history_inject Legion::LLM::Inference::Steps::ToolHistory
-
#step_trigger_match Legion::LLM::Inference::Steps::TriggerMatch
-
steps Legion::LLM::Skills::Base
-
#stop Legion::LLM::Inference::Request
-
#stop Legion::LLM::Inference::Response
-
stop_flush_timer Legion::LLM::Scheduling::Batch
-
#stop_reason Legion::LLM::Types::Chunk
-
stopwords_for_level Legion::LLM::Context::Compressor
-
store_metadata Legion::LLM::Inference::Conversation
-
#stream Legion::LLM::Inference::Request
-
#stream Legion::LLM::Inference::Response
-
streaming_events Legion::LLM::API::Translators::AnthropicResponse
-
#strip_thinking Legion::LLM::Context::Curator
-
structured Legion::LLM
-
structured Legion::LLM::Inference::GaiaCaller
-
structured_direct Legion::LLM
-
#success? Legion::LLM::Types::ToolCall
-
summarize Legion::LLM::Inference::Prompt
-
summarize_messages Legion::LLM::Context::Compressor
-
summary Legion::LLM::Metering::Tokens
-
summary Legion::LLM::Hooks::Reflection
-
summary Legion::LLM::Metering::Recorder
-
summary Legion::LLM::Quality::ShadowEval
-
summary Legion::LLM::EscalationTracker
-
#system Legion::LLM::Inference::Request
-
system_baseline_default Legion::LLM::Settings
-
#target Legion::LLM::Router::Rule
-
#task_id Legion::LLM::Types::Message
-
telemetry_defaults Legion::LLM::Settings
-
#test Legion::LLM::Inference::Request
-
#test Legion::LLM::Inference::Response
-
#text Legion::LLM::Types::Message
-
#text Legion::LLM::Types::ContentBlock
-
text Legion::LLM::Types::ContentBlock
-
#thinking Legion::LLM::Inference::Request
-
#thinking Legion::LLM::Inference::Response
-
thinking Legion::LLM::Types::ContentBlock
-
#tier Legion::LLM::Router::Resolution
-
tier_available? Legion::LLM::Router
-
tier_mappings Legion::LLM::Inference::Steps::TierAssigner
-
#timeline Legion::LLM::Inference::Executor
-
#timeline Legion::LLM::Inference::Response
-
timeout_result Legion::LLM::Fleet::Dispatcher
-
#timestamp Legion::LLM::Types::Chunk
-
#timestamp Legion::LLM::Types::Message
-
#timestamps Legion::LLM::Inference::Response
-
timing_and_context Legion::LLM::Inference::Steps::Metering
-
#to_a Legion::LLM::Router::EscalationChain
-
#to_audit_hash Legion::LLM::Types::ToolCall
-
#to_h Legion::LLM::Types::Chunk
-
#to_h Legion::LLM::Types::Message
-
#to_h Legion::LLM::Types::ToolCall
-
#to_h Legion::LLM::Router::Resolution
-
#to_h Legion::LLM::Types::ContentBlock
-
#to_h Legion::LLM::Quality::Confidence::Score
-
#to_provider_hash Legion::LLM::Types::Message
-
#to_resolution Legion::LLM::Router::Rule
-
token_fields Legion::LLM::Inference::Steps::Metering
-
token_valid? Legion::LLM::Call::CodexConfigLoader
-
#tokens Legion::LLM::Inference::Request
-
#tokens Legion::LLM::Inference::Response
-
#tool_call Legion::LLM::Types::Chunk
-
#tool_call_id Legion::LLM::Types::Message
-
#tool_call_id Legion::LLM::Patches::ToolResultWrapper
-
#tool_calls Legion::LLM::Types::Message
-
#tool_choice Legion::LLM::Inference::Request
-
#tool_event_handler Legion::LLM::Inference::Executor
-
#tool_name Legion::LLM::Patches::ToolResultWrapper
-
tool_result Legion::LLM::Types::ContentBlock
-
tool_trigger_defaults Legion::LLM::Settings
-
tool_use Legion::LLM::Types::ContentBlock
-
#tool_use_id Legion::LLM::Types::ContentBlock
-
#tools Legion::LLM::Inference::Request
-
#tools Legion::LLM::Inference::Response
-
total_input_tokens Legion::LLM::Metering::Tokens
-
total_memory_mb Legion::LLM::Discovery::System
-
total_output_tokens Legion::LLM::Metering::Tokens
-
#total_tokens Legion::LLM::Usage
-
total_tokens Legion::LLM::Metering::Tokens
-
#tracing Legion::LLM::Types::Chunk
-
#tracing Legion::LLM::Inference::Request
-
#tracing Legion::LLM::Inference::Executor
-
#tracing Legion::LLM::Inference::Response
-
#tracing_headers Legion::LLM::Transport::Message
-
track Legion::LLM::Hooks::CostTracking
-
transport_available? Legion::LLM::Fleet::ReplyDispatcher
-
transport_connected? Legion::LLM::Audit
-
transport_connected? Legion::LLM::Metering
-
transport_metering? Legion::LLM::Hooks::Metering
-
transport_ready? Legion::LLM::Fleet::Dispatcher
-
trigger Legion::LLM::Skills::Base
-
trigger_word_index Legion::LLM::Skills::Registry
-
trigger_words Legion::LLM::Skills::Base
-
try_defer Legion::LLM::Inference
-
try_register_native_provider Legion::LLM::Call::Providers
-
#ttl Legion::LLM::Inference::Request
-
#type Legion::LLM::Types::Chunk
-
#type Legion::LLM::Types::ContentBlock
-
#type Legion::LLM::Transport::Messages::ToolEvent
-
#type Legion::LLM::Transport::Messages::FleetError
-
#type Legion::LLM::Transport::Messages::SkillEvent
-
#type Legion::LLM::Transport::Messages::PromptEvent
-
#type Legion::LLM::Transport::Messages::FleetRequest
-
#type Legion::LLM::Transport::Messages::FleetResponse
-
#type Legion::LLM::Transport::Messages::MeteringEvent
-
unavailable_response Legion::LLM::Fleet::Handler
-
usable_setting? Legion::LLM::Call::Providers
-
#usage Legion::LLM::Types::Chunk
-
#usage Legion::LLM::Call::NativeResponseAdapter
-
valid_token? Legion::LLM::Fleet::Handler
-
#validation Legion::LLM::Inference::Response
-
venv_available? Legion::LLM::Tools::Interceptors::PythonVenv
-
verify_providers Legion::LLM::Call::Providers
-
#version Legion::LLM::Types::Message
-
wait_for_response Legion::LLM::Fleet::Dispatcher
-
warn_once Legion::LLM::CompatWarning
-
#warnings Legion::LLM::Inference::Executor
-
#warnings Legion::LLM::Inference::Response
-
when_conditions Legion::LLM::Skills::Base
-
#wire Legion::LLM::Inference::Response
-
#with Legion::LLM::Inference::Response
-
#with_result Legion::LLM::Types::ToolCall
-
#within_schedule? Legion::LLM::Router::Rule
-
wrap Legion::LLM::Types::Message
-
write_sticky_state Legion::LLM::Inference::Conversation