Class: Karafka::Instrumentation::LoggerListener
- Inherits:
-
Object
- Object
- Karafka::Instrumentation::LoggerListener
- Defined in:
- lib/karafka/instrumentation/logger_listener.rb
Overview
Default listener that hooks up to our instrumentation and uses its events for logging It can be removed/replaced or anything without any harm to the Karafka app flow.
Instance Method Summary collapse
-
#initialize(log_polling: true) ⇒ LoggerListener
constructor
A new instance of LoggerListener.
- #on_app_quiet(event) ⇒ Object
- #on_app_quieting(event) ⇒ Object
-
#on_app_running(event) ⇒ Object
Logs info that we’re running Karafka app.
-
#on_app_stopped(event) ⇒ Object
Logs info that we stopped the Karafka server.
-
#on_app_stopping(event) ⇒ Object
Logs info that we’re going to stop the Karafka server.
-
#on_client_pause(event) ⇒ Object
Prints info about a consumer pause occurrence.
-
#on_client_resume(event) ⇒ Object
Prints information about resuming of processing of a given topic partition.
- #on_connection_listener_before_fetch_loop(event) ⇒ Object
-
#on_connection_listener_fetch_loop(event) ⇒ Object
Logs each messages fetching attempt.
-
#on_connection_listener_fetch_loop_received(event) ⇒ Object
Logs about messages that we’ve received from Kafka.
-
#on_consumer_consuming_retry(event) ⇒ Object
Prints info about retry of processing after an error.
-
#on_consumer_consuming_seek(event) ⇒ Object
Prints info about seeking to a particular location.
-
#on_dead_letter_queue_dispatched(event) ⇒ Object
Logs info when we have dispatched a message the the DLQ.
-
#on_error_occurred(event) ⇒ Object
There are many types of errors that can occur in many places, but we provide a single handler for all of them to simplify error instrumentation.
- #on_filtering_seek(event) ⇒ Object
-
#on_filtering_throttled(event) ⇒ Object
Logs info about throttling event.
-
#on_process_notice_signal(event) ⇒ Object
Logs info about system signals that Karafka received and prints backtrace for threads in case of ttin.
-
#on_rebalance_partitions_assigned(event) ⇒ Object
Logs info about partitions that we’ve gained.
-
#on_rebalance_partitions_revoked(event) ⇒ Object
Logs info about partitions we have lost.
- #on_swarm_manager_before_fork(event) ⇒ Object
- #on_swarm_manager_control(event) ⇒ Object
- #on_swarm_manager_stopping(event) ⇒ Object
- #on_swarm_manager_terminating(event) ⇒ Object
- #on_swarm_node_after_fork(_event) ⇒ Object
-
#on_worker_process(event) ⇒ Object
Prints info about the fact that a given job has started.
-
#on_worker_processed(event) ⇒ Object
Prints info about the fact that a given job has finished.
-
#on_worker_scaling_down(event) ⇒ Object
Logs info about the workers pool scaling down.
-
#on_worker_scaling_up(event) ⇒ Object
Logs info about the workers pool scaling up.
Constructor Details
#initialize(log_polling: true) ⇒ LoggerListener
Returns a new instance of LoggerListener.
23 24 25 |
# File 'lib/karafka/instrumentation/logger_listener.rb', line 23 def initialize(log_polling: true) @log_polling = log_polling end |
Instance Method Details
#on_app_quiet(event) ⇒ Object
223 224 225 |
# File 'lib/karafka/instrumentation/logger_listener.rb', line 223 def on_app_quiet(event) info "[#{event[:server_id]}] Reached quiet mode. No messages will be processed anymore" end |
#on_app_quieting(event) ⇒ Object
218 219 220 |
# File 'lib/karafka/instrumentation/logger_listener.rb', line 218 def on_app_quieting(event) info "[#{event[:server_id]}] Switching to quiet mode. New messages will not be processed" end |
#on_app_running(event) ⇒ Object
Logs info that we’re running Karafka app.
206 207 208 209 210 211 212 213 214 215 |
# File 'lib/karafka/instrumentation/logger_listener.rb', line 206 def on_app_running(event) server_id = event[:server_id] info "[#{server_id}] Running in #{RUBY_DESCRIPTION}" info "[#{server_id}] Running Karafka #{Karafka::VERSION} server" return if Karafka.pro? info "[#{server_id}] See LICENSE and the LGPL-3.0 for licensing details" end |
#on_app_stopped(event) ⇒ Object
Logs info that we stopped the Karafka server.
237 238 239 |
# File 'lib/karafka/instrumentation/logger_listener.rb', line 237 def on_app_stopped(event) info "[#{event[:server_id]}] Stopped Karafka server" end |
#on_app_stopping(event) ⇒ Object
Logs info that we’re going to stop the Karafka server.
230 231 232 |
# File 'lib/karafka/instrumentation/logger_listener.rb', line 230 def on_app_stopping(event) info "[#{event[:server_id]}] Stopping Karafka server" end |
#on_client_pause(event) ⇒ Object
There may be no offset provided in case user wants to pause on the consecutive offset position. This can be beneficial when not wanting to purge the buffers.
Prints info about a consumer pause occurrence. Irrelevant if user or system initiated.
120 121 122 123 124 125 126 127 128 129 130 131 |
# File 'lib/karafka/instrumentation/logger_listener.rb', line 120 def on_client_pause(event) topic = event[:topic] partition = event[:partition] offset = event[:offset] client = event[:caller] info <<~MSG.tr("\n", " ").strip! [#{client.id}] Pausing on topic #{topic}-#{partition} on #{offset ? "offset #{offset}" : "the consecutive offset"} MSG end |
#on_client_resume(event) ⇒ Object
Prints information about resuming of processing of a given topic partition
136 137 138 139 140 141 142 143 144 |
# File 'lib/karafka/instrumentation/logger_listener.rb', line 136 def on_client_resume(event) topic = event[:topic] partition = event[:partition] client = event[:caller] info <<~MSG.tr("\n", " ").strip! [#{client.id}] Resuming on topic #{topic}-#{partition} MSG end |
#on_connection_listener_before_fetch_loop(event) ⇒ Object
28 29 30 31 32 33 34 35 36 37 38 |
# File 'lib/karafka/instrumentation/logger_listener.rb', line 28 def on_connection_listener_before_fetch_loop(event) listener_id = event[:caller].id subscription_group = event[:subscription_group] consumer_group_id = subscription_group.consumer_group.id topics = subscription_group.topics.select(&:active?).map(&:name).join(", ") group_details = "#{consumer_group_id}/#{subscription_group.id}" info( "[#{listener_id}] Group #{group_details} subscribing to topics: #{topics}" ) end |
#on_connection_listener_fetch_loop(event) ⇒ Object
Logs each messages fetching attempt
43 44 45 46 47 48 |
# File 'lib/karafka/instrumentation/logger_listener.rb', line 43 def on_connection_listener_fetch_loop(event) return unless log_polling? listener_id = event[:caller].id debug "[#{listener_id}] Polling messages..." end |
#on_connection_listener_fetch_loop_received(event) ⇒ Object
Logs about messages that we’ve received from Kafka
53 54 55 56 57 58 59 60 61 62 63 64 65 66 |
# File 'lib/karafka/instrumentation/logger_listener.rb', line 53 def on_connection_listener_fetch_loop_received(event) return unless log_polling? listener_id = event[:caller].id time = event[:time].round(2) = event[:messages_buffer].size = "[#{listener_id}] Polled #{} messages in #{time}ms" # We don't want the "polled 0" in dev as it would spam the log # Instead we publish only info when there was anything we could poll and fail over to the # zero notifications when in debug mode .zero? ? debug() : info() end |
#on_consumer_consuming_retry(event) ⇒ Object
Prints info about retry of processing after an error
149 150 151 152 153 154 155 156 157 158 159 160 |
# File 'lib/karafka/instrumentation/logger_listener.rb', line 149 def on_consumer_consuming_retry(event) topic = event[:topic] partition = event[:partition] offset = event[:offset] consumer = event[:caller] timeout = event[:timeout] info <<~MSG.tr("\n", " ").strip! [#{consumer.id}] Retrying of #{consumer.class} after #{timeout} ms on topic #{topic}-#{partition} from offset #{offset} MSG end |
#on_consumer_consuming_seek(event) ⇒ Object
Prints info about seeking to a particular location
165 166 167 168 169 170 171 172 173 174 175 |
# File 'lib/karafka/instrumentation/logger_listener.rb', line 165 def on_consumer_consuming_seek(event) topic = event[:topic] partition = event[:partition] seek_offset = event[:message].offset consumer = event[:caller] info <<~MSG.tr("\n", " ").strip! [#{consumer.id}] Seeking from #{consumer.class} on topic #{topic}-#{partition} to offset #{seek_offset} MSG end |
#on_dead_letter_queue_dispatched(event) ⇒ Object
Logs info when we have dispatched a message the the DLQ
280 281 282 283 284 285 286 287 288 289 290 291 292 293 |
# File 'lib/karafka/instrumentation/logger_listener.rb', line 280 def on_dead_letter_queue_dispatched(event) consumer = event[:caller] topic = consumer.topic.name = event[:message] offset = .offset dlq_topic = consumer.topic.dead_letter_queue.topic partition = .partition info <<~MSG.tr("\n", " ").strip! [#{consumer.id}] Dispatched message #{offset} from #{topic}-#{partition} to DLQ topic: #{dlq_topic} MSG end |
#on_error_occurred(event) ⇒ Object
There are many types of errors that can occur in many places, but we provide a single handler for all of them to simplify error instrumentation.
359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 |
# File 'lib/karafka/instrumentation/logger_listener.rb', line 359 def on_error_occurred(event) type = event[:type] error = event[:error] backtrace = (error.backtrace || []).join("\n") details = [error.to_s, error_details(event)].compact.join(" ") case type when "consumer.initialized.error" error "Consumer initialized error: #{details}" error backtrace when "consumer.wrap.error" error "Consumer wrap failed due to an error: #{details}" error backtrace when "consumer.consume.error" error "Consumer consuming error: #{details}" error backtrace when "consumer.revoked.error" error "Consumer on revoked failed due to an error: #{details}" error backtrace when "consumer.idle.error" error "Consumer idle failed due to an error: #{details}" error backtrace when "consumer.shutdown.error" error "Consumer on shutdown failed due to an error: #{details}" error backtrace when "consumer.tick.error" error "Consumer on tick failed due to an error: #{details}" error backtrace when "consumer.eofed.error" error "Consumer on eofed failed due to an error: #{details}" error backtrace when "consumer.after_consume.error" error "Consumer on after_consume failed due to an error: #{details}" error backtrace when "worker.process.error" fatal "Worker processing failed due to an error: #{details}" fatal backtrace when "connection.listener.fetch_loop.error" error "Listener fetch loop error: #{details}" error backtrace when "swarm.supervisor.error" fatal "Swarm supervisor crashed due to an error: #{details}" fatal backtrace when "runner.call.error" fatal "Runner crashed due to an error: #{details}" fatal backtrace when "app.stopping.error" active_listeners = event.payload[:active_listeners] alive_workers = event.payload[:alive_workers] in_processing = event.payload[:in_processing] = <<~MSG.tr("\n", " ").strip! Forceful Karafka server stop with: #{alive_workers.size} active workers and #{active_listeners.size} active listeners MSG error active_listeners.each do |listener| error "Listener #{listener.id} for #{listener.subscription_group.name} still active" end in_processing.each do |group_id, jobs| next if jobs.empty? jobs.each do |job| job_class = job.class.name.split("::").last topic_name = job.executor.topic.name partition = job.executor.partition error "In processing: #{job_class} job for #{topic_name}/#{partition} " \ "(group: #{group_id})" end end when "app.forceful_stopping.error" error "Forceful shutdown error occurred: #{details}" error backtrace when "librdkafka.error" error "librdkafka internal error occurred: #{details}" error backtrace # Those can occur when emitted statistics are consumed by the end user and the processing # of statistics fails. The statistics are emitted from librdkafka main loop thread and # any errors there crash the whole thread when "callbacks.statistics.error" error "callbacks.statistics processing failed due to an error: #{details}" error backtrace when "callbacks.error.error" error "callbacks.error processing failed due to an error: #{details}" error backtrace # Those will only occur when retries in the client fail and when they did not stop after # back-offs when "connection.client.poll.error" error "Data polling error occurred: #{details}" error backtrace when "connection.client.rebalance_callback.error" error "Rebalance callback error occurred: #{details}" error backtrace when "connection.client.unsubscribe.error" error "Client unsubscribe error occurred: #{details}" error backtrace when "parallel_segments.reducer.error" error "Parallel segments reducer error occurred: #{details}" error backtrace when "parallel_segments.partitioner.error" error "Parallel segments partitioner error occurred: #{details}" error backtrace when "virtual_partitions.partitioner.error" error "Virtual partitions partitioner error occurred: #{details}" error backtrace # This handles any custom errors coming from places like Web-UI, etc else error "#{type} error occurred: #{error.class} - #{details}" error backtrace end end |
#on_filtering_seek(event) ⇒ Object
314 315 316 317 318 319 320 321 322 323 324 325 326 |
# File 'lib/karafka/instrumentation/logger_listener.rb', line 314 def on_filtering_seek(event) consumer = event[:caller] topic = consumer.topic.name # Message to which we seek = event[:message] partition = .partition offset = .offset info <<~MSG.tr("\n", " ").strip! [#{consumer.id}] Post-filtering seeking to message #{offset} on #{topic}-#{partition} MSG end |
#on_filtering_throttled(event) ⇒ Object
Logs info about throttling event
298 299 300 301 302 303 304 305 306 307 308 309 310 311 |
# File 'lib/karafka/instrumentation/logger_listener.rb', line 298 def on_filtering_throttled(event) consumer = event[:caller] topic = consumer.topic.name # Here we get last message before throttle = event[:message] partition = .partition offset = .offset info <<~MSG.tr("\n", " ").strip! [#{consumer.id}] Throttled and will resume from message #{offset} on #{topic}-#{partition} MSG end |
#on_process_notice_signal(event) ⇒ Object
Logs info about system signals that Karafka received and prints backtrace for threads in case of ttin
181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 |
# File 'lib/karafka/instrumentation/logger_listener.rb', line 181 def on_process_notice_signal(event) server_id = Karafka::Server.id info "[#{server_id}] Received #{event[:signal]} system signal" # We print backtrace only for ttin return unless event[:signal] == :SIGTTIN # Inspired by Sidekiq Thread.list.each do |thread| tid = (thread.object_id ^ ::Process.pid).to_s(36) warn "" warn "Thread TID-#{tid} #{thread.name}" if thread.backtrace && !thread.backtrace.empty? warn thread.backtrace.join("\n") else warn "<no backtrace available>" end end end |
#on_rebalance_partitions_assigned(event) ⇒ Object
Logs info about partitions that we’ve gained
262 263 264 265 266 267 268 269 270 271 272 273 274 275 |
# File 'lib/karafka/instrumentation/logger_listener.rb', line 262 def on_rebalance_partitions_assigned(event) assigned_partitions = event[:tpl].to_h.transform_values { |part| part.map(&:partition) } group_id = event[:consumer_group_id] client_id = event[:client_id] group_prefix = "[#{client_id}] Group #{group_id} rebalance" if assigned_partitions.empty? info "#{group_prefix}: No partitions assigned" else assigned_partitions.each do |topic, partitions| info "#{group_prefix}: #{topic}-[#{partitions.join(",")}] assigned" end end end |
#on_rebalance_partitions_revoked(event) ⇒ Object
Logs info about partitions we have lost
244 245 246 247 248 249 250 251 252 253 254 255 256 257 |
# File 'lib/karafka/instrumentation/logger_listener.rb', line 244 def on_rebalance_partitions_revoked(event) revoked_partitions = event[:tpl].to_h.transform_values { |part| part.map(&:partition) } group_id = event[:consumer_group_id] client_id = event[:client_id] group_prefix = "[#{client_id}] Group #{group_id} rebalance" if revoked_partitions.empty? info "#{group_prefix}: No partitions revoked" else revoked_partitions.each do |topic, partitions| info "#{group_prefix}: #{topic}-[#{partitions.join(",")}] revoked" end end end |
#on_swarm_manager_before_fork(event) ⇒ Object
341 342 343 |
# File 'lib/karafka/instrumentation/logger_listener.rb', line 341 def on_swarm_manager_before_fork(event) debug "Swarm manager starting node with id: #{event[:node].id}" end |
#on_swarm_manager_control(event) ⇒ Object
351 352 353 354 |
# File 'lib/karafka/instrumentation/logger_listener.rb', line 351 def on_swarm_manager_control(event) pids = event[:caller].nodes.map(&:pid).join(", ") debug "Swarm manager checking nodes: #{pids}" end |
#on_swarm_manager_stopping(event) ⇒ Object
329 330 331 332 |
# File 'lib/karafka/instrumentation/logger_listener.rb', line 329 def on_swarm_manager_stopping(event) node = event[:node] error "Swarm manager detected unhealthy node #{node.pid}. Sending TERM signal..." end |
#on_swarm_manager_terminating(event) ⇒ Object
335 336 337 338 |
# File 'lib/karafka/instrumentation/logger_listener.rb', line 335 def on_swarm_manager_terminating(event) node = event[:node] error "Swarm manager detected unresponsive node #{node.pid}. Sending KILL signal..." end |
#on_swarm_node_after_fork(_event) ⇒ Object
346 347 348 |
# File 'lib/karafka/instrumentation/logger_listener.rb', line 346 def on_swarm_node_after_fork(_event) info "Swarm node #{::Process.pid} forked from #{::Process.ppid}" end |
#on_worker_process(event) ⇒ Object
Prints info about the fact that a given job has started
71 72 73 74 75 76 77 78 |
# File 'lib/karafka/instrumentation/logger_listener.rb', line 71 def on_worker_process(event) job = event[:job] job_type = job.class.to_s.split("::").last consumer = job.executor.topic.consumer topic = job.executor.topic.name partition = job.executor.partition info "[#{job.id}] #{job_type} job for #{consumer} on #{topic}-#{partition} started" end |
#on_worker_processed(event) ⇒ Object
Prints info about the fact that a given job has finished
83 84 85 86 87 88 89 90 91 92 93 94 |
# File 'lib/karafka/instrumentation/logger_listener.rb', line 83 def on_worker_processed(event) job = event[:job] time = event[:time].round(2) job_type = job.class.to_s.split("::").last consumer = job.executor.topic.consumer topic = job.executor.topic.name partition = job.executor.partition info <<~MSG.tr("\n", " ").strip! [#{job.id}] #{job_type} job for #{consumer} on #{topic}-#{partition} finished in #{time} ms MSG end |
#on_worker_scaling_down(event) ⇒ Object
Logs info about the workers pool scaling down. The actual size change happens asynchronously as workers pick up nil sentinels and exit.
109 110 111 112 113 |
# File 'lib/karafka/instrumentation/logger_listener.rb', line 109 def on_worker_scaling_down(event) from = event[:from] to = event[:to] info "Workers pool scaling down from #{from} to #{to} workers" end |
#on_worker_scaling_up(event) ⇒ Object
Logs info about the workers pool scaling up
99 100 101 102 103 |
# File 'lib/karafka/instrumentation/logger_listener.rb', line 99 def on_worker_scaling_up(event) from = event[:from] to = event[:to] info "Workers pool scaled up from #{from} to #{to} workers" end |