Class: Deimos::KafkaTopicInfo
- Inherits:
-
ActiveRecord::Base
- Object
- ActiveRecord::Base
- Deimos::KafkaTopicInfo
- Defined in:
- lib/deimos/kafka_topic_info.rb
Overview
Record that keeps track of which topics are being worked on by DbProducers.
Class Method Summary collapse
-
.clear_lock(topic, lock_id) ⇒ Object
This is called once a producer is finished working on a topic, i.e.
-
.heartbeat(topic, lock_id) ⇒ Object
Update the locked_at timestamp to indicate that the producer is still working on those messages and to continue.
-
.lock(topic, lock_id) ⇒ Boolean
Lock a topic for the given ID.
-
.ping_empty_topics(except_topics) ⇒ Object
Update all topics that aren't currently locked and have no messages waiting.
-
.register_error(topic, lock_id) ⇒ Object
The producer calls this if it gets an error sending messages.
Class Method Details
.clear_lock(topic, lock_id) ⇒ Object
This is called once a producer is finished working on a topic, i.e. there are no more messages to fetch. It unlocks the topic and moves on to the next one.
53 54 55 56 57 58 59 60 |
# File 'lib/deimos/kafka_topic_info.rb', line 53 def clear_lock(topic, lock_id) self.where(topic: topic, locked_by: lock_id). update_all(locked_by: nil, locked_at: nil, error: false, retries: 0, last_processed_at: Time.zone.now) end |
.heartbeat(topic, lock_id) ⇒ Object
Update the locked_at timestamp to indicate that the producer is still working on those messages and to continue.
96 97 98 99 |
# File 'lib/deimos/kafka_topic_info.rb', line 96 def heartbeat(topic, lock_id) self.where(topic: topic, locked_by: lock_id). update_all(locked_at: Time.zone.now) end |
.lock(topic, lock_id) ⇒ Boolean
Lock a topic for the given ID. Returns whether the lock was successful.
13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 |
# File 'lib/deimos/kafka_topic_info.rb', line 13 def lock(topic, lock_id) # Try to create it - it's fine if it already exists begin self.create(topic: topic, last_processed_at: Time.zone.now) rescue ActiveRecord::RecordNotUnique # continue on end # Lock the record qtopic = self.connection.quote(topic) qlock_id = self.connection.quote(lock_id) qtable = self.connection.quote_table_name('kafka_topic_info') qnow = self.connection.quote(Time.zone.now.to_s(:db)) qfalse = self.connection.quoted_false qtime = self.connection.quote(1.minute.ago.to_s(:db)) # If a record is marked as error and less than 1 minute old, # we don't want to pick it up even if not currently locked because # we worry we'll run into the same problem again. # Once it's more than 1 minute old, we figure it's OK to try again # so we can pick up any topic that's that old, even if it was # locked by someone, because it's the job of the producer to keep # updating the locked_at timestamp as they work on messages in that # topic. If the locked_at timestamp is that old, chances are that # the producer crashed. sql = <<~SQL UPDATE #{qtable} SET locked_by=#{qlock_id}, locked_at=#{qnow}, error=#{qfalse} WHERE topic=#{qtopic} AND ((locked_by IS NULL AND error=#{qfalse}) OR locked_at < #{qtime}) SQL self.connection.update(sql) self.where(locked_by: lock_id, topic: topic).any? end |
.ping_empty_topics(except_topics) ⇒ Object
Update all topics that aren't currently locked and have no messages waiting. It's OK if some messages get inserted in the middle of this because the point is that at least within a few milliseconds of each other, it wasn't locked and had no messages, meaning the topic was in a good state. realized had messages in them, meaning all other topics were empty.
69 70 71 72 73 74 75 |
# File 'lib/deimos/kafka_topic_info.rb', line 69 def ping_empty_topics(except_topics) records = KafkaTopicInfo.where(locked_by: nil). where('topic not in(?)', except_topics) records.each do |info| info.update_attribute(:last_processed_at, Time.zone.now) end end |
.register_error(topic, lock_id) ⇒ Object
The producer calls this if it gets an error sending messages. This essentially locks down this topic for 1 minute (for all producers) and allows the caller to continue to the next topic.
82 83 84 85 86 87 88 89 90 |
# File 'lib/deimos/kafka_topic_info.rb', line 82 def register_error(topic, lock_id) record = self.where(topic: topic, locked_by: lock_id).last attr_hash = { locked_by: nil, locked_at: Time.zone.now, error: true, retries: record.retries + 1 } record.attributes = attr_hash record.save! end |