359. Logger Rate Limiter - Detailed Explanation
Problem Statement
Design a logger system that receives a stream of messages with timestamps and prints each unique message at most once every 10 seconds. In other words, if a message has been printed in the last 10 seconds, any new occurrences of that same message should be throttled (ignored). If 10 seconds or more have elapsed since it was last printed, the message can be printed again.
- Input: A series of calls to
logger.shouldPrintMessage(timestamp, message)
wheretimestamp
is in seconds (non-decreasing order) andmessage
is a string (lowercase English letters). - Output: Each call returns
true
if the message should be printed at that timestamp (not printed in the last 10 seconds), orfalse
if it should be skipped due to the rate limit.
Why? This ensures no duplicate message is printed too frequently. You can imagine it like hearing someone shout “Hey!” at a party – you only respond if at least 10 seconds have passed since the last identical shout. If they repeat sooner, you ignore them.
Example
Suppose the logger receives the following calls in order:
logger.shouldPrintMessage(1, "hello") ➞ true logger.shouldPrintMessage(2, "world") ➞ true logger.shouldPrintMessage(3, "hello") ➞ false logger.shouldPrintMessage(11, "hello") ➞ true
Explanation:
- At timestamp 1, message
"hello"
has not been seen before, so it prints (true
). - At timestamp 2, message
"world"
is new, so it prints (true
). - At timestamp 3, message
"hello"
appears again only 2 seconds after it was last printed. Since 2 < 10 seconds, it is throttled (false
). - At timestamp 11, message
"hello"
appears again. Now 10 seconds have passed since it was last printed at t=1, so it can print again (true
).
This meets the requirement that “hello” was printed at most once in any 10-second window.
Key Points and Constraints
- Timestamps are non-decreasing (they won’t go backward, but can be equal or increase).
- Each message is a lowercase string up to 30 characters.
- Timestamps can range from 0 up to 10^9.
- We need efficient performance for potentially many calls. Each
shouldPrintMessage
call should be fast (ideally O(1) time). - We must ensure we don’t use excessive memory (avoid storing unnecessary or expired data).
Hints to Solve the Problem:
-
Use a data structure to remember the last time each message was printed. This structure should allow quick checks and updates. A hash map (dictionary) is ideal for direct lookups by message.
-
On each new message event, check the stored timestamp for that message (if any). Compute the time difference with the current timestamp.
-
If the message was printed in the past 10 seconds (difference < 10), return
false
(skip it). -
Otherwise (message is new or it’s been ≥10 seconds since last print), update the stored timestamp to the current time and return
true
. -
Optionally, consider cleaning up or limiting stored data. Since timestamps grow and messages may stop coming, think about removing old entries if they are no longer needed (especially in certain approach).
With these ideas in mind, let's explore two approaches: one using a HashMap and another using a Queue.
Approach 1: Using a HashMap (Optimal Solution)
Idea: Maintain a hash map (dictionary) where keys are message strings and values are the last timestamp at which that message was printed. This way, whenever a message comes in, we can instantly check when it was last seen.
- When a new message arrives: look up its last printed timestamp in the map.
- If the message is not in the map (never seen before), we can print it. Store the current timestamp in the map for that message.
- If it exists in the map, calculate the difference:
current_timestamp - last_timestamp
.- If this difference is less than 10, it means the last print was within the past 10 seconds – so we return false (do not print again yet).
- If the difference is ≥ 10 seconds, we return true and update the map with the new timestamp (meaning we print it now and reset the 10-second window).
This approach directly enforces the rule by always keeping track of the most recent print time for each message. It uses O(1) time on average for each check since hash map operations (get/put) are constant time. It’s also space-efficient, storing at most one entry per unique message.
Why this is optimal: We only store necessary data (one timestamp per message) and each message check/update is constant time. There’s no need to iterate over other messages or timestamps when processing a new message. This satisfies the efficiency requirements even if there are many calls.
Potential drawback: The map can grow if there are many unique messages over time and we never remove old ones. In a long-running system, if messages stop appearing, their timestamps might remain stored. However, since each entry is small (just an integer timestamp) and the problem constraints don’t demand removal of stale data, this is usually acceptable. If needed, a periodic cleanup of very old entries could be implemented, but it’s not required for correctness.
Approach 1 – Example Walkthrough:
Using the earlier example, let’s see how the HashMap evolves and how decisions are made:
- Start: map is empty
{}
. - t=1, "hello": Not in map, so allow print. Update map:
{"hello": 1}
. - t=2, "world": Not in map, allow print. Update map:
{"hello": 1, "world": 2}
. - t=3, "hello": Map says
"hello"
last seen at 1. Difference = 3-1 = 2, which is < 10, so deny print (false
). (Map remains{"hello": 1, "world": 2}
unchanged since we didn’t print.) - t=11, "hello": Last seen at 1. Difference = 11-1 = 10, which is >= 10, so allow print (
true
). Update map:{"hello": 11, "world": 2}
(now “hello”’s last print time is 11).
At the end, the map tells us “hello” was last printed at 11 and “world” at 2. This matches the rule that “hello” was printed at t=1 and t=11, but not in between.
Why not just use a Set or List?
We need to know when the message was last printed, not just that it was printed. A set of messages wouldn’t tell us how recently it was printed. A list of past messages would let us check the last occurrence, but finding that last occurrence for each message would be inefficient (linear search). A map directly gives the last occurrence time in O(1) time, making it the best choice for this task.
Approach 2: Using a Queue (Alternative Approach)
Idea: An alternative is to use a queue to keep track of messages in the last 10-second window, coupled with a set for quick lookup. This approach explicitly maintains the sliding window of recent messages:
- Use a queue of pairs
(timestamp, message)
representing messages that have been printed and are still within the last 10 seconds window. New messages get added to the back of the queue when printed. - Use a set of messages currently in that queue (the messages printed in the last 10 seconds). This allows us to quickly check if a new message is a duplicate of something in the recent window.
When a new message arrives:
-
Evict old messages: First, we remove any entries from the front of the queue that are older than 10 seconds compared to the current timestamp. For each removed entry, also remove that message from the set. This ensures the queue/set only contain messages from the last <10 seconds. (If current time is
t
, pop from queue whilecurrent_t - front_of_queue.timestamp >= 10
.). -
After cleaning, check if the incoming
message
is in the set of recent messages.- If yes, that means the message was printed within the last 10 seconds (still in our queue window), so we return false (throttle it).
- If no, then it’s not a duplicate in the last 10 seconds. We can print it: push
(timestamp, message)
onto the queue and add the message to the set, then returntrue
.
This approach achieves the same outcome as the hash map: the set ensures no message repeats in a 10-second span, and the queue automatically drops messages once they exit the 10-second window. The logic is a bit more involved, but it has a nice side effect of naturally cleaning up old messages (so the data structure doesn’t grow without bound). At any time, the queue contains at most 10 seconds worth of messages. In the worst case, if messages are very frequent, the queue size could be proportional to all messages in a 10-second span.
Efficiency: Checking and adding to a set is O(1) average. Removing old messages from the queue is also typically O(1) per message (each message enters and leaves the queue at most once). In the average case, each incoming message triggers a constant amount of work (amortized O(1) time per message). In the worst case, if many messages expire at once, a single call might clean up multiple old entries, making that call O(n) where n is the number of messages in the 10-second window. Overall, if there are N messages total, each message is enqueued and dequeued once, so the total work is O(N) (which averages to O(1) per message). Space usage is O(N) in worst case (if many unique messages within 10 seconds).
Approach 2 – Example Walkthrough:
Let’s walk through the same example using the queue-set approach to visualize how it works:
- Start: queue is empty
[]
, set is empty{}
. - t=1, "hello": Clean-up step: queue empty (nothing to remove).
"hello"
not in set, so we print it. Enqueue(1,"hello")
, set becomes{"hello"}
. - t=2, "world": Clean-up: front of queue is
(1,"hello")
. Check age: 2-1 = 1 < 10, so nothing removed."world"
not in set, print it. Enqueue(2,"world")
, set becomes{"hello","world"}
. - t=3, "hello": Clean-up: front
(1,"hello")
, 3-1=2 < 10, nothing removed (still within window). Now check"hello"
– it is in the set ("hello"
was printed at t=1 which is still within 10s). So we do NOT print. (Queue and set remain unchanged: still[(1,"hello"), (2,"world")]
and{"hello","world"}
.) - t=11, "hello": Clean-up: front
(1,"hello")
, 11-1=10, which is >=10, so remove(1,"hello")
from queue and remove"hello"
from set. Now front is(2,"world")
: 11-2=9 < 10, stop removing. After cleanup, queue contains[(2,"world")]
, set is{"world"}
. Now check"hello"
– it’s no longer in the set (we just removed the old occurrence). Therefore, it’s allowed to print. Enqueue(11,"hello")
, set becomes{"world","hello"}
. Returntrue
.
The results (true, true, false, true
) match the requirements. The queue approach ensures that at t=11, the old "hello"
from t=1 was purged, so the new "hello"
is treated as fresh.
Code Implementations
Python – Approach 1 (Using Dictionary)
Explanation: The Logger
class uses a dictionary last_print_time
to track the last timestamp a message was printed. In shouldPrintMessage
, it checks the dictionary:
-
If the message isn’t in the dictionary, it hasn’t been printed before, so we log it now and store the timestamp.
-
If it is in the dictionary, we see when it was last printed. If the difference is under 10, we return False (too soon to print again). If 10 or more, we update the timestamp and return True.
This is efficient since dictionary lookups and updates are O(1) average.
Java – Approach 1 (Using HashMap)
Explanation: This Java code mirrors the Python logic. A HashMap lastPrintTime
stores the last timestamp a message was printed. On each call:
- If the message is absent in the map, insert it with the current timestamp and return true.
- If present, check the difference. If it’s less than 10, return false; otherwise update the timestamp in the map and return true.
Both versions ensure a message won’t be printed again until at least 10 seconds have passed since its last print.
Python – Approach 2 (Using Queue + Set)
Explanation: We use a deque
(double-ended queue) to store messages that have been printed in the last 10 seconds, and a set
for quick membership checks. On each call:
-
We first purge the queue of any messages older than 10 seconds compared to the current timestamp (pop from left until the condition fails). For each popped entry, also remove that message from the set.
-
Then we check the set. If the incoming
message
is present, it means it was printed recently, so return False. If not present, we add it to both the queue and set (as it will now be in the recent window) and return True.
This approach explicitly maintains the 10-second sliding window of messages.
Java – Approach 2 (Using Queue + Set)
Explanation: This Java version uses an ArrayDeque
for the queue and a HashSet for recent messages. It follows the same steps: purge old messages, then check/insert the new message. The logic is identical to the Python version. (Note: Pair<Integer,String>
here is assumed to be a simple tuple class or you could use its equivalent in Java like AbstractMap.SimpleEntry
or similar to store the pair.)
Complexity Analysis
-
Approach 1 (HashMap/Dictionary): Each call to
shouldPrintMessage
does a constant amount of work – a hash lookup and at most one update. So time complexity per call is O(1) on average (amortized), and O(1) for the constructor as well . The space complexity is O(M), where M is the number of distinct messages that have been seen. In the worst case, if all messages are unique, the map will have an entry for each. This is typically not an issue given problem constraints, but it means space grows with the variety of messages. -
Approach 2 (Queue + Set): In the average case, each message gets enqueued and dequeued at most once, so over N calls the total operations are about 2N (each message enters and leaves the queue once). This gives an amortized time complexity of O(1) per call. However, a single call may dequeue many expired messages at once, so worst-case time for one call can be O(N) if N messages were queued and all expire at that moment . Still, across many calls, the work is evenly spread. The space complexity is O(N_window), where N_window is the number of messages in the 10-second window (since we store only recent messages in the queue and set). In the worst-case, if a huge number of messages arrive in a 10-second span (all unique), the space could approach O(N) (similar to approach 1’s worst case). The benefit is that messages older than 10 seconds are automatically removed, so the space usage at any given time is limited to the recent window.
Comparison: Approach 1 uses a bit less bookkeeping and offers straightforward O(1) time per call, but it doesn’t automatically purge old data. Approach 2 manages a sliding window and ensures we only keep recent data, at the cost of a slightly more complex implementation and occasionally extra work to dequeue old entries. For most interview scenarios and given the constraints, Approach 1 (HashMap) is considered optimal and simpler.
Step-by-Step Walkthrough with Visuals
Let’s illustrate how the system behaves over time with an example timeline. We’ll use the same example input and show the internal state changes for Approach 1 (HashMap):
Timeline of events:
-
Time 1: Message = "hello"
- Check: "hello" not seen before (map has no "hello").
- Action: Print "hello".
- Update:
map = {"hello": 1}
(store last seen time 1 for "hello"). - Output: true (message printed).
-
Time 2: Message = "world"
- Check: "world" not in map.
- Action: Print "world".
- Update:
map = {"hello": 1, "world": 2}
. - Output: true.
-
Time 3: Message = "hello"
- Check: "hello" last seen at time 1 (from map). Difference = 3 - 1 = 2 seconds.
- Decision: 2 seconds < 10 seconds, so this is too soon to print "hello" again.
- Action: Do not print "hello". (No map update, "hello" remains last seen at 1.)
- Output: false (throttled).
-
Time 11: Message = "hello"
- Check: "hello" last seen at time 1. Difference = 11 - 1 = 10 seconds.
- Decision: 10 seconds ≥ 10, so it’s been long enough to print "hello" again.
- Action: Print "hello".
- Update:
map = {"hello": 11, "world": 2}
(update "hello"’s last seen to 11). - Output: true.
Visualizing the 10-second rule: between time 1 and time 11, 10 seconds elapsed. The message "hello" was allowed again at time 11, but the attempt at time 3 (only 2 seconds after the initial print) was blocked. The message "world", on the other hand, was printed at time 2 without issue since it was a different message and had no conflict with "hello."
If we were to visualize Approach 2 (queue) over the same timeline, it would look like this: at time 3, "hello" was still in the queue from time 1, so it blocked the new "hello"; by time 11, the entry from time 1 would have been dequeued, making "hello" available to print again. Both approaches yield the same behavior as shown above.
Common Mistakes & Edge Cases
-
Forgetting to update the timestamp after printing: If you return true but don’t update the last seen time in the map (or don’t enqueue the message in the queue approach), subsequent checks will be wrong. Always update the state when a message is printed.
-
Off-by-one errors in time difference: Be careful with the 10-second condition. If a message was last printed at time t, it should be allowed at time t+10 or later. In our logic, we check
if (current - last < 10) then block
. This correctly blocks up to 9 seconds difference and allows exactly 10 seconds difference. For example, last print at 1 allows next at 11, but not at 10. This matches the problem requirement (as seen in examples). Make sure to handle the equality properly (≥10 means allow) to avoid one second off error. -
Not handling the first-time message: Always handle the case where the message has never been seen (map lookup returns null or key not present). The first occurrence should always print.
-
Using a data structure with slow lookup: A common mistake is using a list of recent messages and checking
if message in list
. This results in O(n) time per call in the worst case, which can be too slow if many messages come in. A set or hash map should be used for quick membership tests. -
Memory bloat in Approach 1: In a scenario with an extremely large number of unique messages and none repeating, the hash map will keep growing. While this is typically acceptable within constraint limits, forgetting that old entries remain forever might be an issue in long-running systems. One might consider purging entries that haven’t been seen for a long time if memory is a concern (not required in the context of this problem’s constraints, but relevant in system design).
-
Assuming strict increasing timestamps: The problem says non-decreasing, meaning timestamps can stay the same or jump. Our solutions handle this since if two messages have the same timestamp, they are processed in order given. If the same message appears twice at the exact same timestamp, Approach 1 would allow the first and throttle the second (since the second finds last timestamp equal and 0 difference < 10). Approach 2 would similarly block the second because the first would still be in the queue/set. This is an edge scenario but our logic covers it. Just ensure the implementation doesn’t wrongly assume strictly increasing time (e.g., don’t remove from queue until
>=10
difference as we did, not>10
, to handle exactly 10s properly).
Alternative Variations and Related Problems
-
Different Time Windows or Rates: A variation could ask for a message to be printed at most once every k seconds (not just 10). The solution approach would be the same; just replace 10 with k in the checks. Another twist could be allowing at most N occurrences of the same message in a window of time. That would require counting occurrences, perhaps using a queue to drop old occurrences (similar to “Hit Counter” problems).
-
Design Hit Counter: A very similar problem is Design Hit Counter (LeetCode 362) where you count how many hits (events) occurred in the past 5 minutes. The typical solution also uses a queue to only store recent events and purge old ones, which is analogous to Approach 2. It doesn’t involve distinct messages, just counting events in a window, but the sliding window technique is the same. It’s a good practice problem to understand rate-limiting concepts.
-
Rate Limiting in Distributed Systems: In real-world systems (APIs, web servers), you might need to limit how often a user or client can perform an action. This logger problem is a simplified version of that concept. In distributed systems, a common approach is to use a token bucket or leaky bucket algorithm, or a centralized store (like Redis) to track requests per user. The idea of “allow one action per 10 seconds” could be enforced with a similar timestamp check in a database or in-memory store, albeit with more complexities like synchronization, clock differences, and scaling.
-
Message Throttling: A related interview problem is designing a message throttling or debounce system, where you drop or delay messages that come too frequently. The logic is often similar: track last times or counts in windows. For instance, ensure a notification is sent at most once per minute, etc.
-
Sliding Window Problems: More generally, any problem that involves “in the last X time (or last X items)” can often be tackled with a sliding window technique using a queue or two-pointer method. This logger problem is a good introduction to that pattern, focusing on time instead of array indices.
-
API Rate Limiter: If you want to explore further, think about how you would extend this to multiple instances or a web service. You’d need a shared store to record last message times per user or message, and possibly a more sophisticated strategy if the limits are not just one message per 10 seconds but, say, 5 messages per minute (which involves counting within a window). This leads into system design territory.
GET YOUR FREE
Coding Questions Catalog
