17
17
"""A simple least-recently-used (LRU) cache."""
19
from collections import deque
21
from bzrlib import symbol_versioning
25
class _LRUNode(object):
26
"""This maintains the linked-list which is the lru internals."""
28
__slots__ = ('prev', 'next', 'key', 'value', 'cleanup', 'size')
30
def __init__(self, key, value, cleanup=None):
35
self.cleanup = cleanup
36
# TODO: We could compute this 'on-the-fly' like we used to, and remove
37
# one pointer from this object, we just need to decide if it
38
# actually costs us much of anything in normal usage
50
return '%s(%r n:%r p:%r)' % (self.__class__.__name__, self.key,
53
def run_cleanup(self):
54
if self.cleanup is not None:
55
self.cleanup(self.key, self.value)
57
# Just make sure to break any refcycles, etc
24
61
class LRUCache(object):
33
70
DeprecationWarning)
34
71
after_cleanup_count = after_cleanup_size
37
self._queue = deque() # Track when things are accessed
38
self._refcount = {} # number of entries in self._queue for each key
73
# The "HEAD" of the lru linked list
74
self._most_recently_used = None
75
# The "TAIL" of the lru linked list
76
self._last_recently_used = None
39
77
self._update_max_cache(max_cache, after_cleanup_count)
41
79
def __contains__(self, key):
42
80
return key in self._cache
44
82
def __getitem__(self, key):
45
val = self._cache[key]
46
self._record_access(key)
83
node = self._cache[key]
84
# Inlined from _record_access to decrease the overhead of __getitem__
85
# We also have more knowledge about structure if __getitem__ is
86
# succeeding, then we know that self._most_recently_used must not be
88
mru = self._most_recently_used
90
# Nothing to do, this node is already at the head of the queue
92
elif node is self._last_recently_used:
93
self._last_recently_used = node.prev
94
# Remove this node from the old location
97
if node_prev is not None:
98
node_prev.next = node_next
99
if node_next is not None:
100
node_next.prev = node_prev
101
# Insert this node to the front of the list
104
self._most_recently_used = node
49
108
def __len__(self):
50
109
return len(self._cache)
112
"""Walk the LRU list, only meant to be used in tests."""
113
node = self._most_recently_used
115
if node.prev is not None:
116
raise AssertionError('the _most_recently_used entry is not'
117
' supposed to have a previous entry'
119
while node is not None:
120
if node.next is None:
121
if node is not self._last_recently_used:
122
raise AssertionError('only the last node should have'
123
' no next value: %s' % (node,))
125
if node.next.prev is not node:
126
raise AssertionError('inconsistency found, node.next.prev'
127
' != node: %s' % (node,))
128
if node.prev is None:
129
if node is not self._most_recently_used:
130
raise AssertionError('only the _most_recently_used should'
131
' not have a previous node: %s'
134
if node.prev.next is not node:
135
raise AssertionError('inconsistency found, node.prev.next'
136
' != node: %s' % (node,))
52
140
def add(self, key, value, cleanup=None):
53
141
"""Add a new value to the cache.
55
Also, if the entry is ever removed from the queue, call cleanup.
56
Passing it the key and value being removed.
143
Also, if the entry is ever removed from the cache, call
58
146
:param key: The key to store it under
59
147
:param value: The object to store
60
148
:param cleanup: None or a function taking (key, value) to indicate
61
'value' sohuld be cleaned up.
149
'value' should be cleaned up.
63
151
if key in self._cache:
65
self._cache[key] = value
66
if cleanup is not None:
67
self._cleanup[key] = cleanup
68
self._record_access(key)
152
node = self._cache[key]
155
node.cleanup = cleanup
157
node = _LRUNode(key, value, cleanup=cleanup)
158
self._cache[key] = node
159
self._record_access(node)
70
161
if len(self._cache) > self._max_cache:
71
162
# Trigger the cleanup
165
def cache_size(self):
166
"""Get the number of entries we will cache."""
167
return self._max_cache
74
169
def get(self, key, default=None):
75
if key in self._cache:
170
node = self._cache.get(key, None)
173
self._record_access(node)
80
177
"""Get the list of keys currently cached.
96
197
# Make sure the cache is shrunk to the correct size
97
198
while len(self._cache) > self._after_cleanup_count:
98
199
self._remove_lru()
99
# No need to compact the queue at this point, because the code that
100
# calls this would have already triggered it based on queue length
102
201
def __setitem__(self, key, value):
103
202
"""Add a value to the cache, there will be no cleanup function."""
104
203
self.add(key, value, cleanup=None)
106
def _record_access(self, key):
205
def _record_access(self, node):
107
206
"""Record that key was accessed."""
108
self._queue.append(key)
109
# Can't use setdefault because you can't += 1 the result
110
self._refcount[key] = self._refcount.get(key, 0) + 1
112
# If our access queue is too large, clean it up too
113
if len(self._queue) > self._compact_queue_length:
114
self._compact_queue()
116
def _compact_queue(self):
117
"""Compact the queue, leaving things in sorted last appended order."""
119
for item in self._queue:
120
if self._refcount[item] == 1:
121
new_queue.append(item)
123
self._refcount[item] -= 1
124
self._queue = new_queue
125
# All entries should be of the same size. There should be one entry in
126
# queue for each entry in cache, and all refcounts should == 1
127
if not (len(self._queue) == len(self._cache) ==
128
len(self._refcount) == sum(self._refcount.itervalues())):
129
raise AssertionError()
131
def _remove(self, key):
132
"""Remove an entry, making sure to maintain the invariants."""
133
cleanup = self._cleanup.pop(key, None)
134
val = self._cache.pop(key)
135
if cleanup is not None:
207
# Move 'node' to the front of the queue
208
if self._most_recently_used is None:
209
self._most_recently_used = node
210
self._last_recently_used = node
212
elif node is self._most_recently_used:
213
# Nothing to do, this node is already at the head of the queue
215
elif node is self._last_recently_used:
216
self._last_recently_used = node.prev
217
# We've taken care of the tail pointer, remove the node, and insert it
220
if node.prev is not None:
221
node.prev.next = node.next
222
if node.next is not None:
223
node.next.prev = node.prev
225
node.next = self._most_recently_used
226
self._most_recently_used = node
227
node.next.prev = node
230
def _remove_node(self, node):
231
if node is self._last_recently_used:
232
self._last_recently_used = node.prev
233
self._cache.pop(node.key)
234
# If we have removed all entries, remove the head pointer as well
235
if self._last_recently_used is None:
236
self._most_recently_used = None
139
239
def _remove_lru(self):
140
240
"""Remove one entry from the lru, and handle consequences.
200
294
self._compute_size = compute_size
201
295
if compute_size is None:
202
296
self._compute_size = len
203
# This approximates that texts are > 0.5k in size. It only really
204
# effects when we clean up the queue, so we don't want it to be too
206
297
self._update_max_size(max_size, after_cleanup_size=after_cleanup_size)
207
298
LRUCache.__init__(self, max_cache=max(int(max_size/512), 1))
209
300
def add(self, key, value, cleanup=None):
210
301
"""Add a new value to the cache.
212
Also, if the entry is ever removed from the queue, call cleanup.
213
Passing it the key and value being removed.
303
Also, if the entry is ever removed from the cache, call
215
306
:param key: The key to store it under
216
307
:param value: The object to store
217
308
:param cleanup: None or a function taking (key, value) to indicate
218
'value' sohuld be cleaned up.
309
'value' should be cleaned up.
220
if key in self._cache:
311
node = self._cache.get(key, None)
222
312
value_len = self._compute_size(value)
223
313
if value_len >= self._after_cleanup_size:
314
# The new value is 'too big to fit', as it would fill up/overflow
315
# the cache all by itself
316
trace.mutter('Adding the key %r to an LRUSizeCache failed.'
317
' value %d is too big to fit in a the cache'
318
' with size %d %d', key, value_len,
319
self._after_cleanup_size, self._max_size)
321
# We won't be replacing the old node, so just remove it
322
self._remove_node(node)
323
if cleanup is not None:
327
node = _LRUNode(key, value, cleanup=cleanup)
328
self._cache[key] = node
330
self._value_size -= node.size
331
node.size = value_len
225
332
self._value_size += value_len
226
self._cache[key] = value
227
if cleanup is not None:
228
self._cleanup[key] = cleanup
229
self._record_access(key)
333
self._record_access(node)
231
335
if self._value_size > self._max_size:
232
336
# Time to cleanup
242
346
while self._value_size > self._after_cleanup_size:
243
347
self._remove_lru()
245
def _remove(self, key):
246
"""Remove an entry, making sure to maintain the invariants."""
247
val = LRUCache._remove(self, key)
248
self._value_size -= self._compute_size(val)
349
def _remove_node(self, node):
350
self._value_size -= node.size
351
LRUCache._remove_node(self, node)
250
353
def resize(self, max_size, after_cleanup_size=None):
251
354
"""Change the number of bytes that will be cached."""