~bzr-pqm/bzr/bzr.dev

362 by Martin Pool
- Import stat-cache code
1
# (C) 2005 Canonical Ltd
2
3
# This program is free software; you can redistribute it and/or modify
4
# it under the terms of the GNU General Public License as published by
5
# the Free Software Foundation; either version 2 of the License, or
6
# (at your option) any later version.
7
8
# This program is distributed in the hope that it will be useful,
9
# but WITHOUT ANY WARRANTY; without even the implied warranty of
10
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
11
# GNU General Public License for more details.
12
13
# You should have received a copy of the GNU General Public License
14
# along with this program; if not, write to the Free Software
15
# Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA  02111-1307  USA
16
17
import stat, os, sha, time
18
from binascii import b2a_qp, a2b_qp
19
20
from trace import mutter
21
22
23
# file fingerprints are: (path, size, mtime, ctime, ino, dev).
24
#
25
# if this is the same for this file as in the previous revision, we
26
# assume the content is the same and the SHA-1 is the same.
27
28
# This is stored in a fingerprint file that also contains the file-id
29
# and the content SHA-1.
30
31
# Thus for any given file we can quickly get the SHA-1, either from
32
# the cache or if the cache is out of date.
33
34
# At the moment this is stored in a simple textfile; it might be nice
35
# to use a tdb instead.
36
37
38
# What we need:
39
40
# build a new cache from scratch
41
# load cache, incrementally update it
42
43
# TODO: Have a paranoid mode where we always compare the texts and
44
# always recalculate the digest, to trap modification without stat
45
# change and SHA collisions.
46
47
48
49
def fingerprint(path, abspath):
50
    try:
51
        fs = os.lstat(abspath)
52
    except OSError:
53
        # might be missing, etc
54
        return None
55
56
    if stat.S_ISDIR(fs.st_mode):
57
        return None
58
59
    return (fs.st_size, fs.st_mtime,
60
            fs.st_ctime, fs.st_ino, fs.st_dev)
61
62
63
def write_cache(branch, entry_iter):
64
    outf = branch.controlfile('work-cache.tmp', 'wt')
65
    for entry in entry_iter:
66
        outf.write(entry[0] + ' ' + entry[1] + ' ')
67
        outf.write(b2a_qp(entry[2], True))
68
        outf.write(' %d %d %d %d %d\n' % entry[3:])
69
        
70
    outf.close()
71
    os.rename(branch.controlfilename('work-cache.tmp'),
72
              branch.controlfilename('work-cache'))
73
74
        
75
        
76
def load_cache(branch):
77
    cache = {}
78
79
    try:
80
        cachefile = branch.controlfile('work-cache', 'rt')
81
    except IOError:
82
        return cache
83
    
84
    for l in cachefile:
85
        f = l.split(' ')
86
        file_id = f[0]
87
        if file_id in cache:
88
            raise BzrError("duplicated file_id in cache: {%s}" % file_id)
89
        cache[file_id] = (f[0], f[1], a2b_qp(f[2])) + tuple([long(x) for x in f[3:]])
90
    return cache
91
92
93
94
95
def _files_from_inventory(inv):
96
    for path, ie in inv.iter_entries():
97
        if ie.kind != 'file':
98
            continue
99
        yield ie.file_id, path
100
    
101
102
def build_cache(branch):
103
    inv = branch.read_working_inventory()
104
105
    cache = {}
106
    _update_cache_from_list(branch, cache, _files_from_inventory(inv))
107
    
108
109
110
def update_cache(branch, inv):
111
    # TODO: It's supposed to be faster to stat the files in order by inum.
112
    # We don't directly know the inum of the files of course but we do
113
    # know where they were last sighted, so we can sort by that.
114
115
    cache = load_cache(branch)
116
    return _update_cache_from_list(branch, cache, _files_from_inventory(inv))
117
118
119
120
def _update_cache_from_list(branch, cache, to_update):
121
    """Update the cache to have info on the named files.
122
123
    to_update is a sequence of (file_id, path) pairs.
124
    """
125
    hardcheck = dirty = 0
126
    for file_id, path in to_update:
127
        fap = branch.abspath(path)
128
        fp = fingerprint(fap, path)
129
        cacheentry = cache.get(file_id)
130
131
        if fp == None: # not here
132
            if cacheentry:
133
                del cache[file_id]
134
                dirty += 1
135
            continue
136
137
        if cacheentry and (cacheentry[3:] == fp):
138
            continue                    # all stat fields unchanged
139
140
        hardcheck += 1
141
142
        dig = sha.new(file(fap, 'rb').read()).hexdigest()
143
144
        if cacheentry == None or dig != cacheentry[1]: 
145
            # if there was no previous entry for this file, or if the
146
            # SHA has changed, then update the cache
147
            cacheentry = (file_id, dig, path) + fp
148
            cache[file_id] = cacheentry
149
            dirty += 1
150
151
    mutter('work cache: read %d files, %d changed' % (hardcheck, dirty))
152
        
153
    if dirty:
154
        write_cache(branch, cache.itervalues())
155
156
    return cache