229
227
def make_log_request_dict(direction='reverse', specific_fileids=None,
230
228
start_revision=None, end_revision=None, limit=None,
231
message_search=None, levels=None, generate_tags=True,
229
message_search=None, levels=1, generate_tags=True,
233
231
diff_type=None, _match_using_deltas=True,
234
exclude_common_ancestry=False, match=None,
232
exclude_common_ancestry=False,
237
234
"""Convenience function for making a logging request dictionary.
259
256
matching commit messages
261
258
:param levels: the number of levels of revisions to
262
generate; 1 for just the mainline; 0 for all levels, or None for
259
generate; 1 for just the mainline; 0 for all levels.
265
261
:param generate_tags: If True, include tags for matched revisions.
267
263
:param delta_type: Either 'full', 'partial' or None.
268
264
'full' means generate the complete delta - adds/deletes/modifies/etc;
269
265
'partial' means filter the delta using specific_fileids;
282
278
:param exclude_common_ancestry: Whether -rX..Y should be interpreted as a
283
279
range operator or as a graph difference.
285
:param signature: show digital signature information
287
:param match: Dictionary of list of search strings to use when filtering
288
revisions. Keys can be 'message', 'author', 'committer', 'bugs' or
289
the empty string to match any of the preceding properties.
292
# Take care of old style message_search parameter
295
if 'message' in match:
296
match['message'].append(message_search)
298
match['message'] = [message_search]
300
match={ 'message': [message_search] }
302
282
'direction': direction,
303
283
'specific_fileids': specific_fileids,
304
284
'start_revision': start_revision,
305
285
'end_revision': end_revision,
287
'message_search': message_search,
307
288
'levels': levels,
308
289
'generate_tags': generate_tags,
309
290
'delta_type': delta_type,
310
291
'diff_type': diff_type,
311
292
'exclude_common_ancestry': exclude_common_ancestry,
312
'signature': signature,
314
293
# Add 'private' attributes for features that may be deprecated
315
294
'_match_using_deltas': _match_using_deltas,
327
def format_signature_validity(rev_id, repo):
328
"""get the signature validity
330
:param rev_id: revision id to validate
331
:param repo: repository of revision
332
:return: human readable string to print to log
334
from bzrlib import gpg
336
gpg_strategy = gpg.GPGStrategy(None)
337
result = repo.verify_revision(rev_id, gpg_strategy)
338
if result[0] == gpg.SIGNATURE_VALID:
339
return "valid signature from {0}".format(result[1])
340
if result[0] == gpg.SIGNATURE_KEY_MISSING:
341
return "unknown key {0}".format(result[1])
342
if result[0] == gpg.SIGNATURE_NOT_VALID:
343
return "invalid signature!"
344
if result[0] == gpg.SIGNATURE_NOT_SIGNED:
345
return "no signature"
348
306
class LogGenerator(object):
349
307
"""A generator of log revisions."""
395
353
# Tweak the LogRequest based on what the LogFormatter can handle.
396
354
# (There's no point generating stuff if the formatter can't display it.)
398
if rqst['levels'] is None or lf.get_levels() > rqst['levels']:
399
# user didn't specify levels, use whatever the LF can handle:
400
rqst['levels'] = lf.get_levels()
356
rqst['levels'] = lf.get_levels()
402
357
if not getattr(lf, 'supports_tags', False):
403
358
rqst['generate_tags'] = False
404
359
if not getattr(lf, 'supports_delta', False):
405
360
rqst['delta_type'] = None
406
361
if not getattr(lf, 'supports_diff', False):
407
362
rqst['diff_type'] = None
408
if not getattr(lf, 'supports_signatures', False):
409
rqst['signature'] = False
411
364
# Find and print the interesting revisions
412
365
generator = self._generator_factory(self.branch, rqst)
460
412
diff = self._format_diff(rev, rev_id, diff_type)
462
signature = format_signature_validity(rev_id,
463
self.branch.repository)
466
413
yield LogRevision(rev, revno, merge_depth, delta,
467
self.rev_tag_dict.get(rev_id), diff, signature)
414
self.rev_tag_dict.get(rev_id), diff)
470
417
if log_count >= limit:
526
473
# Apply the other filters
527
474
return make_log_rev_iterator(self.branch, view_revisions,
528
rqst.get('delta_type'), rqst.get('match'),
475
rqst.get('delta_type'), rqst.get('message_search'),
529
476
file_ids=rqst.get('specific_fileids'),
530
477
direction=rqst.get('direction'))
544
491
rqst.get('specific_fileids')[0], view_revisions,
545
492
include_merges=rqst.get('levels') != 1)
546
493
return make_log_rev_iterator(self.branch, view_revisions,
547
rqst.get('delta_type'), rqst.get('match'))
494
rqst.get('delta_type'), rqst.get('message_search'))
550
497
def _calc_view_revisions(branch, start_rev_id, end_rev_id, direction,
732
679
br_revno, br_rev_id = branch.last_revision_info()
733
680
repo = branch.repository
734
graph = repo.get_graph()
735
681
if start_rev_id is None and end_rev_id is None:
736
682
cur_revno = br_revno
737
for revision_id in graph.iter_lefthand_ancestry(br_rev_id,
738
(_mod_revision.NULL_REVISION,)):
683
for revision_id in repo.iter_reverse_revision_history(br_rev_id):
739
684
yield revision_id, str(cur_revno), 0
742
687
if end_rev_id is None:
743
688
end_rev_id = br_rev_id
744
689
found_start = start_rev_id is None
745
for revision_id in graph.iter_lefthand_ancestry(end_rev_id,
746
(_mod_revision.NULL_REVISION,)):
690
for revision_id in repo.iter_reverse_revision_history(end_rev_id):
747
691
revno_str = _compute_revno_str(branch, revision_id)
748
692
if not found_start and revision_id == start_rev_id:
749
693
if not exclude_common_ancestry:
801
745
yield rev_id, '.'.join(map(str, revno)), merge_depth
748
@deprecated_function(deprecated_in((2, 2, 0)))
749
def calculate_view_revisions(branch, start_revision, end_revision, direction,
750
specific_fileid, generate_merge_revisions):
751
"""Calculate the revisions to view.
753
:return: An iterator of (revision_id, dotted_revno, merge_depth) tuples OR
754
a list of the same tuples.
756
start_rev_id, end_rev_id = _get_revision_limits(branch, start_revision,
758
view_revisions = list(_calc_view_revisions(branch, start_rev_id, end_rev_id,
759
direction, generate_merge_revisions or specific_fileid))
761
view_revisions = _filter_revisions_touching_file_id(branch,
762
specific_fileid, view_revisions,
763
include_merges=generate_merge_revisions)
764
return _rebase_merge_depth(view_revisions)
804
767
def _rebase_merge_depth(view_revisions):
805
768
"""Adjust depths upwards so the top level is 0."""
806
769
# If either the first or last revision have a merge_depth of 0, we're done
850
813
return log_rev_iterator
853
def _make_search_filter(branch, generate_delta, match, log_rev_iterator):
816
def _make_search_filter(branch, generate_delta, search, log_rev_iterator):
854
817
"""Create a filtered iterator of log_rev_iterator matching on a regex.
856
819
:param branch: The branch being logged.
857
820
:param generate_delta: Whether to generate a delta for each revision.
858
:param match: A dictionary with properties as keys and lists of strings
859
as values. To match, a revision may match any of the supplied strings
860
within a single property but must match at least one string for each
821
:param search: A user text search string.
862
822
:param log_rev_iterator: An input iterator containing all revisions that
863
823
could be displayed, in lists.
864
824
:return: An iterator over lists of ((rev_id, revno, merge_depth), rev,
868
828
return log_rev_iterator
869
searchRE = [(k, [re.compile(x, re.IGNORECASE) for x in v])
870
for (k,v) in match.iteritems()]
871
return _filter_re(searchRE, log_rev_iterator)
874
def _filter_re(searchRE, log_rev_iterator):
829
searchRE = re.compile(search, re.IGNORECASE)
830
return _filter_message_re(searchRE, log_rev_iterator)
833
def _filter_message_re(searchRE, log_rev_iterator):
875
834
for revs in log_rev_iterator:
876
new_revs = [rev for rev in revs if _match_filter(searchRE, rev[1])]
880
def _match_filter(searchRE, rev):
882
'message': (rev.message,),
883
'committer': (rev.committer,),
884
'author': (rev.get_apparent_authors()),
885
'bugs': list(rev.iter_bugs())
887
strings[''] = [item for inner_list in strings.itervalues()
888
for item in inner_list]
889
for (k,v) in searchRE:
890
if k in strings and not _match_any_filter(strings[k], v):
894
def _match_any_filter(strings, res):
895
return any([filter(None, map(re.search, strings)) for re in res])
836
for (rev_id, revno, merge_depth), rev, delta in revs:
837
if searchRE.search(rev.message):
838
new_revs.append(((rev_id, revno, merge_depth), rev, delta))
897
842
def _make_delta_filter(branch, generate_delta, search, log_rev_iterator,
898
843
fileids=None, direction='reverse'):
1140
1086
cur_revno = branch_revno
1142
1088
mainline_revs = []
1143
graph = branch.repository.get_graph()
1144
for revision_id in graph.iter_lefthand_ancestry(
1145
branch_last_revision, (_mod_revision.NULL_REVISION,)):
1089
for revision_id in branch.repository.iter_reverse_revision_history(
1090
branch_last_revision):
1146
1091
if cur_revno < start_revno:
1147
1092
# We have gone far enough, but we always add 1 more revision
1148
1093
rev_nos[revision_id] = cur_revno
1162
1107
return mainline_revs, rev_nos, start_rev_id, end_rev_id
1110
@deprecated_function(deprecated_in((2, 2, 0)))
1111
def _filter_revision_range(view_revisions, start_rev_id, end_rev_id):
1112
"""Filter view_revisions based on revision ranges.
1114
:param view_revisions: A list of (revision_id, dotted_revno, merge_depth)
1115
tuples to be filtered.
1117
:param start_rev_id: If not NONE specifies the first revision to be logged.
1118
If NONE then all revisions up to the end_rev_id are logged.
1120
:param end_rev_id: If not NONE specifies the last revision to be logged.
1121
If NONE then all revisions up to the end of the log are logged.
1123
:return: The filtered view_revisions.
1125
if start_rev_id or end_rev_id:
1126
revision_ids = [r for r, n, d in view_revisions]
1128
start_index = revision_ids.index(start_rev_id)
1131
if start_rev_id == end_rev_id:
1132
end_index = start_index
1135
end_index = revision_ids.index(end_rev_id)
1137
end_index = len(view_revisions) - 1
1138
# To include the revisions merged into the last revision,
1139
# extend end_rev_id down to, but not including, the next rev
1140
# with the same or lesser merge_depth
1141
end_merge_depth = view_revisions[end_index][2]
1143
for index in xrange(end_index+1, len(view_revisions)+1):
1144
if view_revisions[index][2] <= end_merge_depth:
1145
end_index = index - 1
1148
# if the search falls off the end then log to the end as well
1149
end_index = len(view_revisions) - 1
1150
view_revisions = view_revisions[start_index:end_index+1]
1151
return view_revisions
1165
1154
def _filter_revisions_touching_file_id(branch, file_id, view_revisions,
1166
1155
include_merges=True):
1167
1156
r"""Return the list of revision ids which touch a given file id.
1238
@deprecated_function(deprecated_in((2, 2, 0)))
1239
def get_view_revisions(mainline_revs, rev_nos, branch, direction,
1240
include_merges=True):
1241
"""Produce an iterator of revisions to show
1242
:return: an iterator of (revision_id, revno, merge_depth)
1243
(if there is no revno for a revision, None is supplied)
1245
if not include_merges:
1246
revision_ids = mainline_revs[1:]
1247
if direction == 'reverse':
1248
revision_ids.reverse()
1249
for revision_id in revision_ids:
1250
yield revision_id, str(rev_nos[revision_id]), 0
1252
graph = branch.repository.get_graph()
1253
# This asks for all mainline revisions, which means we only have to spider
1254
# sideways, rather than depth history. That said, its still size-of-history
1255
# and should be addressed.
1256
# mainline_revisions always includes an extra revision at the beginning, so
1258
parent_map = dict(((key, value) for key, value in
1259
graph.iter_ancestry(mainline_revs[1:]) if value is not None))
1260
# filter out ghosts; merge_sort errors on ghosts.
1261
rev_graph = _mod_repository._strip_NULL_ghosts(parent_map)
1262
merge_sorted_revisions = tsort.merge_sort(
1266
generate_revno=True)
1268
if direction == 'forward':
1269
# forward means oldest first.
1270
merge_sorted_revisions = reverse_by_depth(merge_sorted_revisions)
1271
elif direction != 'reverse':
1272
raise ValueError('invalid direction %r' % direction)
1274
for (sequence, rev_id, merge_depth, revno, end_of_merge
1275
) in merge_sorted_revisions:
1276
yield rev_id, '.'.join(map(str, revno)), merge_depth
1249
1279
def reverse_by_depth(merge_sorted_revisions, _depth=0):
1250
1280
"""Reverse revisions by depth.
1348
1374
"""Create a LogFormatter.
1350
1376
:param to_file: the file to output to
1351
:param to_exact_file: if set, gives an output stream to which
1377
:param to_exact_file: if set, gives an output stream to which
1352
1378
non-Unicode diffs are written.
1353
1379
:param show_ids: if True, revision-ids are to be displayed
1354
1380
:param show_timezone: the timezone to use
1590
1612
if revision.delta is not None:
1591
1613
# Use the standard status output to display changes
1592
1614
from bzrlib.delta import report_delta
1593
report_delta(to_file, revision.delta, short_status=False,
1615
report_delta(to_file, revision.delta, short_status=False,
1594
1616
show_ids=self.show_ids, indent=indent)
1595
1617
if revision.diff is not None:
1596
1618
to_file.write(indent + 'diff:\n')
1662
1684
if revision.delta is not None:
1663
1685
# Use the standard status output to display changes
1664
1686
from bzrlib.delta import report_delta
1665
report_delta(to_file, revision.delta,
1666
short_status=self.delta_format==1,
1687
report_delta(to_file, revision.delta,
1688
short_status=self.delta_format==1,
1667
1689
show_ids=self.show_ids, indent=indent + offset)
1668
1690
if revision.diff is not None:
1669
1691
self.show_diff(self.to_exact_file, revision.diff, ' ')
1849
1871
'The committer')
1874
def show_one_log(revno, rev, delta, verbose, to_file, show_timezone):
1875
# deprecated; for compatibility
1876
lf = LongLogFormatter(to_file=to_file, show_timezone=show_timezone)
1877
lf.show(revno, rev, delta)
1852
1880
def show_changed_revisions(branch, old_rh, new_rh, to_file=None,
1853
1881
log_format='long'):
1854
1882
"""Show the change in revision history comparing the old revision history to the new one.