2010-06-01 20:18:57 +04:00
|
|
|
# revset.py - revision set queries for mercurial
|
|
|
|
#
|
|
|
|
# Copyright 2010 Matt Mackall <mpm@selenic.com>
|
|
|
|
#
|
|
|
|
# This software may be used and distributed according to the terms of the
|
|
|
|
# GNU General Public License version 2 or any later version.
|
|
|
|
|
2012-06-02 04:05:31 +04:00
|
|
|
import re
|
2012-01-06 13:04:20 +04:00
|
|
|
import parser, util, error, discovery, hbisect, phases
|
2012-04-14 00:32:49 +04:00
|
|
|
import node
|
2014-02-07 22:32:02 +04:00
|
|
|
import heapq
|
2010-08-30 16:38:24 +04:00
|
|
|
import match as matchmod
|
2013-11-16 20:57:08 +04:00
|
|
|
import ancestor as ancestormod
|
2011-03-12 14:46:31 +03:00
|
|
|
from i18n import _
|
2011-12-25 15:35:16 +04:00
|
|
|
import encoding
|
2012-08-28 22:52:04 +04:00
|
|
|
import obsolete as obsmod
|
2014-01-17 18:42:12 +04:00
|
|
|
import pathutil
|
2013-01-03 21:48:14 +04:00
|
|
|
import repoview
|
2010-06-01 20:18:57 +04:00
|
|
|
|
2012-04-11 13:25:34 +04:00
|
|
|
def _revancestors(repo, revs, followfirst):
|
|
|
|
"""Like revlog.ancestors(), but supports followfirst."""
|
|
|
|
cut = followfirst and 1 or None
|
|
|
|
cl = repo.changelog
|
2014-02-07 22:32:02 +04:00
|
|
|
|
|
|
|
def iterate():
|
2014-02-08 01:44:57 +04:00
|
|
|
revqueue, revsnode = None, None
|
|
|
|
h = []
|
|
|
|
|
|
|
|
revs.descending()
|
|
|
|
revqueue = util.deque(revs)
|
|
|
|
if revqueue:
|
|
|
|
revsnode = revqueue.popleft()
|
|
|
|
heapq.heappush(h, -revsnode)
|
|
|
|
|
|
|
|
seen = set([node.nullrev])
|
2014-02-07 22:32:02 +04:00
|
|
|
while h:
|
|
|
|
current = -heapq.heappop(h)
|
|
|
|
if current not in seen:
|
2014-02-08 01:44:57 +04:00
|
|
|
if revsnode and current == revsnode:
|
|
|
|
if revqueue:
|
|
|
|
revsnode = revqueue.popleft()
|
|
|
|
heapq.heappush(h, -revsnode)
|
2014-02-07 22:32:02 +04:00
|
|
|
seen.add(current)
|
|
|
|
yield current
|
|
|
|
for parent in cl.parentrevs(current)[:cut]:
|
|
|
|
if parent != node.nullrev:
|
|
|
|
heapq.heappush(h, -parent)
|
|
|
|
|
2014-03-13 04:19:46 +04:00
|
|
|
return _descgeneratorset(iterate())
|
2012-04-11 13:25:34 +04:00
|
|
|
|
|
|
|
def _revdescendants(repo, revs, followfirst):
|
|
|
|
"""Like revlog.descendants() but supports followfirst."""
|
|
|
|
cut = followfirst and 1 or None
|
2014-02-11 00:26:45 +04:00
|
|
|
|
|
|
|
def iterate():
|
|
|
|
cl = repo.changelog
|
|
|
|
first = min(revs)
|
|
|
|
nullrev = node.nullrev
|
|
|
|
if first == nullrev:
|
|
|
|
# Are there nodes with a null first parent and a non-null
|
|
|
|
# second one? Maybe. Do we care? Probably not.
|
|
|
|
for i in cl:
|
2012-04-11 13:25:34 +04:00
|
|
|
yield i
|
2014-02-11 00:26:45 +04:00
|
|
|
else:
|
|
|
|
seen = set(revs)
|
|
|
|
for i in cl.revs(first + 1):
|
|
|
|
for x in cl.parentrevs(i)[:cut]:
|
|
|
|
if x != nullrev and x in seen:
|
|
|
|
seen.add(i)
|
|
|
|
yield i
|
|
|
|
break
|
|
|
|
|
2014-03-13 04:18:54 +04:00
|
|
|
return _ascgeneratorset(iterate())
|
2012-04-11 13:25:34 +04:00
|
|
|
|
2012-06-02 02:50:22 +04:00
|
|
|
def _revsbetween(repo, roots, heads):
|
|
|
|
"""Return all paths between roots and heads, inclusive of both endpoint
|
|
|
|
sets."""
|
|
|
|
if not roots:
|
2014-01-21 23:39:26 +04:00
|
|
|
return baseset([])
|
2012-06-02 02:50:22 +04:00
|
|
|
parentrevs = repo.changelog.parentrevs
|
2014-09-17 09:55:49 +04:00
|
|
|
visit = list(heads)
|
2012-06-02 02:50:22 +04:00
|
|
|
reachable = set()
|
|
|
|
seen = {}
|
|
|
|
minroot = min(roots)
|
|
|
|
roots = set(roots)
|
|
|
|
# open-code the post-order traversal due to the tiny size of
|
|
|
|
# sys.getrecursionlimit()
|
|
|
|
while visit:
|
|
|
|
rev = visit.pop()
|
|
|
|
if rev in roots:
|
|
|
|
reachable.add(rev)
|
|
|
|
parents = parentrevs(rev)
|
|
|
|
seen[rev] = parents
|
|
|
|
for parent in parents:
|
|
|
|
if parent >= minroot and parent not in seen:
|
|
|
|
visit.append(parent)
|
|
|
|
if not reachable:
|
2014-01-21 23:39:26 +04:00
|
|
|
return baseset([])
|
2012-06-02 02:50:22 +04:00
|
|
|
for rev in sorted(seen):
|
|
|
|
for parent in seen[rev]:
|
|
|
|
if parent in reachable:
|
|
|
|
reachable.add(rev)
|
2014-01-21 23:39:26 +04:00
|
|
|
return baseset(sorted(reachable))
|
2012-06-02 02:50:22 +04:00
|
|
|
|
2010-06-01 20:18:57 +04:00
|
|
|
elements = {
|
|
|
|
"(": (20, ("group", 1, ")"), ("func", 1, ")")),
|
2011-04-30 19:43:04 +04:00
|
|
|
"~": (18, None, ("ancestor", 18)),
|
|
|
|
"^": (18, None, ("parent", 18), ("parentpost", 18)),
|
2010-10-07 20:45:17 +04:00
|
|
|
"-": (5, ("negate", 19), ("minus", 5)),
|
2010-06-02 23:07:46 +04:00
|
|
|
"::": (17, ("dagrangepre", 17), ("dagrange", 17),
|
|
|
|
("dagrangepost", 17)),
|
|
|
|
"..": (17, ("dagrangepre", 17), ("dagrange", 17),
|
|
|
|
("dagrangepost", 17)),
|
|
|
|
":": (15, ("rangepre", 15), ("range", 15), ("rangepost", 15)),
|
2010-06-01 20:18:57 +04:00
|
|
|
"not": (10, ("not", 10)),
|
|
|
|
"!": (10, ("not", 10)),
|
|
|
|
"and": (5, None, ("and", 5)),
|
|
|
|
"&": (5, None, ("and", 5)),
|
|
|
|
"or": (4, None, ("or", 4)),
|
|
|
|
"|": (4, None, ("or", 4)),
|
|
|
|
"+": (4, None, ("or", 4)),
|
|
|
|
",": (2, None, ("list", 2)),
|
|
|
|
")": (0, None, None),
|
|
|
|
"symbol": (0, ("symbol",), None),
|
|
|
|
"string": (0, ("string",), None),
|
|
|
|
"end": (0, None, None),
|
|
|
|
}
|
|
|
|
|
|
|
|
keywords = set(['and', 'or', 'not'])
|
|
|
|
|
2014-03-19 02:19:44 +04:00
|
|
|
def tokenize(program, lookup=None):
|
2012-10-31 03:48:44 +04:00
|
|
|
'''
|
|
|
|
Parse a revset statement into a stream of tokens
|
|
|
|
|
|
|
|
Check that @ is a valid unquoted token character (issue3686):
|
|
|
|
>>> list(tokenize("@::"))
|
|
|
|
[('symbol', '@', 0), ('::', None, 1), ('end', None, 3)]
|
|
|
|
|
|
|
|
'''
|
|
|
|
|
2010-06-01 20:18:57 +04:00
|
|
|
pos, l = 0, len(program)
|
|
|
|
while pos < l:
|
|
|
|
c = program[pos]
|
|
|
|
if c.isspace(): # skip inter-token whitespace
|
|
|
|
pass
|
2010-06-02 23:07:46 +04:00
|
|
|
elif c == ':' and program[pos:pos + 2] == '::': # look ahead carefully
|
2010-06-05 05:57:52 +04:00
|
|
|
yield ('::', None, pos)
|
2010-06-02 23:07:46 +04:00
|
|
|
pos += 1 # skip ahead
|
2010-06-01 20:18:57 +04:00
|
|
|
elif c == '.' and program[pos:pos + 2] == '..': # look ahead carefully
|
2010-06-05 05:57:52 +04:00
|
|
|
yield ('..', None, pos)
|
2010-06-01 20:18:57 +04:00
|
|
|
pos += 1 # skip ahead
|
2011-04-30 19:43:04 +04:00
|
|
|
elif c in "():,-|&+!~^": # handle simple operators
|
2010-06-05 05:57:52 +04:00
|
|
|
yield (c, None, pos)
|
2010-09-25 00:36:53 +04:00
|
|
|
elif (c in '"\'' or c == 'r' and
|
|
|
|
program[pos:pos + 2] in ("r'", 'r"')): # handle quoted strings
|
|
|
|
if c == 'r':
|
|
|
|
pos += 1
|
|
|
|
c = program[pos]
|
|
|
|
decode = lambda x: x
|
|
|
|
else:
|
|
|
|
decode = lambda x: x.decode('string-escape')
|
2010-06-01 20:18:57 +04:00
|
|
|
pos += 1
|
|
|
|
s = pos
|
|
|
|
while pos < l: # find closing quote
|
|
|
|
d = program[pos]
|
|
|
|
if d == '\\': # skip over escaped characters
|
|
|
|
pos += 2
|
|
|
|
continue
|
|
|
|
if d == c:
|
2010-09-25 00:36:53 +04:00
|
|
|
yield ('string', decode(program[s:pos]), s)
|
2010-06-01 20:18:57 +04:00
|
|
|
break
|
|
|
|
pos += 1
|
|
|
|
else:
|
2010-06-18 23:31:19 +04:00
|
|
|
raise error.ParseError(_("unterminated string"), s)
|
2012-05-12 17:54:54 +04:00
|
|
|
# gather up a symbol/keyword
|
2012-10-31 03:48:44 +04:00
|
|
|
elif c.isalnum() or c in '._@' or ord(c) > 127:
|
2010-06-01 20:18:57 +04:00
|
|
|
s = pos
|
|
|
|
pos += 1
|
|
|
|
while pos < l: # find end of symbol
|
|
|
|
d = program[pos]
|
2014-03-19 02:54:42 +04:00
|
|
|
if not (d.isalnum() or d in "-._/@" or ord(d) > 127):
|
2010-06-01 20:18:57 +04:00
|
|
|
break
|
|
|
|
if d == '.' and program[pos - 1] == '.': # special case for ..
|
|
|
|
pos -= 1
|
|
|
|
break
|
|
|
|
pos += 1
|
|
|
|
sym = program[s:pos]
|
|
|
|
if sym in keywords: # operator keywords
|
2010-06-05 05:57:52 +04:00
|
|
|
yield (sym, None, s)
|
2014-03-19 02:54:42 +04:00
|
|
|
elif '-' in sym:
|
|
|
|
# some jerk gave us foo-bar-baz, try to check if it's a symbol
|
|
|
|
if lookup and lookup(sym):
|
|
|
|
# looks like a real symbol
|
|
|
|
yield ('symbol', sym, s)
|
|
|
|
else:
|
|
|
|
# looks like an expression
|
|
|
|
parts = sym.split('-')
|
|
|
|
for p in parts[:-1]:
|
|
|
|
if p: # possible consecutive -
|
|
|
|
yield ('symbol', p, s)
|
|
|
|
s += len(p)
|
|
|
|
yield ('-', None, pos)
|
|
|
|
s += 1
|
|
|
|
if parts[-1]: # possible trailing -
|
|
|
|
yield ('symbol', parts[-1], s)
|
2010-06-01 20:18:57 +04:00
|
|
|
else:
|
2010-06-05 05:57:52 +04:00
|
|
|
yield ('symbol', sym, s)
|
2010-06-01 20:18:57 +04:00
|
|
|
pos -= 1
|
|
|
|
else:
|
2010-06-18 23:31:19 +04:00
|
|
|
raise error.ParseError(_("syntax error"), pos)
|
2010-06-01 20:18:57 +04:00
|
|
|
pos += 1
|
2010-06-05 05:57:52 +04:00
|
|
|
yield ('end', None, pos)
|
2010-06-01 20:18:57 +04:00
|
|
|
|
|
|
|
# helpers
|
|
|
|
|
|
|
|
def getstring(x, err):
|
2010-09-24 21:46:54 +04:00
|
|
|
if x and (x[0] == 'string' or x[0] == 'symbol'):
|
2010-06-01 20:18:57 +04:00
|
|
|
return x[1]
|
2010-06-05 05:57:52 +04:00
|
|
|
raise error.ParseError(err)
|
2010-06-01 20:18:57 +04:00
|
|
|
|
|
|
|
def getlist(x):
|
|
|
|
if not x:
|
|
|
|
return []
|
|
|
|
if x[0] == 'list':
|
|
|
|
return getlist(x[1]) + [x[2]]
|
|
|
|
return [x]
|
|
|
|
|
2010-06-12 00:30:12 +04:00
|
|
|
def getargs(x, min, max, err):
|
2010-06-01 20:18:57 +04:00
|
|
|
l = getlist(x)
|
2012-02-23 21:05:20 +04:00
|
|
|
if len(l) < min or (max >= 0 and len(l) > max):
|
2010-06-05 05:57:52 +04:00
|
|
|
raise error.ParseError(err)
|
2010-06-01 20:18:57 +04:00
|
|
|
return l
|
|
|
|
|
|
|
|
def getset(repo, subset, x):
|
|
|
|
if not x:
|
2010-06-18 23:31:19 +04:00
|
|
|
raise error.ParseError(_("missing argument"))
|
2014-02-19 03:54:46 +04:00
|
|
|
s = methods[x[0]](repo, subset, *x[1:])
|
|
|
|
if util.safehasattr(s, 'set'):
|
|
|
|
return s
|
|
|
|
return baseset(s)
|
2010-06-01 20:18:57 +04:00
|
|
|
|
2012-06-06 04:35:34 +04:00
|
|
|
def _getrevsource(repo, r):
|
|
|
|
extra = repo[r].extra()
|
|
|
|
for label in ('source', 'transplant_source', 'rebase_source'):
|
|
|
|
if label in extra:
|
|
|
|
try:
|
|
|
|
return repo[extra[label]].rev()
|
|
|
|
except error.RepoLookupError:
|
|
|
|
pass
|
|
|
|
return None
|
|
|
|
|
2010-06-01 20:18:57 +04:00
|
|
|
# operator methods
|
|
|
|
|
|
|
|
def stringset(repo, subset, x):
|
|
|
|
x = repo[x].rev()
|
2010-06-04 05:32:41 +04:00
|
|
|
if x == -1 and len(subset) == len(repo):
|
2014-01-21 23:39:26 +04:00
|
|
|
return baseset([-1])
|
2011-04-15 21:07:44 +04:00
|
|
|
if len(subset) == len(repo) or x in subset:
|
2014-01-21 23:39:26 +04:00
|
|
|
return baseset([x])
|
|
|
|
return baseset([])
|
2010-06-01 20:18:57 +04:00
|
|
|
|
|
|
|
def symbolset(repo, subset, x):
|
|
|
|
if x in symbols:
|
2010-06-18 23:31:19 +04:00
|
|
|
raise error.ParseError(_("can't use %s here") % x)
|
2010-06-01 20:18:57 +04:00
|
|
|
return stringset(repo, subset, x)
|
|
|
|
|
|
|
|
def rangeset(repo, subset, x, y):
|
2014-01-21 23:39:26 +04:00
|
|
|
cl = baseset(repo.changelog)
|
revset: evaluate sub expressions correctly (issue3775)
Before this patch, sub expression may return unexpected result, if it
is joined with another expression by "or":
- "^"/parentspec():
"R or R^1" is not equal to "R^1 or R". the former returns only "R".
- "~"/ancestorspec():
"R or R~1" is not equal to "R~1 or R". the former returns only "R".
- ":"/rangeset():
"10 or (10 or 15):" is not equal to "(10 or 15): or 10". the
former returns only 10 and 15 or grater (11 to 14 are not
included).
In "or"-ed expression "A or B", the "subset" passed to evaluation of
"B" doesn't contain revisions gotten from evaluation of "A", for
efficiency.
In the other hand, "stringset()" fails to look corresponding revision
for specified string/symbol up, if "subset" doesn't contain that
revision.
So, predicates looking revisions up indirectly should evaluate sub
expressions of themselves not with passed "subset" but with "entire
revisions in the repository", to prevent "stringset()" from unexpected
failing to look symbols in them up.
But predicates in above example don't so. For example, in the case of
"R or R^1":
1. "R^1" is evaluated with "subset" containing revisions other than
"R", because "R" is already gotten by the former of "or"-ed
expressions
2. "parentspec()" evaluates "R" of "R^1" with such "subset"
3. "stringset()" fails to look "R" up, because "R" is not contained
in "subset"
4. so, evaluation of "R^1" returns no revision
This patch evaluates sub expressions for predicates above with "entire
revisions in the repository".
2013-01-23 17:52:55 +04:00
|
|
|
m = getset(repo, cl, x)
|
|
|
|
n = getset(repo, cl, y)
|
2010-06-28 20:07:27 +04:00
|
|
|
|
|
|
|
if not m or not n:
|
2014-01-21 23:39:26 +04:00
|
|
|
return baseset([])
|
2010-06-28 20:07:27 +04:00
|
|
|
m, n = m[0], n[-1]
|
|
|
|
|
2010-06-01 20:18:57 +04:00
|
|
|
if m < n:
|
2014-02-03 22:15:15 +04:00
|
|
|
r = spanset(repo, m, n + 1)
|
2010-06-28 20:07:27 +04:00
|
|
|
else:
|
2014-02-03 22:15:15 +04:00
|
|
|
r = spanset(repo, m, n - 1)
|
|
|
|
return r & subset
|
2010-06-01 20:18:57 +04:00
|
|
|
|
2012-06-02 02:50:22 +04:00
|
|
|
def dagrange(repo, subset, x, y):
|
2014-02-03 22:15:15 +04:00
|
|
|
r = spanset(repo)
|
2013-04-17 00:29:54 +04:00
|
|
|
xs = _revsbetween(repo, getset(repo, r, x), getset(repo, r, y))
|
2014-01-22 22:46:02 +04:00
|
|
|
s = subset.set()
|
2014-05-02 01:07:04 +04:00
|
|
|
return xs.filter(s.__contains__)
|
2012-06-02 02:50:22 +04:00
|
|
|
|
2010-06-01 20:18:57 +04:00
|
|
|
def andset(repo, subset, x, y):
|
|
|
|
return getset(repo, getset(repo, subset, x), y)
|
|
|
|
|
|
|
|
def orset(repo, subset, x, y):
|
2011-04-13 21:30:41 +04:00
|
|
|
xl = getset(repo, subset, x)
|
2014-01-24 02:20:58 +04:00
|
|
|
yl = getset(repo, subset - xl, y)
|
2014-02-07 02:57:25 +04:00
|
|
|
return xl + yl
|
2010-06-01 20:18:57 +04:00
|
|
|
|
|
|
|
def notset(repo, subset, x):
|
2014-01-24 02:20:58 +04:00
|
|
|
return subset - getset(repo, subset, x)
|
2010-06-01 20:18:57 +04:00
|
|
|
|
|
|
|
def listset(repo, subset, a, b):
|
2010-06-18 23:31:19 +04:00
|
|
|
raise error.ParseError(_("can't use a list in this context"))
|
2010-06-01 20:18:57 +04:00
|
|
|
|
|
|
|
def func(repo, subset, a, b):
|
|
|
|
if a[0] == 'symbol' and a[1] in symbols:
|
|
|
|
return symbols[a[1]](repo, subset, b)
|
2010-06-18 23:31:19 +04:00
|
|
|
raise error.ParseError(_("not a function: %s") % a[1])
|
2010-06-01 20:18:57 +04:00
|
|
|
|
|
|
|
# functions
|
|
|
|
|
2011-04-08 18:47:58 +04:00
|
|
|
def adds(repo, subset, x):
|
|
|
|
"""``adds(pattern)``
|
|
|
|
Changesets that add a file matching pattern.
|
2014-01-17 18:55:11 +04:00
|
|
|
|
|
|
|
The pattern without explicit kind like ``glob:`` is expected to be
|
|
|
|
relative to the current directory and match against a file or a
|
|
|
|
directory.
|
2010-10-23 21:21:51 +04:00
|
|
|
"""
|
2011-04-08 18:47:58 +04:00
|
|
|
# i18n: "adds" is a keyword
|
|
|
|
pat = getstring(x, _("adds requires a pattern"))
|
|
|
|
return checkstatus(repo, subset, pat, 1)
|
2010-06-01 20:18:57 +04:00
|
|
|
|
2011-04-08 18:47:58 +04:00
|
|
|
def ancestor(repo, subset, x):
|
2013-01-29 00:19:21 +04:00
|
|
|
"""``ancestor(*changeset)``
|
2014-04-08 01:17:51 +04:00
|
|
|
A greatest common ancestor of the changesets.
|
2013-01-29 00:19:21 +04:00
|
|
|
|
|
|
|
Accepts 0 or more changesets.
|
|
|
|
Will return empty list when passed no args.
|
|
|
|
Greatest common ancestor of a single changeset is that changeset.
|
2010-10-23 21:21:51 +04:00
|
|
|
"""
|
2011-04-08 18:47:58 +04:00
|
|
|
# i18n: "ancestor" is a keyword
|
2013-01-29 00:19:21 +04:00
|
|
|
l = getlist(x)
|
2014-02-03 22:15:15 +04:00
|
|
|
rl = spanset(repo)
|
2013-01-29 00:19:21 +04:00
|
|
|
anc = None
|
|
|
|
|
|
|
|
# (getset(repo, rl, i) for i in l) generates a list of lists
|
|
|
|
for revs in (getset(repo, rl, i) for i in l):
|
|
|
|
for r in revs:
|
|
|
|
if anc is None:
|
2014-04-08 01:17:51 +04:00
|
|
|
anc = repo[r]
|
2013-01-29 00:19:21 +04:00
|
|
|
else:
|
2014-04-08 01:17:51 +04:00
|
|
|
anc = anc.ancestor(repo[r])
|
2013-01-29 00:19:21 +04:00
|
|
|
|
2014-04-08 01:17:51 +04:00
|
|
|
if anc is not None and anc.rev() in subset:
|
|
|
|
return baseset([anc.rev()])
|
2014-01-21 23:39:26 +04:00
|
|
|
return baseset([])
|
2010-06-01 20:18:57 +04:00
|
|
|
|
2012-04-11 13:25:34 +04:00
|
|
|
def _ancestors(repo, subset, x, followfirst=False):
|
2014-02-03 22:15:15 +04:00
|
|
|
args = getset(repo, spanset(repo), x)
|
2011-04-08 18:47:58 +04:00
|
|
|
if not args:
|
2014-01-21 23:39:26 +04:00
|
|
|
return baseset([])
|
2014-02-07 22:32:02 +04:00
|
|
|
s = _revancestors(repo, args, followfirst)
|
2014-05-02 01:07:04 +04:00
|
|
|
return subset.filter(s.__contains__)
|
2010-06-01 20:18:57 +04:00
|
|
|
|
2012-04-11 13:25:34 +04:00
|
|
|
def ancestors(repo, subset, x):
|
|
|
|
"""``ancestors(set)``
|
|
|
|
Changesets that are ancestors of a changeset in set.
|
|
|
|
"""
|
|
|
|
return _ancestors(repo, subset, x)
|
|
|
|
|
|
|
|
def _firstancestors(repo, subset, x):
|
|
|
|
# ``_firstancestors(set)``
|
|
|
|
# Like ``ancestors(set)`` but follows only the first parents.
|
|
|
|
return _ancestors(repo, subset, x, followfirst=True)
|
|
|
|
|
2011-04-30 19:43:04 +04:00
|
|
|
def ancestorspec(repo, subset, x, n):
|
|
|
|
"""``set~n``
|
2012-05-12 17:54:54 +04:00
|
|
|
Changesets that are the Nth ancestor (first parents only) of a changeset
|
|
|
|
in set.
|
2011-04-30 19:43:04 +04:00
|
|
|
"""
|
|
|
|
try:
|
|
|
|
n = int(n[1])
|
2011-07-12 21:35:03 +04:00
|
|
|
except (TypeError, ValueError):
|
2011-04-30 19:43:04 +04:00
|
|
|
raise error.ParseError(_("~ expects a number"))
|
|
|
|
ps = set()
|
|
|
|
cl = repo.changelog
|
2014-02-07 02:57:25 +04:00
|
|
|
for r in getset(repo, baseset(cl), x):
|
2011-04-30 19:43:04 +04:00
|
|
|
for i in range(n):
|
|
|
|
r = cl.parentrevs(r)[0]
|
|
|
|
ps.add(r)
|
2014-09-17 21:57:47 +04:00
|
|
|
return subset & ps
|
2011-04-30 19:43:04 +04:00
|
|
|
|
2011-04-08 18:47:58 +04:00
|
|
|
def author(repo, subset, x):
|
|
|
|
"""``author(string)``
|
|
|
|
Alias for ``user(string)``.
|
2010-10-23 21:21:51 +04:00
|
|
|
"""
|
2011-04-08 18:47:58 +04:00
|
|
|
# i18n: "author" is a keyword
|
2011-12-25 15:35:16 +04:00
|
|
|
n = encoding.lower(getstring(x, _("author requires a string")))
|
2012-05-31 02:13:58 +04:00
|
|
|
kind, pattern, matcher = _substringmatcher(n)
|
2014-02-06 21:28:41 +04:00
|
|
|
return subset.filter(lambda x: matcher(encoding.lower(repo[x].user())))
|
2010-07-30 05:07:46 +04:00
|
|
|
|
2013-11-16 20:57:08 +04:00
|
|
|
def only(repo, subset, x):
|
|
|
|
"""``only(set, [set])``
|
|
|
|
Changesets that are ancestors of the first set that are not ancestors
|
|
|
|
of any other head in the repo. If a second set is specified, the result
|
|
|
|
is ancestors of the first set that are not ancestors of the second set
|
|
|
|
(i.e. ::<set1> - ::<set2>).
|
|
|
|
"""
|
|
|
|
cl = repo.changelog
|
2014-04-22 17:12:13 +04:00
|
|
|
# i18n: "only" is a keyword
|
2013-11-16 20:57:08 +04:00
|
|
|
args = getargs(x, 1, 2, _('only takes one or two arguments'))
|
|
|
|
include = getset(repo, spanset(repo), args[0]).set()
|
|
|
|
if len(args) == 1:
|
2014-07-19 03:46:56 +04:00
|
|
|
if len(include) == 0:
|
|
|
|
return baseset([])
|
|
|
|
|
2013-11-16 20:57:08 +04:00
|
|
|
descendants = set(_revdescendants(repo, include, False))
|
|
|
|
exclude = [rev for rev in cl.headrevs()
|
|
|
|
if not rev in descendants and not rev in include]
|
|
|
|
else:
|
|
|
|
exclude = getset(repo, spanset(repo), args[1])
|
|
|
|
|
|
|
|
results = set(ancestormod.missingancestors(include, exclude, cl.parentrevs))
|
2014-05-02 01:07:04 +04:00
|
|
|
return lazyset(subset, results.__contains__)
|
2013-11-16 20:57:08 +04:00
|
|
|
|
2011-09-19 00:54:11 +04:00
|
|
|
def bisect(repo, subset, x):
|
|
|
|
"""``bisect(string)``
|
2011-09-24 03:32:50 +04:00
|
|
|
Changesets marked in the specified bisect status:
|
|
|
|
|
|
|
|
- ``good``, ``bad``, ``skip``: csets explicitly marked as good/bad/skip
|
2012-08-16 00:38:42 +04:00
|
|
|
- ``goods``, ``bads`` : csets topologically good/bad
|
2011-09-24 03:32:50 +04:00
|
|
|
- ``range`` : csets taking part in the bisection
|
|
|
|
- ``pruned`` : csets that are goods, bads or skipped
|
|
|
|
- ``untested`` : csets whose fate is yet unknown
|
|
|
|
- ``ignored`` : csets ignored due to DAG topology
|
2012-05-09 02:29:09 +04:00
|
|
|
- ``current`` : the cset currently being bisected
|
2010-10-23 21:21:51 +04:00
|
|
|
"""
|
2012-07-26 08:58:43 +04:00
|
|
|
# i18n: "bisect" is a keyword
|
2011-09-17 02:20:45 +04:00
|
|
|
status = getstring(x, _("bisect requires a string")).lower()
|
2012-04-19 08:27:35 +04:00
|
|
|
state = set(hbisect.get(repo, status))
|
2014-09-17 21:57:57 +04:00
|
|
|
return subset & state
|
2010-06-01 20:18:57 +04:00
|
|
|
|
2011-09-19 00:54:11 +04:00
|
|
|
# Backward-compatibility
|
|
|
|
# - no help entry so that we do not advertise it any more
|
|
|
|
def bisected(repo, subset, x):
|
|
|
|
return bisect(repo, subset, x)
|
|
|
|
|
2011-04-08 18:47:58 +04:00
|
|
|
def bookmark(repo, subset, x):
|
|
|
|
"""``bookmark([name])``
|
|
|
|
The named bookmark or all bookmarks.
|
2012-05-31 02:13:33 +04:00
|
|
|
|
|
|
|
If `name` starts with `re:`, the remainder of the name is treated as
|
|
|
|
a regular expression. To match a bookmark that actually starts with `re:`,
|
|
|
|
use the prefix `literal:`.
|
2010-10-23 21:21:51 +04:00
|
|
|
"""
|
2011-04-08 18:47:58 +04:00
|
|
|
# i18n: "bookmark" is a keyword
|
|
|
|
args = getargs(x, 0, 1, _('bookmark takes one or no arguments'))
|
|
|
|
if args:
|
|
|
|
bm = getstring(args[0],
|
|
|
|
# i18n: "bookmark" is a keyword
|
|
|
|
_('the argument to bookmark must be a string'))
|
2012-05-31 02:13:33 +04:00
|
|
|
kind, pattern, matcher = _stringmatcher(bm)
|
2014-09-17 21:58:25 +04:00
|
|
|
bms = set()
|
2012-05-31 02:13:33 +04:00
|
|
|
if kind == 'literal':
|
2014-08-12 07:45:08 +04:00
|
|
|
bmrev = repo._bookmarks.get(pattern, None)
|
2012-05-31 02:13:33 +04:00
|
|
|
if not bmrev:
|
|
|
|
raise util.Abort(_("bookmark '%s' does not exist") % bm)
|
2014-09-17 21:58:25 +04:00
|
|
|
bms.add(repo[bmrev].rev())
|
2012-05-31 02:13:33 +04:00
|
|
|
else:
|
|
|
|
matchrevs = set()
|
2013-01-28 00:24:37 +04:00
|
|
|
for name, bmrev in repo._bookmarks.iteritems():
|
2012-05-31 02:13:33 +04:00
|
|
|
if matcher(name):
|
|
|
|
matchrevs.add(bmrev)
|
|
|
|
if not matchrevs:
|
|
|
|
raise util.Abort(_("no bookmarks exist that match '%s'")
|
|
|
|
% pattern)
|
|
|
|
for bmrev in matchrevs:
|
2014-09-17 21:58:25 +04:00
|
|
|
bms.add(repo[bmrev].rev())
|
|
|
|
else:
|
|
|
|
bms = set([repo[r].rev()
|
|
|
|
for r in repo._bookmarks.values()])
|
2014-09-18 06:56:59 +04:00
|
|
|
bms -= set([node.nullrev])
|
2014-09-18 06:57:09 +04:00
|
|
|
return subset & bms
|
2010-06-01 20:18:57 +04:00
|
|
|
|
|
|
|
def branch(repo, subset, x):
|
# User Dan Villiom Podlaski Christiansen <danchr@gmail.com>
# Date 1289564504 -3600
# Node ID b75264c15cc888cf38c3c7b8f619801e3c2589c7
# Parent 89b2e5d940f669e590096c6be70eee61c9172fff
revsets: overload the branch() revset to also take a branch name.
This should only change semantics in the specific case of a tag/branch
conflict where the tag wasn't done on the branch with the same
name. Previously, branch(whatever) would resolve to the branch of the
tag in that case, whereas now it will resolve to the branch of the
name. The previous behaviour, while documented, seemed very
counter-intuitive to me.
An alternate approach would be to introduce a new revset such as
branchname() or namedbranch(). While this would retain backwards
compatibility, the distinction between it and branch() would not be
readily apparent to users. The most intuitive behaviour would be to
have branch(x) require 'x' to be a branch name, and something like
branchof(x) or samebranch(x) do what branch(x) currently
does. Unfortunately, our backwards compatibility guarantees prevent us
from doing that.
Please note that while 'hg tag' guards against shadowing a branch, 'hg
branch' does not. Besides, even if it did, that wouldn't solve the
issue of conversions with such tags and branches...
2011-03-24 03:28:16 +03:00
|
|
|
"""``branch(string or set)``
|
|
|
|
All changesets belonging to the given branch or the branches of the given
|
|
|
|
changesets.
|
2012-05-31 02:13:33 +04:00
|
|
|
|
|
|
|
If `string` starts with `re:`, the remainder of the name is treated as
|
|
|
|
a regular expression. To match a branch that actually starts with `re:`,
|
|
|
|
use the prefix `literal:`.
|
2010-10-23 21:21:51 +04:00
|
|
|
"""
|
# User Dan Villiom Podlaski Christiansen <danchr@gmail.com>
# Date 1289564504 -3600
# Node ID b75264c15cc888cf38c3c7b8f619801e3c2589c7
# Parent 89b2e5d940f669e590096c6be70eee61c9172fff
revsets: overload the branch() revset to also take a branch name.
This should only change semantics in the specific case of a tag/branch
conflict where the tag wasn't done on the branch with the same
name. Previously, branch(whatever) would resolve to the branch of the
tag in that case, whereas now it will resolve to the branch of the
name. The previous behaviour, while documented, seemed very
counter-intuitive to me.
An alternate approach would be to introduce a new revset such as
branchname() or namedbranch(). While this would retain backwards
compatibility, the distinction between it and branch() would not be
readily apparent to users. The most intuitive behaviour would be to
have branch(x) require 'x' to be a branch name, and something like
branchof(x) or samebranch(x) do what branch(x) currently
does. Unfortunately, our backwards compatibility guarantees prevent us
from doing that.
Please note that while 'hg tag' guards against shadowing a branch, 'hg
branch' does not. Besides, even if it did, that wouldn't solve the
issue of conversions with such tags and branches...
2011-03-24 03:28:16 +03:00
|
|
|
try:
|
|
|
|
b = getstring(x, '')
|
|
|
|
except error.ParseError:
|
|
|
|
# not a string, but another revspec, e.g. tip()
|
|
|
|
pass
|
2012-05-31 02:13:33 +04:00
|
|
|
else:
|
|
|
|
kind, pattern, matcher = _stringmatcher(b)
|
|
|
|
if kind == 'literal':
|
|
|
|
# note: falls through to the revspec case if no branch with
|
|
|
|
# this name exists
|
|
|
|
if pattern in repo.branchmap():
|
2014-02-06 21:28:41 +04:00
|
|
|
return subset.filter(lambda r: matcher(repo[r].branch()))
|
2012-05-31 02:13:33 +04:00
|
|
|
else:
|
2014-02-06 21:28:41 +04:00
|
|
|
return subset.filter(lambda r: matcher(repo[r].branch()))
|
# User Dan Villiom Podlaski Christiansen <danchr@gmail.com>
# Date 1289564504 -3600
# Node ID b75264c15cc888cf38c3c7b8f619801e3c2589c7
# Parent 89b2e5d940f669e590096c6be70eee61c9172fff
revsets: overload the branch() revset to also take a branch name.
This should only change semantics in the specific case of a tag/branch
conflict where the tag wasn't done on the branch with the same
name. Previously, branch(whatever) would resolve to the branch of the
tag in that case, whereas now it will resolve to the branch of the
name. The previous behaviour, while documented, seemed very
counter-intuitive to me.
An alternate approach would be to introduce a new revset such as
branchname() or namedbranch(). While this would retain backwards
compatibility, the distinction between it and branch() would not be
readily apparent to users. The most intuitive behaviour would be to
have branch(x) require 'x' to be a branch name, and something like
branchof(x) or samebranch(x) do what branch(x) currently
does. Unfortunately, our backwards compatibility guarantees prevent us
from doing that.
Please note that while 'hg tag' guards against shadowing a branch, 'hg
branch' does not. Besides, even if it did, that wouldn't solve the
issue of conversions with such tags and branches...
2011-03-24 03:28:16 +03:00
|
|
|
|
2014-02-03 22:15:15 +04:00
|
|
|
s = getset(repo, spanset(repo), x)
|
2010-06-01 20:18:57 +04:00
|
|
|
b = set()
|
|
|
|
for r in s:
|
|
|
|
b.add(repo[r].branch())
|
2014-01-22 22:46:02 +04:00
|
|
|
s = s.set()
|
2014-02-06 21:28:41 +04:00
|
|
|
return subset.filter(lambda r: r in s or repo[r].branch() in b)
|
2010-06-01 20:18:57 +04:00
|
|
|
|
2012-10-19 02:39:06 +04:00
|
|
|
def bumped(repo, subset, x):
|
|
|
|
"""``bumped()``
|
|
|
|
Mutable changesets marked as successors of public changesets.
|
|
|
|
|
|
|
|
Only non-public and non-obsolete changesets can be `bumped`.
|
|
|
|
"""
|
|
|
|
# i18n: "bumped" is a keyword
|
|
|
|
getargs(x, 0, 0, _("bumped takes no arguments"))
|
|
|
|
bumped = obsmod.getrevs(repo, 'bumped')
|
2014-01-25 04:57:44 +04:00
|
|
|
return subset & bumped
|
2012-10-19 02:39:06 +04:00
|
|
|
|
2012-11-01 03:23:23 +04:00
|
|
|
def bundle(repo, subset, x):
|
|
|
|
"""``bundle()``
|
|
|
|
Changesets in the bundle.
|
|
|
|
|
|
|
|
Bundle must be specified by the -R option."""
|
|
|
|
|
|
|
|
try:
|
2013-01-16 23:41:34 +04:00
|
|
|
bundlerevs = repo.changelog.bundlerevs
|
2012-11-01 03:23:23 +04:00
|
|
|
except AttributeError:
|
|
|
|
raise util.Abort(_("no bundle provided - specify with -R"))
|
2014-01-25 04:57:44 +04:00
|
|
|
return subset & bundlerevs
|
2012-11-01 03:23:23 +04:00
|
|
|
|
2011-04-08 18:47:58 +04:00
|
|
|
def checkstatus(repo, subset, pat, field):
|
2012-04-26 16:24:46 +04:00
|
|
|
hasset = matchmod.patkind(pat) == 'set'
|
2014-01-31 22:47:51 +04:00
|
|
|
|
|
|
|
def matches(x):
|
|
|
|
m = None
|
|
|
|
fname = None
|
|
|
|
c = repo[x]
|
2012-04-26 16:24:46 +04:00
|
|
|
if not m or hasset:
|
|
|
|
m = matchmod.match(repo.root, repo.getcwd(), [pat], ctx=c)
|
|
|
|
if not m.anypats() and len(m.files()) == 1:
|
|
|
|
fname = m.files()[0]
|
|
|
|
if fname is not None:
|
|
|
|
if fname not in c.files():
|
2014-01-31 22:47:51 +04:00
|
|
|
return False
|
2011-04-08 18:47:58 +04:00
|
|
|
else:
|
|
|
|
for f in c.files():
|
|
|
|
if m(f):
|
|
|
|
break
|
|
|
|
else:
|
2014-01-31 22:47:51 +04:00
|
|
|
return False
|
2011-04-08 18:47:58 +04:00
|
|
|
files = repo.status(c.p1().node(), c.node())[field]
|
2012-04-26 16:24:46 +04:00
|
|
|
if fname is not None:
|
|
|
|
if fname in files:
|
2014-01-31 22:47:51 +04:00
|
|
|
return True
|
2011-04-08 18:47:58 +04:00
|
|
|
else:
|
|
|
|
for f in files:
|
|
|
|
if m(f):
|
2014-01-31 22:47:51 +04:00
|
|
|
return True
|
|
|
|
|
2014-02-06 21:28:41 +04:00
|
|
|
return subset.filter(matches)
|
2011-04-08 18:47:58 +04:00
|
|
|
|
2012-04-08 13:14:56 +04:00
|
|
|
def _children(repo, narrow, parentset):
|
2014-03-13 23:34:32 +04:00
|
|
|
cs = set()
|
2012-12-07 22:37:43 +04:00
|
|
|
if not parentset:
|
2014-03-13 23:34:32 +04:00
|
|
|
return baseset(cs)
|
2012-01-16 11:21:30 +04:00
|
|
|
pr = repo.changelog.parentrevs
|
2012-12-07 22:37:43 +04:00
|
|
|
minrev = min(parentset)
|
2014-03-13 23:34:32 +04:00
|
|
|
for r in narrow:
|
|
|
|
if r <= minrev:
|
|
|
|
continue
|
|
|
|
for p in pr(r):
|
2012-04-08 13:14:56 +04:00
|
|
|
if p in parentset:
|
2014-03-13 23:34:32 +04:00
|
|
|
cs.add(r)
|
|
|
|
return baseset(cs)
|
2012-01-16 11:21:30 +04:00
|
|
|
|
2011-04-08 18:47:58 +04:00
|
|
|
def children(repo, subset, x):
|
|
|
|
"""``children(set)``
|
|
|
|
Child changesets of changesets in set.
|
2010-10-23 21:21:51 +04:00
|
|
|
"""
|
2014-01-22 22:46:02 +04:00
|
|
|
s = getset(repo, baseset(repo), x).set()
|
2012-01-16 11:21:30 +04:00
|
|
|
cs = _children(repo, subset, s)
|
2014-01-25 04:57:44 +04:00
|
|
|
return subset & cs
|
2010-07-22 17:17:38 +04:00
|
|
|
|
2011-04-08 18:47:58 +04:00
|
|
|
def closed(repo, subset, x):
|
|
|
|
"""``closed()``
|
|
|
|
Changeset is closed.
|
|
|
|
"""
|
|
|
|
# i18n: "closed" is a keyword
|
|
|
|
getargs(x, 0, 0, _("closed takes no arguments"))
|
2014-02-06 21:28:41 +04:00
|
|
|
return subset.filter(lambda r: repo[r].closesbranch())
|
2010-06-01 20:18:57 +04:00
|
|
|
|
2011-04-08 18:47:58 +04:00
|
|
|
def contains(repo, subset, x):
|
|
|
|
"""``contains(pattern)``
|
2014-04-29 02:09:23 +04:00
|
|
|
The revision's manifest contains a file matching pattern (but might not
|
|
|
|
modify it). See :hg:`help patterns` for information about file patterns.
|
2014-01-17 18:55:11 +04:00
|
|
|
|
|
|
|
The pattern without explicit kind like ``glob:`` is expected to be
|
|
|
|
relative to the current directory and match against a file exactly
|
|
|
|
for efficiency.
|
2010-10-23 21:21:51 +04:00
|
|
|
"""
|
2011-04-08 18:47:58 +04:00
|
|
|
# i18n: "contains" is a keyword
|
|
|
|
pat = getstring(x, _("contains requires a pattern"))
|
2014-02-05 03:07:03 +04:00
|
|
|
|
|
|
|
def matches(x):
|
|
|
|
if not matchmod.patkind(pat):
|
|
|
|
pats = pathutil.canonpath(repo.root, repo.getcwd(), pat)
|
|
|
|
if pats in repo[x]:
|
|
|
|
return True
|
|
|
|
else:
|
|
|
|
c = repo[x]
|
|
|
|
m = matchmod.match(repo.root, repo.getcwd(), [pat], ctx=c)
|
2012-01-21 09:05:04 +04:00
|
|
|
for f in c.manifest():
|
2011-04-08 18:47:58 +04:00
|
|
|
if m(f):
|
2014-02-05 03:07:03 +04:00
|
|
|
return True
|
|
|
|
return False
|
|
|
|
|
2014-02-06 21:28:41 +04:00
|
|
|
return subset.filter(matches)
|
2011-04-08 18:47:58 +04:00
|
|
|
|
2012-05-13 09:12:26 +04:00
|
|
|
def converted(repo, subset, x):
|
|
|
|
"""``converted([id])``
|
|
|
|
Changesets converted from the given identifier in the old repository if
|
|
|
|
present, or all converted changesets if no identifier is specified.
|
|
|
|
"""
|
|
|
|
|
|
|
|
# There is exactly no chance of resolving the revision, so do a simple
|
|
|
|
# string compare and hope for the best
|
|
|
|
|
|
|
|
rev = None
|
2012-07-26 08:58:43 +04:00
|
|
|
# i18n: "converted" is a keyword
|
2012-05-13 09:12:26 +04:00
|
|
|
l = getargs(x, 0, 1, _('converted takes one or no arguments'))
|
|
|
|
if l:
|
2012-07-26 08:58:43 +04:00
|
|
|
# i18n: "converted" is a keyword
|
2012-05-13 09:12:26 +04:00
|
|
|
rev = getstring(l[0], _('converted requires a revision'))
|
|
|
|
|
|
|
|
def _matchvalue(r):
|
|
|
|
source = repo[r].extra().get('convert_revision', None)
|
|
|
|
return source is not None and (rev is None or source.startswith(rev))
|
|
|
|
|
2014-02-06 21:28:41 +04:00
|
|
|
return subset.filter(lambda r: _matchvalue(r))
|
2012-05-13 09:12:26 +04:00
|
|
|
|
2011-04-08 18:47:58 +04:00
|
|
|
def date(repo, subset, x):
|
|
|
|
"""``date(interval)``
|
|
|
|
Changesets within the interval, see :hg:`help dates`.
|
|
|
|
"""
|
|
|
|
# i18n: "date" is a keyword
|
|
|
|
ds = getstring(x, _("date requires a string"))
|
|
|
|
dm = util.matchdate(ds)
|
2014-02-06 21:28:41 +04:00
|
|
|
return subset.filter(lambda x: dm(repo[x].date()[0]))
|
2010-06-01 20:18:57 +04:00
|
|
|
|
2011-06-17 00:47:34 +04:00
|
|
|
def desc(repo, subset, x):
|
|
|
|
"""``desc(string)``
|
|
|
|
Search commit message for string. The match is case-insensitive.
|
|
|
|
"""
|
|
|
|
# i18n: "desc" is a keyword
|
2011-12-25 15:35:16 +04:00
|
|
|
ds = encoding.lower(getstring(x, _("desc requires a string")))
|
2014-01-31 03:39:56 +04:00
|
|
|
|
|
|
|
def matches(x):
|
|
|
|
c = repo[x]
|
|
|
|
return ds in encoding.lower(c.description())
|
|
|
|
|
2014-02-06 21:28:41 +04:00
|
|
|
return subset.filter(matches)
|
2011-06-17 00:47:34 +04:00
|
|
|
|
2012-04-11 13:25:34 +04:00
|
|
|
def _descendants(repo, subset, x, followfirst=False):
|
2014-02-03 22:15:15 +04:00
|
|
|
args = getset(repo, spanset(repo), x)
|
2010-06-28 20:07:27 +04:00
|
|
|
if not args:
|
2014-01-21 23:39:26 +04:00
|
|
|
return baseset([])
|
2014-02-11 00:26:45 +04:00
|
|
|
s = _revdescendants(repo, args, followfirst)
|
2014-03-26 01:10:01 +04:00
|
|
|
|
|
|
|
# Both sets need to be ascending in order to lazily return the union
|
|
|
|
# in the correct order.
|
|
|
|
args.ascending()
|
2014-09-13 03:21:13 +04:00
|
|
|
result = (orderedlazyset(s, subset.__contains__, ascending=True) +
|
|
|
|
orderedlazyset(args, subset.__contains__, ascending=True))
|
2014-03-26 01:10:01 +04:00
|
|
|
|
|
|
|
# Wrap result in a lazyset since it's an _addset, which doesn't implement
|
|
|
|
# all the necessary functions to be consumed by callers.
|
|
|
|
return orderedlazyset(result, lambda r: True, ascending=True)
|
2010-06-01 20:18:57 +04:00
|
|
|
|
2012-04-11 13:25:34 +04:00
|
|
|
def descendants(repo, subset, x):
|
|
|
|
"""``descendants(set)``
|
|
|
|
Changesets which are descendants of changesets in set.
|
|
|
|
"""
|
|
|
|
return _descendants(repo, subset, x)
|
|
|
|
|
|
|
|
def _firstdescendants(repo, subset, x):
|
|
|
|
# ``_firstdescendants(set)``
|
|
|
|
# Like ``descendants(set)`` but follows only the first parents.
|
|
|
|
return _descendants(repo, subset, x, followfirst=True)
|
|
|
|
|
revset: add destination() predicate
This predicate is used to find csets that were created because of a graft,
transplant or rebase --keep. An optional revset can be supplied, in which case
the result will be limited to those copies which specified one of the revs as
the source for the command.
hg log -r destination() # csets copied from anywhere
hg log -r destination(branch(default)) # all csets copied from default
hg log -r origin(x) or destination(origin(x)) # all instances of x
This predicate will follow a cset through different types of copies. Given a
repo with a cset 'S' that is grafted to create G(S), which itself is
transplanted to become T(G(S)):
o-S
/
o-o-G(S)
\
o-T(G(S))
hg log -r destination( S ) # { G(S), T(G(S)) }
hg log -r destination( G(S) ) # { T(G(S)) }
The implementation differences between the three different copy commands (see
the origin() predicate) are not intentionally exposed, however if the
transplant was a graft instead:
hg log -r destination( G(S) ) # {}
because the 'extra' field in G(G(S)) is S, not G(S). The implementation cannot
correct this by following sources before G(S) and then select the csets that
reference those sources because the cset provided to the predicate would also
end up selected. If there were more than two copies, sources of the argument
would also get selected.
Note that the convert extension does not currently update the 'extra' map in its
destination csets, and therefore copies made prior to the convert will be
missing from the resulting set.
Instead of the loop over 'subset', the following almost works, but does not
select a transplant of a transplant. That is, 'destination(S)' will only
select T(S).
dests = set([r for r in subset if _getrevsource(repo, r) in args])
2012-07-07 08:47:55 +04:00
|
|
|
def destination(repo, subset, x):
|
|
|
|
"""``destination([set])``
|
|
|
|
Changesets that were created by a graft, transplant or rebase operation,
|
|
|
|
with the given revisions specified as the source. Omitting the optional set
|
|
|
|
is the same as passing all().
|
|
|
|
"""
|
|
|
|
if x is not None:
|
2014-02-03 22:15:15 +04:00
|
|
|
args = getset(repo, spanset(repo), x).set()
|
revset: add destination() predicate
This predicate is used to find csets that were created because of a graft,
transplant or rebase --keep. An optional revset can be supplied, in which case
the result will be limited to those copies which specified one of the revs as
the source for the command.
hg log -r destination() # csets copied from anywhere
hg log -r destination(branch(default)) # all csets copied from default
hg log -r origin(x) or destination(origin(x)) # all instances of x
This predicate will follow a cset through different types of copies. Given a
repo with a cset 'S' that is grafted to create G(S), which itself is
transplanted to become T(G(S)):
o-S
/
o-o-G(S)
\
o-T(G(S))
hg log -r destination( S ) # { G(S), T(G(S)) }
hg log -r destination( G(S) ) # { T(G(S)) }
The implementation differences between the three different copy commands (see
the origin() predicate) are not intentionally exposed, however if the
transplant was a graft instead:
hg log -r destination( G(S) ) # {}
because the 'extra' field in G(G(S)) is S, not G(S). The implementation cannot
correct this by following sources before G(S) and then select the csets that
reference those sources because the cset provided to the predicate would also
end up selected. If there were more than two copies, sources of the argument
would also get selected.
Note that the convert extension does not currently update the 'extra' map in its
destination csets, and therefore copies made prior to the convert will be
missing from the resulting set.
Instead of the loop over 'subset', the following almost works, but does not
select a transplant of a transplant. That is, 'destination(S)' will only
select T(S).
dests = set([r for r in subset if _getrevsource(repo, r) in args])
2012-07-07 08:47:55 +04:00
|
|
|
else:
|
2014-02-03 22:15:15 +04:00
|
|
|
args = getall(repo, spanset(repo), x).set()
|
revset: add destination() predicate
This predicate is used to find csets that were created because of a graft,
transplant or rebase --keep. An optional revset can be supplied, in which case
the result will be limited to those copies which specified one of the revs as
the source for the command.
hg log -r destination() # csets copied from anywhere
hg log -r destination(branch(default)) # all csets copied from default
hg log -r origin(x) or destination(origin(x)) # all instances of x
This predicate will follow a cset through different types of copies. Given a
repo with a cset 'S' that is grafted to create G(S), which itself is
transplanted to become T(G(S)):
o-S
/
o-o-G(S)
\
o-T(G(S))
hg log -r destination( S ) # { G(S), T(G(S)) }
hg log -r destination( G(S) ) # { T(G(S)) }
The implementation differences between the three different copy commands (see
the origin() predicate) are not intentionally exposed, however if the
transplant was a graft instead:
hg log -r destination( G(S) ) # {}
because the 'extra' field in G(G(S)) is S, not G(S). The implementation cannot
correct this by following sources before G(S) and then select the csets that
reference those sources because the cset provided to the predicate would also
end up selected. If there were more than two copies, sources of the argument
would also get selected.
Note that the convert extension does not currently update the 'extra' map in its
destination csets, and therefore copies made prior to the convert will be
missing from the resulting set.
Instead of the loop over 'subset', the following almost works, but does not
select a transplant of a transplant. That is, 'destination(S)' will only
select T(S).
dests = set([r for r in subset if _getrevsource(repo, r) in args])
2012-07-07 08:47:55 +04:00
|
|
|
|
|
|
|
dests = set()
|
|
|
|
|
|
|
|
# subset contains all of the possible destinations that can be returned, so
|
|
|
|
# iterate over them and see if their source(s) were provided in the args.
|
|
|
|
# Even if the immediate src of r is not in the args, src's source (or
|
|
|
|
# further back) may be. Scanning back further than the immediate src allows
|
|
|
|
# transitive transplants and rebases to yield the same results as transitive
|
|
|
|
# grafts.
|
|
|
|
for r in subset:
|
|
|
|
src = _getrevsource(repo, r)
|
|
|
|
lineage = None
|
|
|
|
|
|
|
|
while src is not None:
|
|
|
|
if lineage is None:
|
|
|
|
lineage = list()
|
|
|
|
|
|
|
|
lineage.append(r)
|
|
|
|
|
|
|
|
# The visited lineage is a match if the current source is in the arg
|
|
|
|
# set. Since every candidate dest is visited by way of iterating
|
2012-08-16 00:38:42 +04:00
|
|
|
# subset, any dests further back in the lineage will be tested by a
|
revset: add destination() predicate
This predicate is used to find csets that were created because of a graft,
transplant or rebase --keep. An optional revset can be supplied, in which case
the result will be limited to those copies which specified one of the revs as
the source for the command.
hg log -r destination() # csets copied from anywhere
hg log -r destination(branch(default)) # all csets copied from default
hg log -r origin(x) or destination(origin(x)) # all instances of x
This predicate will follow a cset through different types of copies. Given a
repo with a cset 'S' that is grafted to create G(S), which itself is
transplanted to become T(G(S)):
o-S
/
o-o-G(S)
\
o-T(G(S))
hg log -r destination( S ) # { G(S), T(G(S)) }
hg log -r destination( G(S) ) # { T(G(S)) }
The implementation differences between the three different copy commands (see
the origin() predicate) are not intentionally exposed, however if the
transplant was a graft instead:
hg log -r destination( G(S) ) # {}
because the 'extra' field in G(G(S)) is S, not G(S). The implementation cannot
correct this by following sources before G(S) and then select the csets that
reference those sources because the cset provided to the predicate would also
end up selected. If there were more than two copies, sources of the argument
would also get selected.
Note that the convert extension does not currently update the 'extra' map in its
destination csets, and therefore copies made prior to the convert will be
missing from the resulting set.
Instead of the loop over 'subset', the following almost works, but does not
select a transplant of a transplant. That is, 'destination(S)' will only
select T(S).
dests = set([r for r in subset if _getrevsource(repo, r) in args])
2012-07-07 08:47:55 +04:00
|
|
|
# different iteration over subset. Likewise, if the src was already
|
|
|
|
# selected, the current lineage can be selected without going back
|
|
|
|
# further.
|
|
|
|
if src in args or src in dests:
|
|
|
|
dests.update(lineage)
|
|
|
|
break
|
|
|
|
|
|
|
|
r = src
|
|
|
|
src = _getrevsource(repo, r)
|
|
|
|
|
2014-05-02 01:07:04 +04:00
|
|
|
return subset.filter(dests.__contains__)
|
revset: add destination() predicate
This predicate is used to find csets that were created because of a graft,
transplant or rebase --keep. An optional revset can be supplied, in which case
the result will be limited to those copies which specified one of the revs as
the source for the command.
hg log -r destination() # csets copied from anywhere
hg log -r destination(branch(default)) # all csets copied from default
hg log -r origin(x) or destination(origin(x)) # all instances of x
This predicate will follow a cset through different types of copies. Given a
repo with a cset 'S' that is grafted to create G(S), which itself is
transplanted to become T(G(S)):
o-S
/
o-o-G(S)
\
o-T(G(S))
hg log -r destination( S ) # { G(S), T(G(S)) }
hg log -r destination( G(S) ) # { T(G(S)) }
The implementation differences between the three different copy commands (see
the origin() predicate) are not intentionally exposed, however if the
transplant was a graft instead:
hg log -r destination( G(S) ) # {}
because the 'extra' field in G(G(S)) is S, not G(S). The implementation cannot
correct this by following sources before G(S) and then select the csets that
reference those sources because the cset provided to the predicate would also
end up selected. If there were more than two copies, sources of the argument
would also get selected.
Note that the convert extension does not currently update the 'extra' map in its
destination csets, and therefore copies made prior to the convert will be
missing from the resulting set.
Instead of the loop over 'subset', the following almost works, but does not
select a transplant of a transplant. That is, 'destination(S)' will only
select T(S).
dests = set([r for r in subset if _getrevsource(repo, r) in args])
2012-07-07 08:47:55 +04:00
|
|
|
|
2012-12-12 06:12:55 +04:00
|
|
|
def divergent(repo, subset, x):
|
|
|
|
"""``divergent()``
|
|
|
|
Final successors of changesets with an alternative set of final successors.
|
|
|
|
"""
|
|
|
|
# i18n: "divergent" is a keyword
|
|
|
|
getargs(x, 0, 0, _("divergent takes no arguments"))
|
|
|
|
divergent = obsmod.getrevs(repo, 'divergent')
|
2014-09-17 21:58:39 +04:00
|
|
|
return subset & divergent
|
2012-12-12 06:12:55 +04:00
|
|
|
|
2012-01-06 13:04:20 +04:00
|
|
|
def draft(repo, subset, x):
|
|
|
|
"""``draft()``
|
|
|
|
Changeset in draft phase."""
|
2012-07-26 08:58:43 +04:00
|
|
|
# i18n: "draft" is a keyword
|
2012-01-06 13:04:20 +04:00
|
|
|
getargs(x, 0, 0, _("draft takes no arguments"))
|
2012-05-12 02:24:07 +04:00
|
|
|
pc = repo._phasecache
|
2014-02-06 21:28:41 +04:00
|
|
|
return subset.filter(lambda r: pc.phase(repo, r) == phases.draft)
|
2012-01-06 13:04:20 +04:00
|
|
|
|
2012-07-06 21:34:09 +04:00
|
|
|
def extinct(repo, subset, x):
|
|
|
|
"""``extinct()``
|
2012-07-30 17:48:04 +04:00
|
|
|
Obsolete changesets with obsolete descendants only.
|
|
|
|
"""
|
2012-07-26 08:58:43 +04:00
|
|
|
# i18n: "extinct" is a keyword
|
2012-07-26 08:58:43 +04:00
|
|
|
getargs(x, 0, 0, _("extinct takes no arguments"))
|
2012-10-19 02:28:13 +04:00
|
|
|
extincts = obsmod.getrevs(repo, 'extinct')
|
2014-01-25 04:57:44 +04:00
|
|
|
return subset & extincts
|
2012-07-06 21:34:09 +04:00
|
|
|
|
2012-05-12 12:20:57 +04:00
|
|
|
def extra(repo, subset, x):
|
|
|
|
"""``extra(label, [value])``
|
|
|
|
Changesets with the given label in the extra metadata, with the given
|
2012-05-31 02:14:04 +04:00
|
|
|
optional value.
|
|
|
|
|
|
|
|
If `value` starts with `re:`, the remainder of the value is treated as
|
|
|
|
a regular expression. To match a value that actually starts with `re:`,
|
|
|
|
use the prefix `literal:`.
|
|
|
|
"""
|
2012-05-12 12:20:57 +04:00
|
|
|
|
2012-07-26 08:58:43 +04:00
|
|
|
# i18n: "extra" is a keyword
|
2012-05-12 12:20:57 +04:00
|
|
|
l = getargs(x, 1, 2, _('extra takes at least 1 and at most 2 arguments'))
|
2012-07-26 08:58:43 +04:00
|
|
|
# i18n: "extra" is a keyword
|
2012-05-12 12:20:57 +04:00
|
|
|
label = getstring(l[0], _('first argument to extra must be a string'))
|
|
|
|
value = None
|
|
|
|
|
|
|
|
if len(l) > 1:
|
2012-07-26 08:58:43 +04:00
|
|
|
# i18n: "extra" is a keyword
|
2012-05-12 12:20:57 +04:00
|
|
|
value = getstring(l[1], _('second argument to extra must be a string'))
|
2012-05-31 02:14:04 +04:00
|
|
|
kind, value, matcher = _stringmatcher(value)
|
2012-05-12 12:20:57 +04:00
|
|
|
|
|
|
|
def _matchvalue(r):
|
|
|
|
extra = repo[r].extra()
|
2012-05-31 02:14:04 +04:00
|
|
|
return label in extra and (value is None or matcher(extra[label]))
|
2012-05-12 12:20:57 +04:00
|
|
|
|
2014-02-06 21:28:41 +04:00
|
|
|
return subset.filter(lambda r: _matchvalue(r))
|
2012-05-12 12:20:57 +04:00
|
|
|
|
2011-05-17 02:02:35 +04:00
|
|
|
def filelog(repo, subset, x):
|
|
|
|
"""``filelog(pattern)``
|
|
|
|
Changesets connected to the specified filelog.
|
2012-07-25 11:15:28 +04:00
|
|
|
|
2014-04-29 02:09:23 +04:00
|
|
|
For performance reasons, visits only revisions mentioned in the file-level
|
|
|
|
filelog, rather than filtering through all changesets (much faster, but
|
|
|
|
doesn't include deletes or duplicate changes). For a slower, more accurate
|
|
|
|
result, use ``file()``.
|
2014-01-17 18:55:11 +04:00
|
|
|
|
|
|
|
The pattern without explicit kind like ``glob:`` is expected to be
|
|
|
|
relative to the current directory and match against a file exactly
|
|
|
|
for efficiency.
|
2011-05-17 02:02:35 +04:00
|
|
|
"""
|
|
|
|
|
2012-07-26 08:58:43 +04:00
|
|
|
# i18n: "filelog" is a keyword
|
2011-05-17 02:02:35 +04:00
|
|
|
pat = getstring(x, _("filelog requires a pattern"))
|
|
|
|
s = set()
|
|
|
|
|
2012-01-21 09:05:04 +04:00
|
|
|
if not matchmod.patkind(pat):
|
2014-01-17 18:55:03 +04:00
|
|
|
f = pathutil.canonpath(repo.root, repo.getcwd(), pat)
|
2014-01-17 18:42:12 +04:00
|
|
|
fl = repo.file(f)
|
|
|
|
for fr in fl:
|
|
|
|
s.add(fl.linkrev(fr))
|
2011-05-17 02:02:35 +04:00
|
|
|
else:
|
2014-01-17 18:55:03 +04:00
|
|
|
m = matchmod.match(repo.root, repo.getcwd(), [pat], ctx=repo[None])
|
2011-05-17 02:02:35 +04:00
|
|
|
for f in repo[None]:
|
|
|
|
if m(f):
|
|
|
|
fl = repo.file(f)
|
|
|
|
for fr in fl:
|
|
|
|
s.add(fl.linkrev(fr))
|
|
|
|
|
2014-09-17 21:58:50 +04:00
|
|
|
return subset & s
|
2011-05-17 02:02:35 +04:00
|
|
|
|
2011-09-17 21:34:47 +04:00
|
|
|
def first(repo, subset, x):
|
|
|
|
"""``first(set, [n])``
|
|
|
|
An alias for limit().
|
|
|
|
"""
|
|
|
|
return limit(repo, subset, x)
|
|
|
|
|
2012-02-26 20:10:57 +04:00
|
|
|
def _follow(repo, subset, x, name, followfirst=False):
|
|
|
|
l = getargs(x, 0, 1, _("%s takes no arguments or a filename") % name)
|
2012-01-21 09:52:31 +04:00
|
|
|
c = repo['.']
|
2011-05-17 02:02:35 +04:00
|
|
|
if l:
|
2012-02-26 20:10:57 +04:00
|
|
|
x = getstring(l[0], _("%s expected a filename") % name)
|
2012-01-21 09:52:31 +04:00
|
|
|
if x in c:
|
|
|
|
cx = c[x]
|
2012-02-26 20:10:57 +04:00
|
|
|
s = set(ctx.rev() for ctx in cx.ancestors(followfirst=followfirst))
|
2012-01-21 09:52:31 +04:00
|
|
|
# include the revision responsible for the most recent version
|
|
|
|
s.add(cx.linkrev())
|
2011-11-21 01:29:55 +04:00
|
|
|
else:
|
2014-01-21 23:39:26 +04:00
|
|
|
return baseset([])
|
2011-05-17 02:02:35 +04:00
|
|
|
else:
|
2014-02-07 22:32:02 +04:00
|
|
|
s = _revancestors(repo, baseset([c.rev()]), followfirst)
|
2012-02-26 01:11:36 +04:00
|
|
|
|
2014-09-17 21:59:16 +04:00
|
|
|
return subset & s
|
2012-02-26 01:11:36 +04:00
|
|
|
|
2012-02-26 20:10:57 +04:00
|
|
|
def follow(repo, subset, x):
|
|
|
|
"""``follow([file])``
|
|
|
|
An alias for ``::.`` (ancestors of the working copy's first parent).
|
|
|
|
If a filename is specified, the history of the given file is followed,
|
|
|
|
including copies.
|
|
|
|
"""
|
|
|
|
return _follow(repo, subset, x, 'follow')
|
|
|
|
|
|
|
|
def _followfirst(repo, subset, x):
|
|
|
|
# ``followfirst([file])``
|
|
|
|
# Like ``follow([file])`` but follows only the first parent of
|
|
|
|
# every revision or file revision.
|
|
|
|
return _follow(repo, subset, x, '_followfirst', followfirst=True)
|
|
|
|
|
2011-04-08 18:47:58 +04:00
|
|
|
def getall(repo, subset, x):
|
|
|
|
"""``all()``
|
|
|
|
All changesets, the same as ``0:tip``.
|
2010-10-23 21:21:51 +04:00
|
|
|
"""
|
2011-04-08 18:47:58 +04:00
|
|
|
# i18n: "all" is a keyword
|
|
|
|
getargs(x, 0, 0, _("all takes no arguments"))
|
|
|
|
return subset
|
2010-06-01 20:18:57 +04:00
|
|
|
|
|
|
|
def grep(repo, subset, x):
|
2010-10-23 21:21:51 +04:00
|
|
|
"""``grep(regex)``
|
|
|
|
Like ``keyword(string)`` but accepts a regex. Use ``grep(r'...')``
|
2011-05-18 11:28:11 +04:00
|
|
|
to ensure special escape characters are handled correctly. Unlike
|
|
|
|
``keyword(string)``, the match is case-sensitive.
|
2010-10-23 21:21:51 +04:00
|
|
|
"""
|
2010-09-17 19:21:02 +04:00
|
|
|
try:
|
2010-10-23 16:59:19 +04:00
|
|
|
# i18n: "grep" is a keyword
|
2010-10-16 20:50:53 +04:00
|
|
|
gr = re.compile(getstring(x, _("grep requires a string")))
|
2010-09-17 19:21:02 +04:00
|
|
|
except re.error, e:
|
|
|
|
raise error.ParseError(_('invalid match pattern: %s') % e)
|
2014-01-31 04:03:18 +04:00
|
|
|
|
|
|
|
def matches(x):
|
|
|
|
c = repo[x]
|
2010-06-01 20:18:57 +04:00
|
|
|
for e in c.files() + [c.user(), c.description()]:
|
|
|
|
if gr.search(e):
|
2014-01-31 04:03:18 +04:00
|
|
|
return True
|
|
|
|
return False
|
|
|
|
|
2014-02-06 21:28:41 +04:00
|
|
|
return subset.filter(matches)
|
2010-06-01 20:18:57 +04:00
|
|
|
|
2012-02-23 21:05:20 +04:00
|
|
|
def _matchfiles(repo, subset, x):
|
|
|
|
# _matchfiles takes a revset list of prefixed arguments:
|
|
|
|
#
|
|
|
|
# [p:foo, i:bar, x:baz]
|
|
|
|
#
|
|
|
|
# builds a match object from them and filters subset. Allowed
|
|
|
|
# prefixes are 'p:' for regular patterns, 'i:' for include
|
2012-02-26 20:10:51 +04:00
|
|
|
# patterns and 'x:' for exclude patterns. Use 'r:' prefix to pass
|
|
|
|
# a revision identifier, or the empty string to reference the
|
|
|
|
# working directory, from which the match object is
|
2012-04-11 13:32:00 +04:00
|
|
|
# initialized. Use 'd:' to set the default matching mode, default
|
|
|
|
# to 'glob'. At most one 'r:' and 'd:' argument can be passed.
|
2012-02-23 21:05:20 +04:00
|
|
|
|
|
|
|
# i18n: "_matchfiles" is a keyword
|
|
|
|
l = getargs(x, 1, -1, _("_matchfiles requires at least one argument"))
|
|
|
|
pats, inc, exc = [], [], []
|
|
|
|
hasset = False
|
2012-04-11 13:32:00 +04:00
|
|
|
rev, default = None, None
|
2012-02-23 21:05:20 +04:00
|
|
|
for arg in l:
|
2012-07-26 08:58:43 +04:00
|
|
|
# i18n: "_matchfiles" is a keyword
|
2012-02-23 21:05:20 +04:00
|
|
|
s = getstring(arg, _("_matchfiles requires string arguments"))
|
|
|
|
prefix, value = s[:2], s[2:]
|
|
|
|
if prefix == 'p:':
|
|
|
|
pats.append(value)
|
|
|
|
elif prefix == 'i:':
|
|
|
|
inc.append(value)
|
|
|
|
elif prefix == 'x:':
|
|
|
|
exc.append(value)
|
2012-02-26 20:10:51 +04:00
|
|
|
elif prefix == 'r:':
|
|
|
|
if rev is not None:
|
2012-07-26 08:58:43 +04:00
|
|
|
# i18n: "_matchfiles" is a keyword
|
2012-02-26 20:10:51 +04:00
|
|
|
raise error.ParseError(_('_matchfiles expected at most one '
|
|
|
|
'revision'))
|
|
|
|
rev = value
|
2012-04-11 13:32:00 +04:00
|
|
|
elif prefix == 'd:':
|
|
|
|
if default is not None:
|
2012-07-26 08:58:43 +04:00
|
|
|
# i18n: "_matchfiles" is a keyword
|
2012-04-11 13:32:00 +04:00
|
|
|
raise error.ParseError(_('_matchfiles expected at most one '
|
|
|
|
'default mode'))
|
|
|
|
default = value
|
2012-02-23 21:05:20 +04:00
|
|
|
else:
|
2012-07-26 08:58:43 +04:00
|
|
|
# i18n: "_matchfiles" is a keyword
|
2012-02-23 21:05:20 +04:00
|
|
|
raise error.ParseError(_('invalid _matchfiles prefix: %s') % prefix)
|
|
|
|
if not hasset and matchmod.patkind(value) == 'set':
|
|
|
|
hasset = True
|
2012-04-11 13:32:00 +04:00
|
|
|
if not default:
|
|
|
|
default = 'glob'
|
2014-02-04 20:51:07 +04:00
|
|
|
|
|
|
|
def matches(x):
|
|
|
|
m = None
|
|
|
|
c = repo[x]
|
2012-02-26 20:10:51 +04:00
|
|
|
if not m or (hasset and rev is None):
|
|
|
|
ctx = c
|
|
|
|
if rev is not None:
|
|
|
|
ctx = repo[rev or None]
|
2012-02-23 21:05:20 +04:00
|
|
|
m = matchmod.match(repo.root, repo.getcwd(), pats, include=inc,
|
2012-04-11 13:32:00 +04:00
|
|
|
exclude=exc, ctx=ctx, default=default)
|
2012-01-21 09:05:04 +04:00
|
|
|
for f in c.files():
|
2010-06-01 20:18:57 +04:00
|
|
|
if m(f):
|
2014-02-04 20:51:07 +04:00
|
|
|
return True
|
|
|
|
return False
|
|
|
|
|
2014-02-06 21:28:41 +04:00
|
|
|
return subset.filter(matches)
|
2010-06-01 20:18:57 +04:00
|
|
|
|
2012-02-23 21:05:20 +04:00
|
|
|
def hasfile(repo, subset, x):
|
|
|
|
"""``file(pattern)``
|
|
|
|
Changesets affecting files matched by pattern.
|
2012-07-25 11:15:28 +04:00
|
|
|
|
2012-07-26 06:41:26 +04:00
|
|
|
For a faster but less accurate result, consider using ``filelog()``
|
|
|
|
instead.
|
2014-01-17 18:55:11 +04:00
|
|
|
|
|
|
|
This predicate uses ``glob:`` as the default kind of pattern.
|
2012-02-23 21:05:20 +04:00
|
|
|
"""
|
|
|
|
# i18n: "file" is a keyword
|
|
|
|
pat = getstring(x, _("file requires a pattern"))
|
|
|
|
return _matchfiles(repo, subset, ('string', 'p:' + pat))
|
|
|
|
|
2011-04-08 18:47:58 +04:00
|
|
|
def head(repo, subset, x):
|
|
|
|
"""``head()``
|
|
|
|
Changeset is a named branch head.
|
2010-10-23 21:21:51 +04:00
|
|
|
"""
|
2011-04-08 18:47:58 +04:00
|
|
|
# i18n: "head" is a keyword
|
|
|
|
getargs(x, 0, 0, _("head takes no arguments"))
|
|
|
|
hs = set()
|
|
|
|
for b, ls in repo.branchmap().iteritems():
|
|
|
|
hs.update(repo[h].rev() for h in ls)
|
2014-03-14 00:47:21 +04:00
|
|
|
return baseset(hs).filter(subset.__contains__)
|
2010-06-01 20:18:57 +04:00
|
|
|
|
2011-04-08 18:47:58 +04:00
|
|
|
def heads(repo, subset, x):
|
|
|
|
"""``heads(set)``
|
|
|
|
Members of set with no children in set.
|
|
|
|
"""
|
2014-01-24 02:20:58 +04:00
|
|
|
s = getset(repo, subset, x)
|
|
|
|
ps = parents(repo, subset, x)
|
|
|
|
return s - ps
|
2011-04-08 18:47:58 +04:00
|
|
|
|
2012-08-04 22:20:48 +04:00
|
|
|
def hidden(repo, subset, x):
|
|
|
|
"""``hidden()``
|
|
|
|
Hidden changesets.
|
|
|
|
"""
|
|
|
|
# i18n: "hidden" is a keyword
|
|
|
|
getargs(x, 0, 0, _("hidden takes no arguments"))
|
2013-01-13 11:39:16 +04:00
|
|
|
hiddenrevs = repoview.filterrevs(repo, 'visible')
|
2014-01-25 04:57:44 +04:00
|
|
|
return subset & hiddenrevs
|
2012-08-04 22:20:48 +04:00
|
|
|
|
2011-04-08 18:47:58 +04:00
|
|
|
def keyword(repo, subset, x):
|
|
|
|
"""``keyword(string)``
|
|
|
|
Search commit message, user name, and names of changed files for
|
2011-05-18 11:56:27 +04:00
|
|
|
string. The match is case-insensitive.
|
2011-04-08 18:47:58 +04:00
|
|
|
"""
|
|
|
|
# i18n: "keyword" is a keyword
|
2011-12-25 15:35:16 +04:00
|
|
|
kw = encoding.lower(getstring(x, _("keyword requires a string")))
|
2014-01-29 21:04:03 +04:00
|
|
|
|
|
|
|
def matches(r):
|
2010-06-01 20:18:57 +04:00
|
|
|
c = repo[r]
|
2014-01-29 21:04:03 +04:00
|
|
|
return util.any(kw in encoding.lower(t) for t in c.files() + [c.user(),
|
|
|
|
c.description()])
|
|
|
|
|
2014-02-06 21:28:41 +04:00
|
|
|
return subset.filter(matches)
|
2011-04-08 18:47:58 +04:00
|
|
|
|
|
|
|
def limit(repo, subset, x):
|
2011-09-17 07:57:47 +04:00
|
|
|
"""``limit(set, [n])``
|
|
|
|
First n members of set, defaulting to 1.
|
2011-04-08 18:47:58 +04:00
|
|
|
"""
|
|
|
|
# i18n: "limit" is a keyword
|
2011-09-17 07:57:47 +04:00
|
|
|
l = getargs(x, 1, 2, _("limit requires one or two arguments"))
|
2011-04-08 18:47:58 +04:00
|
|
|
try:
|
2011-09-17 07:57:47 +04:00
|
|
|
lim = 1
|
|
|
|
if len(l) == 2:
|
|
|
|
# i18n: "limit" is a keyword
|
|
|
|
lim = int(getstring(l[1], _("limit requires a number")))
|
2011-07-12 21:35:03 +04:00
|
|
|
except (TypeError, ValueError):
|
2011-04-08 18:47:58 +04:00
|
|
|
# i18n: "limit" is a keyword
|
|
|
|
raise error.ParseError(_("limit expects a number"))
|
2014-01-22 22:46:02 +04:00
|
|
|
ss = subset.set()
|
2014-02-03 22:15:15 +04:00
|
|
|
os = getset(repo, spanset(repo), l[0])
|
2014-01-29 04:19:30 +04:00
|
|
|
bs = baseset([])
|
|
|
|
it = iter(os)
|
|
|
|
for x in xrange(lim):
|
|
|
|
try:
|
|
|
|
y = it.next()
|
|
|
|
if y in ss:
|
|
|
|
bs.append(y)
|
|
|
|
except (StopIteration):
|
|
|
|
break
|
|
|
|
return bs
|
2011-04-08 18:47:58 +04:00
|
|
|
|
2011-04-30 19:56:43 +04:00
|
|
|
def last(repo, subset, x):
|
2011-09-17 07:57:47 +04:00
|
|
|
"""``last(set, [n])``
|
|
|
|
Last n members of set, defaulting to 1.
|
2011-04-30 19:56:43 +04:00
|
|
|
"""
|
|
|
|
# i18n: "last" is a keyword
|
2011-09-17 07:57:47 +04:00
|
|
|
l = getargs(x, 1, 2, _("last requires one or two arguments"))
|
2011-04-30 19:56:43 +04:00
|
|
|
try:
|
2011-09-17 07:57:47 +04:00
|
|
|
lim = 1
|
|
|
|
if len(l) == 2:
|
|
|
|
# i18n: "last" is a keyword
|
|
|
|
lim = int(getstring(l[1], _("last requires a number")))
|
2011-07-12 21:35:03 +04:00
|
|
|
except (TypeError, ValueError):
|
2011-04-30 19:56:43 +04:00
|
|
|
# i18n: "last" is a keyword
|
|
|
|
raise error.ParseError(_("last expects a number"))
|
2014-01-22 22:46:02 +04:00
|
|
|
ss = subset.set()
|
2014-02-20 00:56:41 +04:00
|
|
|
os = getset(repo, spanset(repo), l[0])
|
|
|
|
os.reverse()
|
|
|
|
bs = baseset([])
|
|
|
|
it = iter(os)
|
|
|
|
for x in xrange(lim):
|
|
|
|
try:
|
|
|
|
y = it.next()
|
|
|
|
if y in ss:
|
|
|
|
bs.append(y)
|
|
|
|
except (StopIteration):
|
|
|
|
break
|
|
|
|
return bs
|
2011-04-30 19:56:43 +04:00
|
|
|
|
2011-04-08 18:47:58 +04:00
|
|
|
def maxrev(repo, subset, x):
|
|
|
|
"""``max(set)``
|
|
|
|
Changeset with highest revision number in set.
|
|
|
|
"""
|
2014-02-03 22:15:15 +04:00
|
|
|
os = getset(repo, spanset(repo), x)
|
2011-05-01 19:35:05 +04:00
|
|
|
if os:
|
2014-02-18 23:35:03 +04:00
|
|
|
m = os.max()
|
2011-04-08 18:47:58 +04:00
|
|
|
if m in subset:
|
2014-01-21 23:39:26 +04:00
|
|
|
return baseset([m])
|
|
|
|
return baseset([])
|
2011-04-08 18:47:58 +04:00
|
|
|
|
|
|
|
def merge(repo, subset, x):
|
|
|
|
"""``merge()``
|
|
|
|
Changeset is a merge changeset.
|
|
|
|
"""
|
|
|
|
# i18n: "merge" is a keyword
|
|
|
|
getargs(x, 0, 0, _("merge takes no arguments"))
|
|
|
|
cl = repo.changelog
|
2014-02-06 21:28:41 +04:00
|
|
|
return subset.filter(lambda r: cl.parentrevs(r)[1] != -1)
|
2011-04-08 18:47:58 +04:00
|
|
|
|
2012-08-13 23:50:45 +04:00
|
|
|
def branchpoint(repo, subset, x):
|
|
|
|
"""``branchpoint()``
|
|
|
|
Changesets with more than one child.
|
|
|
|
"""
|
|
|
|
# i18n: "branchpoint" is a keyword
|
|
|
|
getargs(x, 0, 0, _("branchpoint takes no arguments"))
|
|
|
|
cl = repo.changelog
|
|
|
|
if not subset:
|
2014-01-21 23:39:26 +04:00
|
|
|
return baseset([])
|
2012-08-13 23:50:45 +04:00
|
|
|
baserev = min(subset)
|
|
|
|
parentscount = [0]*(len(repo) - baserev)
|
2012-10-15 19:42:40 +04:00
|
|
|
for r in cl.revs(start=baserev + 1):
|
2012-08-13 23:50:45 +04:00
|
|
|
for p in cl.parentrevs(r):
|
|
|
|
if p >= baserev:
|
|
|
|
parentscount[p - baserev] += 1
|
2014-02-06 21:28:41 +04:00
|
|
|
return subset.filter(lambda r: parentscount[r - baserev] > 1)
|
2012-08-13 23:50:45 +04:00
|
|
|
|
2011-04-08 18:47:58 +04:00
|
|
|
def minrev(repo, subset, x):
|
|
|
|
"""``min(set)``
|
|
|
|
Changeset with lowest revision number in set.
|
|
|
|
"""
|
2014-02-03 22:15:15 +04:00
|
|
|
os = getset(repo, spanset(repo), x)
|
2011-05-01 19:35:05 +04:00
|
|
|
if os:
|
2014-02-18 23:35:03 +04:00
|
|
|
m = os.min()
|
2011-04-08 18:47:58 +04:00
|
|
|
if m in subset:
|
2014-01-21 23:39:26 +04:00
|
|
|
return baseset([m])
|
|
|
|
return baseset([])
|
2010-06-01 20:18:57 +04:00
|
|
|
|
|
|
|
def modifies(repo, subset, x):
|
2010-10-23 21:21:51 +04:00
|
|
|
"""``modifies(pattern)``
|
|
|
|
Changesets modifying files matched by pattern.
|
2014-01-17 18:55:11 +04:00
|
|
|
|
|
|
|
The pattern without explicit kind like ``glob:`` is expected to be
|
|
|
|
relative to the current directory and match against a file or a
|
|
|
|
directory.
|
2010-10-23 21:21:51 +04:00
|
|
|
"""
|
2010-10-23 16:59:19 +04:00
|
|
|
# i18n: "modifies" is a keyword
|
2010-10-16 20:50:53 +04:00
|
|
|
pat = getstring(x, _("modifies requires a pattern"))
|
2010-06-01 20:18:57 +04:00
|
|
|
return checkstatus(repo, subset, pat, 0)
|
|
|
|
|
2012-04-14 00:32:49 +04:00
|
|
|
def node_(repo, subset, x):
|
2011-04-08 18:47:58 +04:00
|
|
|
"""``id(string)``
|
|
|
|
Revision non-ambiguously specified by the given hex string prefix.
|
2010-10-23 21:21:51 +04:00
|
|
|
"""
|
2011-04-08 18:47:58 +04:00
|
|
|
# i18n: "id" is a keyword
|
|
|
|
l = getargs(x, 1, 1, _("id requires one argument"))
|
|
|
|
# i18n: "id" is a keyword
|
|
|
|
n = getstring(l[0], _("id requires a string"))
|
|
|
|
if len(n) == 40:
|
|
|
|
rn = repo[n].rev()
|
|
|
|
else:
|
2012-05-15 03:25:13 +04:00
|
|
|
rn = None
|
|
|
|
pm = repo.changelog._partialmatch(n)
|
|
|
|
if pm is not None:
|
|
|
|
rn = repo.changelog.rev(pm)
|
|
|
|
|
2014-02-06 21:28:41 +04:00
|
|
|
return subset.filter(lambda r: r == rn)
|
2010-06-01 20:18:57 +04:00
|
|
|
|
2012-07-06 21:29:10 +04:00
|
|
|
def obsolete(repo, subset, x):
|
|
|
|
"""``obsolete()``
|
|
|
|
Mutable changeset with a newer version."""
|
2012-07-26 08:58:43 +04:00
|
|
|
# i18n: "obsolete" is a keyword
|
2012-07-06 21:29:10 +04:00
|
|
|
getargs(x, 0, 0, _("obsolete takes no arguments"))
|
2012-10-19 02:28:13 +04:00
|
|
|
obsoletes = obsmod.getrevs(repo, 'obsolete')
|
2014-01-25 04:57:44 +04:00
|
|
|
return subset & obsoletes
|
2012-07-06 21:29:10 +04:00
|
|
|
|
2012-07-07 08:47:30 +04:00
|
|
|
def origin(repo, subset, x):
|
|
|
|
"""``origin([set])``
|
|
|
|
Changesets that were specified as a source for the grafts, transplants or
|
|
|
|
rebases that created the given revisions. Omitting the optional set is the
|
|
|
|
same as passing all(). If a changeset created by these operations is itself
|
|
|
|
specified as a source for one of these operations, only the source changeset
|
|
|
|
for the first operation is selected.
|
|
|
|
"""
|
|
|
|
if x is not None:
|
2014-02-03 22:15:15 +04:00
|
|
|
args = getset(repo, spanset(repo), x).set()
|
2012-07-07 08:47:30 +04:00
|
|
|
else:
|
2014-02-03 22:15:15 +04:00
|
|
|
args = getall(repo, spanset(repo), x).set()
|
2012-07-07 08:47:30 +04:00
|
|
|
|
|
|
|
def _firstsrc(rev):
|
|
|
|
src = _getrevsource(repo, rev)
|
|
|
|
if src is None:
|
|
|
|
return None
|
|
|
|
|
|
|
|
while True:
|
|
|
|
prev = _getrevsource(repo, src)
|
|
|
|
|
|
|
|
if prev is None:
|
|
|
|
return src
|
|
|
|
src = prev
|
|
|
|
|
|
|
|
o = set([_firstsrc(r) for r in args])
|
2014-09-17 21:59:30 +04:00
|
|
|
o -= set([None])
|
2014-09-18 06:52:34 +04:00
|
|
|
return subset & o
|
2012-07-07 08:47:30 +04:00
|
|
|
|
2011-04-08 18:47:58 +04:00
|
|
|
def outgoing(repo, subset, x):
|
|
|
|
"""``outgoing([path])``
|
|
|
|
Changesets not found in the specified destination repository, or the
|
|
|
|
default push location.
|
2010-10-23 21:21:51 +04:00
|
|
|
"""
|
2011-04-08 18:47:58 +04:00
|
|
|
import hg # avoid start-up nasties
|
|
|
|
# i18n: "outgoing" is a keyword
|
2011-06-22 03:55:00 +04:00
|
|
|
l = getargs(x, 0, 1, _("outgoing takes one or no arguments"))
|
2011-04-08 18:47:58 +04:00
|
|
|
# i18n: "outgoing" is a keyword
|
|
|
|
dest = l and getstring(l[0], _("outgoing requires a repository path")) or ''
|
|
|
|
dest = repo.ui.expandpath(dest or 'default-push', dest or 'default')
|
|
|
|
dest, branches = hg.parseurl(dest)
|
|
|
|
revs, checkout = hg.addbranchrevs(repo, repo, branches, [])
|
|
|
|
if revs:
|
|
|
|
revs = [repo.lookup(rev) for rev in revs]
|
2011-06-10 20:43:38 +04:00
|
|
|
other = hg.peer(repo, {}, dest)
|
2011-04-08 18:47:58 +04:00
|
|
|
repo.ui.pushbuffer()
|
2012-01-09 06:47:16 +04:00
|
|
|
outgoing = discovery.findcommonoutgoing(repo, other, onlyheads=revs)
|
2011-04-08 18:47:58 +04:00
|
|
|
repo.ui.popbuffer()
|
|
|
|
cl = repo.changelog
|
2012-01-09 06:47:16 +04:00
|
|
|
o = set([cl.rev(r) for r in outgoing.missing])
|
2014-09-17 21:59:40 +04:00
|
|
|
return subset & o
|
2010-06-01 20:18:57 +04:00
|
|
|
|
2011-04-08 18:47:58 +04:00
|
|
|
def p1(repo, subset, x):
|
|
|
|
"""``p1([set])``
|
|
|
|
First parent of changesets in set, or the working directory.
|
2010-10-23 21:21:51 +04:00
|
|
|
"""
|
2011-04-08 18:47:58 +04:00
|
|
|
if x is None:
|
|
|
|
p = repo[x].p1().rev()
|
2014-09-17 21:59:52 +04:00
|
|
|
if p >= 0:
|
|
|
|
return subset & baseset([p])
|
|
|
|
return baseset([])
|
2011-04-08 18:47:58 +04:00
|
|
|
|
|
|
|
ps = set()
|
2010-06-01 20:18:57 +04:00
|
|
|
cl = repo.changelog
|
2014-02-03 22:15:15 +04:00
|
|
|
for r in getset(repo, spanset(repo), x):
|
2011-04-08 18:47:58 +04:00
|
|
|
ps.add(cl.parentrevs(r)[0])
|
2014-09-17 15:40:30 +04:00
|
|
|
ps -= set([node.nullrev])
|
2014-01-25 04:57:44 +04:00
|
|
|
return subset & ps
|
2010-06-01 20:18:57 +04:00
|
|
|
|
2011-04-08 18:47:58 +04:00
|
|
|
def p2(repo, subset, x):
|
|
|
|
"""``p2([set])``
|
|
|
|
Second parent of changesets in set, or the working directory.
|
2010-10-23 21:21:51 +04:00
|
|
|
"""
|
2011-04-08 18:47:58 +04:00
|
|
|
if x is None:
|
|
|
|
ps = repo[x].parents()
|
|
|
|
try:
|
|
|
|
p = ps[1].rev()
|
2014-09-17 22:00:03 +04:00
|
|
|
if p >= 0:
|
|
|
|
return subset & baseset([p])
|
|
|
|
return baseset([])
|
2011-04-08 18:47:58 +04:00
|
|
|
except IndexError:
|
2014-01-21 23:39:26 +04:00
|
|
|
return baseset([])
|
2010-06-01 20:18:57 +04:00
|
|
|
|
2011-04-08 18:47:58 +04:00
|
|
|
ps = set()
|
|
|
|
cl = repo.changelog
|
2014-02-03 22:15:15 +04:00
|
|
|
for r in getset(repo, spanset(repo), x):
|
2011-04-08 18:47:58 +04:00
|
|
|
ps.add(cl.parentrevs(r)[1])
|
2014-09-17 15:40:30 +04:00
|
|
|
ps -= set([node.nullrev])
|
2014-01-25 04:57:44 +04:00
|
|
|
return subset & ps
|
2010-06-01 20:18:57 +04:00
|
|
|
|
2011-04-08 18:47:58 +04:00
|
|
|
def parents(repo, subset, x):
|
|
|
|
"""``parents([set])``
|
|
|
|
The set of all parents for all changesets in set, or the working directory.
|
2010-10-23 21:21:51 +04:00
|
|
|
"""
|
2011-04-08 18:47:58 +04:00
|
|
|
if x is None:
|
2014-09-18 06:44:03 +04:00
|
|
|
ps = set(p.rev() for p in repo[x].parents())
|
|
|
|
else:
|
|
|
|
ps = set()
|
|
|
|
cl = repo.changelog
|
|
|
|
for r in getset(repo, spanset(repo), x):
|
|
|
|
ps.update(cl.parentrevs(r))
|
2014-09-18 06:49:26 +04:00
|
|
|
ps -= set([node.nullrev])
|
2014-09-13 02:00:51 +04:00
|
|
|
return baseset(ps) & subset
|
2010-06-01 20:18:57 +04:00
|
|
|
|
2011-04-30 19:43:04 +04:00
|
|
|
def parentspec(repo, subset, x, n):
|
|
|
|
"""``set^0``
|
|
|
|
The set.
|
|
|
|
``set^1`` (or ``set^``), ``set^2``
|
|
|
|
First or second parent, respectively, of all changesets in set.
|
|
|
|
"""
|
|
|
|
try:
|
|
|
|
n = int(n[1])
|
2011-04-30 20:25:45 +04:00
|
|
|
if n not in (0, 1, 2):
|
2011-04-30 19:43:04 +04:00
|
|
|
raise ValueError
|
2011-07-12 21:35:03 +04:00
|
|
|
except (TypeError, ValueError):
|
2011-04-30 19:43:04 +04:00
|
|
|
raise error.ParseError(_("^ expects a number 0, 1, or 2"))
|
|
|
|
ps = set()
|
|
|
|
cl = repo.changelog
|
2014-01-21 23:39:26 +04:00
|
|
|
for r in getset(repo, baseset(cl), x):
|
2011-04-30 19:43:04 +04:00
|
|
|
if n == 0:
|
|
|
|
ps.add(r)
|
|
|
|
elif n == 1:
|
|
|
|
ps.add(cl.parentrevs(r)[0])
|
|
|
|
elif n == 2:
|
|
|
|
parents = cl.parentrevs(r)
|
|
|
|
if len(parents) > 1:
|
|
|
|
ps.add(parents[1])
|
2014-01-25 04:57:44 +04:00
|
|
|
return subset & ps
|
2011-04-30 19:43:04 +04:00
|
|
|
|
2010-08-13 20:11:41 +04:00
|
|
|
def present(repo, subset, x):
|
2010-10-23 21:21:51 +04:00
|
|
|
"""``present(set)``
|
|
|
|
An empty set, if any revision in set isn't found; otherwise,
|
|
|
|
all revisions in set.
|
2012-05-16 12:02:30 +04:00
|
|
|
|
|
|
|
If any of specified revisions is not present in the local repository,
|
|
|
|
the query is normally aborted. But this predicate allows the query
|
|
|
|
to continue even in such cases.
|
2010-10-23 21:21:51 +04:00
|
|
|
"""
|
2010-08-13 20:11:41 +04:00
|
|
|
try:
|
|
|
|
return getset(repo, subset, x)
|
|
|
|
except error.RepoLookupError:
|
2014-01-21 23:39:26 +04:00
|
|
|
return baseset([])
|
2010-08-13 20:11:41 +04:00
|
|
|
|
2012-01-06 13:04:20 +04:00
|
|
|
def public(repo, subset, x):
|
|
|
|
"""``public()``
|
|
|
|
Changeset in public phase."""
|
2012-07-26 08:58:43 +04:00
|
|
|
# i18n: "public" is a keyword
|
2012-01-06 13:04:20 +04:00
|
|
|
getargs(x, 0, 0, _("public takes no arguments"))
|
2012-05-12 02:24:07 +04:00
|
|
|
pc = repo._phasecache
|
2014-02-06 21:28:41 +04:00
|
|
|
return subset.filter(lambda r: pc.phase(repo, r) == phases.public)
|
2012-01-06 13:04:20 +04:00
|
|
|
|
2012-01-20 00:31:05 +04:00
|
|
|
def remote(repo, subset, x):
|
2012-01-27 17:29:58 +04:00
|
|
|
"""``remote([id [,path]])``
|
2012-01-20 00:31:05 +04:00
|
|
|
Local revision that corresponds to the given identifier in a
|
|
|
|
remote repository, if present. Here, the '.' identifier is a
|
|
|
|
synonym for the current local branch.
|
|
|
|
"""
|
|
|
|
|
|
|
|
import hg # avoid start-up nasties
|
|
|
|
# i18n: "remote" is a keyword
|
2012-01-27 17:29:58 +04:00
|
|
|
l = getargs(x, 0, 2, _("remote takes one, two or no arguments"))
|
2012-01-20 00:31:05 +04:00
|
|
|
|
|
|
|
q = '.'
|
|
|
|
if len(l) > 0:
|
|
|
|
# i18n: "remote" is a keyword
|
|
|
|
q = getstring(l[0], _("remote requires a string id"))
|
|
|
|
if q == '.':
|
|
|
|
q = repo['.'].branch()
|
|
|
|
|
|
|
|
dest = ''
|
|
|
|
if len(l) > 1:
|
|
|
|
# i18n: "remote" is a keyword
|
|
|
|
dest = getstring(l[1], _("remote requires a repository path"))
|
|
|
|
dest = repo.ui.expandpath(dest or 'default')
|
|
|
|
dest, branches = hg.parseurl(dest)
|
|
|
|
revs, checkout = hg.addbranchrevs(repo, repo, branches, [])
|
|
|
|
if revs:
|
|
|
|
revs = [repo.lookup(rev) for rev in revs]
|
|
|
|
other = hg.peer(repo, {}, dest)
|
|
|
|
n = other.lookup(q)
|
|
|
|
if n in repo:
|
|
|
|
r = repo[n].rev()
|
2012-01-27 17:29:58 +04:00
|
|
|
if r in subset:
|
2014-01-21 23:39:26 +04:00
|
|
|
return baseset([r])
|
|
|
|
return baseset([])
|
2012-01-20 00:31:05 +04:00
|
|
|
|
2011-04-08 18:47:58 +04:00
|
|
|
def removes(repo, subset, x):
|
|
|
|
"""``removes(pattern)``
|
|
|
|
Changesets which remove files matching pattern.
|
2014-01-17 18:55:11 +04:00
|
|
|
|
|
|
|
The pattern without explicit kind like ``glob:`` is expected to be
|
|
|
|
relative to the current directory and match against a file or a
|
|
|
|
directory.
|
2011-04-08 18:47:58 +04:00
|
|
|
"""
|
|
|
|
# i18n: "removes" is a keyword
|
|
|
|
pat = getstring(x, _("removes requires a pattern"))
|
|
|
|
return checkstatus(repo, subset, pat, 2)
|
|
|
|
|
|
|
|
def rev(repo, subset, x):
|
|
|
|
"""``rev(number)``
|
|
|
|
Revision with the given numeric identifier.
|
|
|
|
"""
|
|
|
|
# i18n: "rev" is a keyword
|
|
|
|
l = getargs(x, 1, 1, _("rev requires one argument"))
|
|
|
|
try:
|
|
|
|
# i18n: "rev" is a keyword
|
|
|
|
l = int(getstring(l[0], _("rev requires a number")))
|
2011-07-12 21:35:03 +04:00
|
|
|
except (TypeError, ValueError):
|
2011-04-08 18:47:58 +04:00
|
|
|
# i18n: "rev" is a keyword
|
|
|
|
raise error.ParseError(_("rev expects a number"))
|
2014-09-17 22:00:09 +04:00
|
|
|
return subset & baseset([l])
|
2011-04-08 18:47:58 +04:00
|
|
|
|
revset: add "matching" keyword
This keyword can be used to find revisions that "match" one or more fields of a
given set of revisions.
A revision matches another if all the selected fields (description, author,
branch, date, files, phase, parents, substate, user, summary and/or metadata)
match the corresponding values of those fields on the source revision.
By default this keyword looks for revisions that whose metadata match
(description, author and date) making it ideal to look for duplicate revisions.
matching takes 2 arguments (the second being optional):
1.- rev: a revset represeting a _single_ revision (e.g. tip, ., p1(.), etc)
2.- [field(s) to match]: an optional string containing the field or fields
(separated by spaces) to match.
Valid fields are most regular context fields and some special fields:
* regular fields:
- description, author, branch, date, files, phase, parents,
substate, user.
Note that author and user are synonyms.
* special fields: summary, metadata.
- summary: matches the first line of the description.
- metatadata: It is equivalent to matching 'description user date'
(i.e. it matches the main metadata fields).
Examples:
1.- Look for revisions with the same metadata (author, description and date)
as the 11th revision:
hg log -r "matching(11)"
2.- Look for revisions with the same description as the 11th revision:
hg log -r "matching(11, description)"
3.- Look for revisions with the same 'summary' (i.e. same first line on their
description) as the 11th revision:
hg log -r "matching(11, summary)"
4.- Look for revisions with the same author as the current revision:
hg log -r "matching(., author)"
You could use 'user' rather than 'author' to get the same result.
5.- Look for revisions with the same description _AND_ author as the tip of the
repository:
hg log -r "matching(tip, 'author description')"
6.- Look for revisions touching the same files as the parent of the tip of the
repository
hg log -r "matching(p1(tip), files)"
7.- Look for revisions whose subrepos are on the same state as the tip of the
repository or its parent
hg log -r "matching(p1(tip):tip, substate)"
8.- Look for revisions whose author and subrepo states both match those of any
of the revisions on the stable branch:
hg log -r "matching(branch(stable), 'author substate')"
2012-04-01 16:12:14 +04:00
|
|
|
def matching(repo, subset, x):
|
|
|
|
"""``matching(revision [, field])``
|
|
|
|
Changesets in which a given set of fields match the set of fields in the
|
|
|
|
selected revision or set.
|
2012-04-26 16:32:48 +04:00
|
|
|
|
revset: add "matching" keyword
This keyword can be used to find revisions that "match" one or more fields of a
given set of revisions.
A revision matches another if all the selected fields (description, author,
branch, date, files, phase, parents, substate, user, summary and/or metadata)
match the corresponding values of those fields on the source revision.
By default this keyword looks for revisions that whose metadata match
(description, author and date) making it ideal to look for duplicate revisions.
matching takes 2 arguments (the second being optional):
1.- rev: a revset represeting a _single_ revision (e.g. tip, ., p1(.), etc)
2.- [field(s) to match]: an optional string containing the field or fields
(separated by spaces) to match.
Valid fields are most regular context fields and some special fields:
* regular fields:
- description, author, branch, date, files, phase, parents,
substate, user.
Note that author and user are synonyms.
* special fields: summary, metadata.
- summary: matches the first line of the description.
- metatadata: It is equivalent to matching 'description user date'
(i.e. it matches the main metadata fields).
Examples:
1.- Look for revisions with the same metadata (author, description and date)
as the 11th revision:
hg log -r "matching(11)"
2.- Look for revisions with the same description as the 11th revision:
hg log -r "matching(11, description)"
3.- Look for revisions with the same 'summary' (i.e. same first line on their
description) as the 11th revision:
hg log -r "matching(11, summary)"
4.- Look for revisions with the same author as the current revision:
hg log -r "matching(., author)"
You could use 'user' rather than 'author' to get the same result.
5.- Look for revisions with the same description _AND_ author as the tip of the
repository:
hg log -r "matching(tip, 'author description')"
6.- Look for revisions touching the same files as the parent of the tip of the
repository
hg log -r "matching(p1(tip), files)"
7.- Look for revisions whose subrepos are on the same state as the tip of the
repository or its parent
hg log -r "matching(p1(tip):tip, substate)"
8.- Look for revisions whose author and subrepo states both match those of any
of the revisions on the stable branch:
hg log -r "matching(branch(stable), 'author substate')"
2012-04-01 16:12:14 +04:00
|
|
|
To match more than one field pass the list of fields to match separated
|
2012-04-26 16:32:48 +04:00
|
|
|
by spaces (e.g. ``author description``).
|
|
|
|
|
|
|
|
Valid fields are most regular revision fields and some special fields.
|
|
|
|
|
|
|
|
Regular revision fields are ``description``, ``author``, ``branch``,
|
2012-06-14 01:32:58 +04:00
|
|
|
``date``, ``files``, ``phase``, ``parents``, ``substate``, ``user``
|
|
|
|
and ``diff``.
|
|
|
|
Note that ``author`` and ``user`` are synonyms. ``diff`` refers to the
|
|
|
|
contents of the revision. Two revisions matching their ``diff`` will
|
|
|
|
also match their ``files``.
|
2012-04-26 16:32:48 +04:00
|
|
|
|
|
|
|
Special fields are ``summary`` and ``metadata``:
|
|
|
|
``summary`` matches the first line of the description.
|
2012-05-10 22:17:05 +04:00
|
|
|
``metadata`` is equivalent to matching ``description user date``
|
2012-04-26 16:32:48 +04:00
|
|
|
(i.e. it matches the main metadata fields).
|
|
|
|
|
|
|
|
``metadata`` is the default field which is used when no fields are
|
|
|
|
specified. You can match more than one field at a time.
|
revset: add "matching" keyword
This keyword can be used to find revisions that "match" one or more fields of a
given set of revisions.
A revision matches another if all the selected fields (description, author,
branch, date, files, phase, parents, substate, user, summary and/or metadata)
match the corresponding values of those fields on the source revision.
By default this keyword looks for revisions that whose metadata match
(description, author and date) making it ideal to look for duplicate revisions.
matching takes 2 arguments (the second being optional):
1.- rev: a revset represeting a _single_ revision (e.g. tip, ., p1(.), etc)
2.- [field(s) to match]: an optional string containing the field or fields
(separated by spaces) to match.
Valid fields are most regular context fields and some special fields:
* regular fields:
- description, author, branch, date, files, phase, parents,
substate, user.
Note that author and user are synonyms.
* special fields: summary, metadata.
- summary: matches the first line of the description.
- metatadata: It is equivalent to matching 'description user date'
(i.e. it matches the main metadata fields).
Examples:
1.- Look for revisions with the same metadata (author, description and date)
as the 11th revision:
hg log -r "matching(11)"
2.- Look for revisions with the same description as the 11th revision:
hg log -r "matching(11, description)"
3.- Look for revisions with the same 'summary' (i.e. same first line on their
description) as the 11th revision:
hg log -r "matching(11, summary)"
4.- Look for revisions with the same author as the current revision:
hg log -r "matching(., author)"
You could use 'user' rather than 'author' to get the same result.
5.- Look for revisions with the same description _AND_ author as the tip of the
repository:
hg log -r "matching(tip, 'author description')"
6.- Look for revisions touching the same files as the parent of the tip of the
repository
hg log -r "matching(p1(tip), files)"
7.- Look for revisions whose subrepos are on the same state as the tip of the
repository or its parent
hg log -r "matching(p1(tip):tip, substate)"
8.- Look for revisions whose author and subrepo states both match those of any
of the revisions on the stable branch:
hg log -r "matching(branch(stable), 'author substate')"
2012-04-01 16:12:14 +04:00
|
|
|
"""
|
2012-07-26 08:58:43 +04:00
|
|
|
# i18n: "matching" is a keyword
|
revset: add "matching" keyword
This keyword can be used to find revisions that "match" one or more fields of a
given set of revisions.
A revision matches another if all the selected fields (description, author,
branch, date, files, phase, parents, substate, user, summary and/or metadata)
match the corresponding values of those fields on the source revision.
By default this keyword looks for revisions that whose metadata match
(description, author and date) making it ideal to look for duplicate revisions.
matching takes 2 arguments (the second being optional):
1.- rev: a revset represeting a _single_ revision (e.g. tip, ., p1(.), etc)
2.- [field(s) to match]: an optional string containing the field or fields
(separated by spaces) to match.
Valid fields are most regular context fields and some special fields:
* regular fields:
- description, author, branch, date, files, phase, parents,
substate, user.
Note that author and user are synonyms.
* special fields: summary, metadata.
- summary: matches the first line of the description.
- metatadata: It is equivalent to matching 'description user date'
(i.e. it matches the main metadata fields).
Examples:
1.- Look for revisions with the same metadata (author, description and date)
as the 11th revision:
hg log -r "matching(11)"
2.- Look for revisions with the same description as the 11th revision:
hg log -r "matching(11, description)"
3.- Look for revisions with the same 'summary' (i.e. same first line on their
description) as the 11th revision:
hg log -r "matching(11, summary)"
4.- Look for revisions with the same author as the current revision:
hg log -r "matching(., author)"
You could use 'user' rather than 'author' to get the same result.
5.- Look for revisions with the same description _AND_ author as the tip of the
repository:
hg log -r "matching(tip, 'author description')"
6.- Look for revisions touching the same files as the parent of the tip of the
repository
hg log -r "matching(p1(tip), files)"
7.- Look for revisions whose subrepos are on the same state as the tip of the
repository or its parent
hg log -r "matching(p1(tip):tip, substate)"
8.- Look for revisions whose author and subrepo states both match those of any
of the revisions on the stable branch:
hg log -r "matching(branch(stable), 'author substate')"
2012-04-01 16:12:14 +04:00
|
|
|
l = getargs(x, 1, 2, _("matching takes 1 or 2 arguments"))
|
|
|
|
|
2014-01-21 23:39:26 +04:00
|
|
|
revs = getset(repo, baseset(repo.changelog), l[0])
|
revset: add "matching" keyword
This keyword can be used to find revisions that "match" one or more fields of a
given set of revisions.
A revision matches another if all the selected fields (description, author,
branch, date, files, phase, parents, substate, user, summary and/or metadata)
match the corresponding values of those fields on the source revision.
By default this keyword looks for revisions that whose metadata match
(description, author and date) making it ideal to look for duplicate revisions.
matching takes 2 arguments (the second being optional):
1.- rev: a revset represeting a _single_ revision (e.g. tip, ., p1(.), etc)
2.- [field(s) to match]: an optional string containing the field or fields
(separated by spaces) to match.
Valid fields are most regular context fields and some special fields:
* regular fields:
- description, author, branch, date, files, phase, parents,
substate, user.
Note that author and user are synonyms.
* special fields: summary, metadata.
- summary: matches the first line of the description.
- metatadata: It is equivalent to matching 'description user date'
(i.e. it matches the main metadata fields).
Examples:
1.- Look for revisions with the same metadata (author, description and date)
as the 11th revision:
hg log -r "matching(11)"
2.- Look for revisions with the same description as the 11th revision:
hg log -r "matching(11, description)"
3.- Look for revisions with the same 'summary' (i.e. same first line on their
description) as the 11th revision:
hg log -r "matching(11, summary)"
4.- Look for revisions with the same author as the current revision:
hg log -r "matching(., author)"
You could use 'user' rather than 'author' to get the same result.
5.- Look for revisions with the same description _AND_ author as the tip of the
repository:
hg log -r "matching(tip, 'author description')"
6.- Look for revisions touching the same files as the parent of the tip of the
repository
hg log -r "matching(p1(tip), files)"
7.- Look for revisions whose subrepos are on the same state as the tip of the
repository or its parent
hg log -r "matching(p1(tip):tip, substate)"
8.- Look for revisions whose author and subrepo states both match those of any
of the revisions on the stable branch:
hg log -r "matching(branch(stable), 'author substate')"
2012-04-01 16:12:14 +04:00
|
|
|
|
|
|
|
fieldlist = ['metadata']
|
|
|
|
if len(l) > 1:
|
|
|
|
fieldlist = getstring(l[1],
|
2012-07-26 08:58:43 +04:00
|
|
|
# i18n: "matching" is a keyword
|
revset: add "matching" keyword
This keyword can be used to find revisions that "match" one or more fields of a
given set of revisions.
A revision matches another if all the selected fields (description, author,
branch, date, files, phase, parents, substate, user, summary and/or metadata)
match the corresponding values of those fields on the source revision.
By default this keyword looks for revisions that whose metadata match
(description, author and date) making it ideal to look for duplicate revisions.
matching takes 2 arguments (the second being optional):
1.- rev: a revset represeting a _single_ revision (e.g. tip, ., p1(.), etc)
2.- [field(s) to match]: an optional string containing the field or fields
(separated by spaces) to match.
Valid fields are most regular context fields and some special fields:
* regular fields:
- description, author, branch, date, files, phase, parents,
substate, user.
Note that author and user are synonyms.
* special fields: summary, metadata.
- summary: matches the first line of the description.
- metatadata: It is equivalent to matching 'description user date'
(i.e. it matches the main metadata fields).
Examples:
1.- Look for revisions with the same metadata (author, description and date)
as the 11th revision:
hg log -r "matching(11)"
2.- Look for revisions with the same description as the 11th revision:
hg log -r "matching(11, description)"
3.- Look for revisions with the same 'summary' (i.e. same first line on their
description) as the 11th revision:
hg log -r "matching(11, summary)"
4.- Look for revisions with the same author as the current revision:
hg log -r "matching(., author)"
You could use 'user' rather than 'author' to get the same result.
5.- Look for revisions with the same description _AND_ author as the tip of the
repository:
hg log -r "matching(tip, 'author description')"
6.- Look for revisions touching the same files as the parent of the tip of the
repository
hg log -r "matching(p1(tip), files)"
7.- Look for revisions whose subrepos are on the same state as the tip of the
repository or its parent
hg log -r "matching(p1(tip):tip, substate)"
8.- Look for revisions whose author and subrepo states both match those of any
of the revisions on the stable branch:
hg log -r "matching(branch(stable), 'author substate')"
2012-04-01 16:12:14 +04:00
|
|
|
_("matching requires a string "
|
|
|
|
"as its second argument")).split()
|
|
|
|
|
2012-06-14 01:32:58 +04:00
|
|
|
# Make sure that there are no repeated fields,
|
|
|
|
# expand the 'special' 'metadata' field type
|
|
|
|
# and check the 'files' whenever we check the 'diff'
|
revset: add "matching" keyword
This keyword can be used to find revisions that "match" one or more fields of a
given set of revisions.
A revision matches another if all the selected fields (description, author,
branch, date, files, phase, parents, substate, user, summary and/or metadata)
match the corresponding values of those fields on the source revision.
By default this keyword looks for revisions that whose metadata match
(description, author and date) making it ideal to look for duplicate revisions.
matching takes 2 arguments (the second being optional):
1.- rev: a revset represeting a _single_ revision (e.g. tip, ., p1(.), etc)
2.- [field(s) to match]: an optional string containing the field or fields
(separated by spaces) to match.
Valid fields are most regular context fields and some special fields:
* regular fields:
- description, author, branch, date, files, phase, parents,
substate, user.
Note that author and user are synonyms.
* special fields: summary, metadata.
- summary: matches the first line of the description.
- metatadata: It is equivalent to matching 'description user date'
(i.e. it matches the main metadata fields).
Examples:
1.- Look for revisions with the same metadata (author, description and date)
as the 11th revision:
hg log -r "matching(11)"
2.- Look for revisions with the same description as the 11th revision:
hg log -r "matching(11, description)"
3.- Look for revisions with the same 'summary' (i.e. same first line on their
description) as the 11th revision:
hg log -r "matching(11, summary)"
4.- Look for revisions with the same author as the current revision:
hg log -r "matching(., author)"
You could use 'user' rather than 'author' to get the same result.
5.- Look for revisions with the same description _AND_ author as the tip of the
repository:
hg log -r "matching(tip, 'author description')"
6.- Look for revisions touching the same files as the parent of the tip of the
repository
hg log -r "matching(p1(tip), files)"
7.- Look for revisions whose subrepos are on the same state as the tip of the
repository or its parent
hg log -r "matching(p1(tip):tip, substate)"
8.- Look for revisions whose author and subrepo states both match those of any
of the revisions on the stable branch:
hg log -r "matching(branch(stable), 'author substate')"
2012-04-01 16:12:14 +04:00
|
|
|
fields = []
|
|
|
|
for field in fieldlist:
|
|
|
|
if field == 'metadata':
|
|
|
|
fields += ['user', 'description', 'date']
|
2012-06-14 01:32:58 +04:00
|
|
|
elif field == 'diff':
|
|
|
|
# a revision matching the diff must also match the files
|
|
|
|
# since matching the diff is very costly, make sure to
|
|
|
|
# also match the files first
|
|
|
|
fields += ['files', 'diff']
|
revset: add "matching" keyword
This keyword can be used to find revisions that "match" one or more fields of a
given set of revisions.
A revision matches another if all the selected fields (description, author,
branch, date, files, phase, parents, substate, user, summary and/or metadata)
match the corresponding values of those fields on the source revision.
By default this keyword looks for revisions that whose metadata match
(description, author and date) making it ideal to look for duplicate revisions.
matching takes 2 arguments (the second being optional):
1.- rev: a revset represeting a _single_ revision (e.g. tip, ., p1(.), etc)
2.- [field(s) to match]: an optional string containing the field or fields
(separated by spaces) to match.
Valid fields are most regular context fields and some special fields:
* regular fields:
- description, author, branch, date, files, phase, parents,
substate, user.
Note that author and user are synonyms.
* special fields: summary, metadata.
- summary: matches the first line of the description.
- metatadata: It is equivalent to matching 'description user date'
(i.e. it matches the main metadata fields).
Examples:
1.- Look for revisions with the same metadata (author, description and date)
as the 11th revision:
hg log -r "matching(11)"
2.- Look for revisions with the same description as the 11th revision:
hg log -r "matching(11, description)"
3.- Look for revisions with the same 'summary' (i.e. same first line on their
description) as the 11th revision:
hg log -r "matching(11, summary)"
4.- Look for revisions with the same author as the current revision:
hg log -r "matching(., author)"
You could use 'user' rather than 'author' to get the same result.
5.- Look for revisions with the same description _AND_ author as the tip of the
repository:
hg log -r "matching(tip, 'author description')"
6.- Look for revisions touching the same files as the parent of the tip of the
repository
hg log -r "matching(p1(tip), files)"
7.- Look for revisions whose subrepos are on the same state as the tip of the
repository or its parent
hg log -r "matching(p1(tip):tip, substate)"
8.- Look for revisions whose author and subrepo states both match those of any
of the revisions on the stable branch:
hg log -r "matching(branch(stable), 'author substate')"
2012-04-01 16:12:14 +04:00
|
|
|
else:
|
|
|
|
if field == 'author':
|
|
|
|
field = 'user'
|
|
|
|
fields.append(field)
|
|
|
|
fields = set(fields)
|
2012-04-13 15:35:45 +04:00
|
|
|
if 'summary' in fields and 'description' in fields:
|
|
|
|
# If a revision matches its description it also matches its summary
|
|
|
|
fields.discard('summary')
|
revset: add "matching" keyword
This keyword can be used to find revisions that "match" one or more fields of a
given set of revisions.
A revision matches another if all the selected fields (description, author,
branch, date, files, phase, parents, substate, user, summary and/or metadata)
match the corresponding values of those fields on the source revision.
By default this keyword looks for revisions that whose metadata match
(description, author and date) making it ideal to look for duplicate revisions.
matching takes 2 arguments (the second being optional):
1.- rev: a revset represeting a _single_ revision (e.g. tip, ., p1(.), etc)
2.- [field(s) to match]: an optional string containing the field or fields
(separated by spaces) to match.
Valid fields are most regular context fields and some special fields:
* regular fields:
- description, author, branch, date, files, phase, parents,
substate, user.
Note that author and user are synonyms.
* special fields: summary, metadata.
- summary: matches the first line of the description.
- metatadata: It is equivalent to matching 'description user date'
(i.e. it matches the main metadata fields).
Examples:
1.- Look for revisions with the same metadata (author, description and date)
as the 11th revision:
hg log -r "matching(11)"
2.- Look for revisions with the same description as the 11th revision:
hg log -r "matching(11, description)"
3.- Look for revisions with the same 'summary' (i.e. same first line on their
description) as the 11th revision:
hg log -r "matching(11, summary)"
4.- Look for revisions with the same author as the current revision:
hg log -r "matching(., author)"
You could use 'user' rather than 'author' to get the same result.
5.- Look for revisions with the same description _AND_ author as the tip of the
repository:
hg log -r "matching(tip, 'author description')"
6.- Look for revisions touching the same files as the parent of the tip of the
repository
hg log -r "matching(p1(tip), files)"
7.- Look for revisions whose subrepos are on the same state as the tip of the
repository or its parent
hg log -r "matching(p1(tip):tip, substate)"
8.- Look for revisions whose author and subrepo states both match those of any
of the revisions on the stable branch:
hg log -r "matching(branch(stable), 'author substate')"
2012-04-01 16:12:14 +04:00
|
|
|
|
|
|
|
# We may want to match more than one field
|
revset: speedup matching() by first matching fields that take less time to
match
This patch sorts the fields that are passed to the matching function so that it
always starts by matching those fields that take less time to match.
Not all fields take the same amount of time to match. I've done several
measurements running the following command:
hg --time log -r "matching(1, field)"
on the mercurial repository, and where 'field' was each one of the fields
accepted by match. In order to avoid the print overhead (which could be
different for different fields, given the different number of matches) I used a
modified version of the matching() function which always returns no matches.
These tests showed that different fields take wildly different amounts of time
to match. Particulary the substate field takes up to 25 seconds to match on my
machine, compared to the 0.3 seconds that takes to match the phase field or the
2 seconds (approx) that takes to match most fields. With this patch, matching
both the phase and the substate of a revision takes the same amount of time as
matching the phase.
The field match order introduced by this patch is as follows:
phase, parents, user, date, branch, summary, files, description, substate
An extra nice thing about this patch is that it makes the match time stable.
2012-04-14 03:41:03 +04:00
|
|
|
# Not all fields take the same amount of time to be matched
|
|
|
|
# Sort the selected fields in order of increasing matching cost
|
2012-04-17 12:33:47 +04:00
|
|
|
fieldorder = ['phase', 'parents', 'user', 'date', 'branch', 'summary',
|
2012-06-14 01:32:58 +04:00
|
|
|
'files', 'description', 'substate', 'diff']
|
revset: speedup matching() by first matching fields that take less time to
match
This patch sorts the fields that are passed to the matching function so that it
always starts by matching those fields that take less time to match.
Not all fields take the same amount of time to match. I've done several
measurements running the following command:
hg --time log -r "matching(1, field)"
on the mercurial repository, and where 'field' was each one of the fields
accepted by match. In order to avoid the print overhead (which could be
different for different fields, given the different number of matches) I used a
modified version of the matching() function which always returns no matches.
These tests showed that different fields take wildly different amounts of time
to match. Particulary the substate field takes up to 25 seconds to match on my
machine, compared to the 0.3 seconds that takes to match the phase field or the
2 seconds (approx) that takes to match most fields. With this patch, matching
both the phase and the substate of a revision takes the same amount of time as
matching the phase.
The field match order introduced by this patch is as follows:
phase, parents, user, date, branch, summary, files, description, substate
An extra nice thing about this patch is that it makes the match time stable.
2012-04-14 03:41:03 +04:00
|
|
|
def fieldkeyfunc(f):
|
|
|
|
try:
|
|
|
|
return fieldorder.index(f)
|
|
|
|
except ValueError:
|
|
|
|
# assume an unknown field is very costly
|
|
|
|
return len(fieldorder)
|
|
|
|
fields = list(fields)
|
|
|
|
fields.sort(key=fieldkeyfunc)
|
|
|
|
|
revset: add "matching" keyword
This keyword can be used to find revisions that "match" one or more fields of a
given set of revisions.
A revision matches another if all the selected fields (description, author,
branch, date, files, phase, parents, substate, user, summary and/or metadata)
match the corresponding values of those fields on the source revision.
By default this keyword looks for revisions that whose metadata match
(description, author and date) making it ideal to look for duplicate revisions.
matching takes 2 arguments (the second being optional):
1.- rev: a revset represeting a _single_ revision (e.g. tip, ., p1(.), etc)
2.- [field(s) to match]: an optional string containing the field or fields
(separated by spaces) to match.
Valid fields are most regular context fields and some special fields:
* regular fields:
- description, author, branch, date, files, phase, parents,
substate, user.
Note that author and user are synonyms.
* special fields: summary, metadata.
- summary: matches the first line of the description.
- metatadata: It is equivalent to matching 'description user date'
(i.e. it matches the main metadata fields).
Examples:
1.- Look for revisions with the same metadata (author, description and date)
as the 11th revision:
hg log -r "matching(11)"
2.- Look for revisions with the same description as the 11th revision:
hg log -r "matching(11, description)"
3.- Look for revisions with the same 'summary' (i.e. same first line on their
description) as the 11th revision:
hg log -r "matching(11, summary)"
4.- Look for revisions with the same author as the current revision:
hg log -r "matching(., author)"
You could use 'user' rather than 'author' to get the same result.
5.- Look for revisions with the same description _AND_ author as the tip of the
repository:
hg log -r "matching(tip, 'author description')"
6.- Look for revisions touching the same files as the parent of the tip of the
repository
hg log -r "matching(p1(tip), files)"
7.- Look for revisions whose subrepos are on the same state as the tip of the
repository or its parent
hg log -r "matching(p1(tip):tip, substate)"
8.- Look for revisions whose author and subrepo states both match those of any
of the revisions on the stable branch:
hg log -r "matching(branch(stable), 'author substate')"
2012-04-01 16:12:14 +04:00
|
|
|
# Each field will be matched with its own "getfield" function
|
|
|
|
# which will be added to the getfieldfuncs array of functions
|
|
|
|
getfieldfuncs = []
|
|
|
|
_funcs = {
|
|
|
|
'user': lambda r: repo[r].user(),
|
|
|
|
'branch': lambda r: repo[r].branch(),
|
|
|
|
'date': lambda r: repo[r].date(),
|
|
|
|
'description': lambda r: repo[r].description(),
|
|
|
|
'files': lambda r: repo[r].files(),
|
|
|
|
'parents': lambda r: repo[r].parents(),
|
|
|
|
'phase': lambda r: repo[r].phase(),
|
|
|
|
'substate': lambda r: repo[r].substate,
|
|
|
|
'summary': lambda r: repo[r].description().splitlines()[0],
|
2012-06-14 01:32:58 +04:00
|
|
|
'diff': lambda r: list(repo[r].diff(git=True),)
|
revset: add "matching" keyword
This keyword can be used to find revisions that "match" one or more fields of a
given set of revisions.
A revision matches another if all the selected fields (description, author,
branch, date, files, phase, parents, substate, user, summary and/or metadata)
match the corresponding values of those fields on the source revision.
By default this keyword looks for revisions that whose metadata match
(description, author and date) making it ideal to look for duplicate revisions.
matching takes 2 arguments (the second being optional):
1.- rev: a revset represeting a _single_ revision (e.g. tip, ., p1(.), etc)
2.- [field(s) to match]: an optional string containing the field or fields
(separated by spaces) to match.
Valid fields are most regular context fields and some special fields:
* regular fields:
- description, author, branch, date, files, phase, parents,
substate, user.
Note that author and user are synonyms.
* special fields: summary, metadata.
- summary: matches the first line of the description.
- metatadata: It is equivalent to matching 'description user date'
(i.e. it matches the main metadata fields).
Examples:
1.- Look for revisions with the same metadata (author, description and date)
as the 11th revision:
hg log -r "matching(11)"
2.- Look for revisions with the same description as the 11th revision:
hg log -r "matching(11, description)"
3.- Look for revisions with the same 'summary' (i.e. same first line on their
description) as the 11th revision:
hg log -r "matching(11, summary)"
4.- Look for revisions with the same author as the current revision:
hg log -r "matching(., author)"
You could use 'user' rather than 'author' to get the same result.
5.- Look for revisions with the same description _AND_ author as the tip of the
repository:
hg log -r "matching(tip, 'author description')"
6.- Look for revisions touching the same files as the parent of the tip of the
repository
hg log -r "matching(p1(tip), files)"
7.- Look for revisions whose subrepos are on the same state as the tip of the
repository or its parent
hg log -r "matching(p1(tip):tip, substate)"
8.- Look for revisions whose author and subrepo states both match those of any
of the revisions on the stable branch:
hg log -r "matching(branch(stable), 'author substate')"
2012-04-01 16:12:14 +04:00
|
|
|
}
|
|
|
|
for info in fields:
|
|
|
|
getfield = _funcs.get(info, None)
|
|
|
|
if getfield is None:
|
|
|
|
raise error.ParseError(
|
2012-07-26 08:58:43 +04:00
|
|
|
# i18n: "matching" is a keyword
|
revset: add "matching" keyword
This keyword can be used to find revisions that "match" one or more fields of a
given set of revisions.
A revision matches another if all the selected fields (description, author,
branch, date, files, phase, parents, substate, user, summary and/or metadata)
match the corresponding values of those fields on the source revision.
By default this keyword looks for revisions that whose metadata match
(description, author and date) making it ideal to look for duplicate revisions.
matching takes 2 arguments (the second being optional):
1.- rev: a revset represeting a _single_ revision (e.g. tip, ., p1(.), etc)
2.- [field(s) to match]: an optional string containing the field or fields
(separated by spaces) to match.
Valid fields are most regular context fields and some special fields:
* regular fields:
- description, author, branch, date, files, phase, parents,
substate, user.
Note that author and user are synonyms.
* special fields: summary, metadata.
- summary: matches the first line of the description.
- metatadata: It is equivalent to matching 'description user date'
(i.e. it matches the main metadata fields).
Examples:
1.- Look for revisions with the same metadata (author, description and date)
as the 11th revision:
hg log -r "matching(11)"
2.- Look for revisions with the same description as the 11th revision:
hg log -r "matching(11, description)"
3.- Look for revisions with the same 'summary' (i.e. same first line on their
description) as the 11th revision:
hg log -r "matching(11, summary)"
4.- Look for revisions with the same author as the current revision:
hg log -r "matching(., author)"
You could use 'user' rather than 'author' to get the same result.
5.- Look for revisions with the same description _AND_ author as the tip of the
repository:
hg log -r "matching(tip, 'author description')"
6.- Look for revisions touching the same files as the parent of the tip of the
repository
hg log -r "matching(p1(tip), files)"
7.- Look for revisions whose subrepos are on the same state as the tip of the
repository or its parent
hg log -r "matching(p1(tip):tip, substate)"
8.- Look for revisions whose author and subrepo states both match those of any
of the revisions on the stable branch:
hg log -r "matching(branch(stable), 'author substate')"
2012-04-01 16:12:14 +04:00
|
|
|
_("unexpected field name passed to matching: %s") % info)
|
|
|
|
getfieldfuncs.append(getfield)
|
|
|
|
# convert the getfield array of functions into a "getinfo" function
|
|
|
|
# which returns an array of field values (or a single value if there
|
|
|
|
# is only one field to match)
|
2012-04-13 15:46:49 +04:00
|
|
|
getinfo = lambda r: [f(r) for f in getfieldfuncs]
|
revset: add "matching" keyword
This keyword can be used to find revisions that "match" one or more fields of a
given set of revisions.
A revision matches another if all the selected fields (description, author,
branch, date, files, phase, parents, substate, user, summary and/or metadata)
match the corresponding values of those fields on the source revision.
By default this keyword looks for revisions that whose metadata match
(description, author and date) making it ideal to look for duplicate revisions.
matching takes 2 arguments (the second being optional):
1.- rev: a revset represeting a _single_ revision (e.g. tip, ., p1(.), etc)
2.- [field(s) to match]: an optional string containing the field or fields
(separated by spaces) to match.
Valid fields are most regular context fields and some special fields:
* regular fields:
- description, author, branch, date, files, phase, parents,
substate, user.
Note that author and user are synonyms.
* special fields: summary, metadata.
- summary: matches the first line of the description.
- metatadata: It is equivalent to matching 'description user date'
(i.e. it matches the main metadata fields).
Examples:
1.- Look for revisions with the same metadata (author, description and date)
as the 11th revision:
hg log -r "matching(11)"
2.- Look for revisions with the same description as the 11th revision:
hg log -r "matching(11, description)"
3.- Look for revisions with the same 'summary' (i.e. same first line on their
description) as the 11th revision:
hg log -r "matching(11, summary)"
4.- Look for revisions with the same author as the current revision:
hg log -r "matching(., author)"
You could use 'user' rather than 'author' to get the same result.
5.- Look for revisions with the same description _AND_ author as the tip of the
repository:
hg log -r "matching(tip, 'author description')"
6.- Look for revisions touching the same files as the parent of the tip of the
repository
hg log -r "matching(p1(tip), files)"
7.- Look for revisions whose subrepos are on the same state as the tip of the
repository or its parent
hg log -r "matching(p1(tip):tip, substate)"
8.- Look for revisions whose author and subrepo states both match those of any
of the revisions on the stable branch:
hg log -r "matching(branch(stable), 'author substate')"
2012-04-01 16:12:14 +04:00
|
|
|
|
2014-02-04 21:14:45 +04:00
|
|
|
def matches(x):
|
|
|
|
for rev in revs:
|
|
|
|
target = getinfo(rev)
|
2012-04-13 15:46:49 +04:00
|
|
|
match = True
|
|
|
|
for n, f in enumerate(getfieldfuncs):
|
2014-02-04 21:14:45 +04:00
|
|
|
if target[n] != f(x):
|
2012-04-13 15:46:49 +04:00
|
|
|
match = False
|
|
|
|
if match:
|
2014-02-04 21:14:45 +04:00
|
|
|
return True
|
|
|
|
return False
|
|
|
|
|
2014-02-06 21:28:41 +04:00
|
|
|
return subset.filter(matches)
|
revset: add "matching" keyword
This keyword can be used to find revisions that "match" one or more fields of a
given set of revisions.
A revision matches another if all the selected fields (description, author,
branch, date, files, phase, parents, substate, user, summary and/or metadata)
match the corresponding values of those fields on the source revision.
By default this keyword looks for revisions that whose metadata match
(description, author and date) making it ideal to look for duplicate revisions.
matching takes 2 arguments (the second being optional):
1.- rev: a revset represeting a _single_ revision (e.g. tip, ., p1(.), etc)
2.- [field(s) to match]: an optional string containing the field or fields
(separated by spaces) to match.
Valid fields are most regular context fields and some special fields:
* regular fields:
- description, author, branch, date, files, phase, parents,
substate, user.
Note that author and user are synonyms.
* special fields: summary, metadata.
- summary: matches the first line of the description.
- metatadata: It is equivalent to matching 'description user date'
(i.e. it matches the main metadata fields).
Examples:
1.- Look for revisions with the same metadata (author, description and date)
as the 11th revision:
hg log -r "matching(11)"
2.- Look for revisions with the same description as the 11th revision:
hg log -r "matching(11, description)"
3.- Look for revisions with the same 'summary' (i.e. same first line on their
description) as the 11th revision:
hg log -r "matching(11, summary)"
4.- Look for revisions with the same author as the current revision:
hg log -r "matching(., author)"
You could use 'user' rather than 'author' to get the same result.
5.- Look for revisions with the same description _AND_ author as the tip of the
repository:
hg log -r "matching(tip, 'author description')"
6.- Look for revisions touching the same files as the parent of the tip of the
repository
hg log -r "matching(p1(tip), files)"
7.- Look for revisions whose subrepos are on the same state as the tip of the
repository or its parent
hg log -r "matching(p1(tip):tip, substate)"
8.- Look for revisions whose author and subrepo states both match those of any
of the revisions on the stable branch:
hg log -r "matching(branch(stable), 'author substate')"
2012-04-01 16:12:14 +04:00
|
|
|
|
2011-04-08 18:47:58 +04:00
|
|
|
def reverse(repo, subset, x):
|
|
|
|
"""``reverse(set)``
|
|
|
|
Reverse order of set.
|
|
|
|
"""
|
|
|
|
l = getset(repo, subset, x)
|
|
|
|
l.reverse()
|
|
|
|
return l
|
|
|
|
|
|
|
|
def roots(repo, subset, x):
|
|
|
|
"""``roots(set)``
|
2012-04-08 13:11:30 +04:00
|
|
|
Changesets in set with no parent changeset in set.
|
2011-04-08 18:47:58 +04:00
|
|
|
"""
|
2014-04-01 03:03:34 +04:00
|
|
|
s = getset(repo, spanset(repo), x).set()
|
|
|
|
subset = baseset([r for r in s if r in subset.set()])
|
2012-04-08 13:11:30 +04:00
|
|
|
cs = _children(repo, subset, s)
|
2014-01-24 02:20:58 +04:00
|
|
|
return subset - cs
|
2011-04-08 18:47:58 +04:00
|
|
|
|
2012-01-06 13:04:20 +04:00
|
|
|
def secret(repo, subset, x):
|
|
|
|
"""``secret()``
|
|
|
|
Changeset in secret phase."""
|
2012-07-26 08:58:43 +04:00
|
|
|
# i18n: "secret" is a keyword
|
2012-01-06 13:04:20 +04:00
|
|
|
getargs(x, 0, 0, _("secret takes no arguments"))
|
2012-05-12 02:24:07 +04:00
|
|
|
pc = repo._phasecache
|
2014-02-06 21:28:41 +04:00
|
|
|
return subset.filter(lambda x: pc.phase(repo, x) == phases.secret)
|
2012-01-06 13:04:20 +04:00
|
|
|
|
2010-06-01 20:18:57 +04:00
|
|
|
def sort(repo, subset, x):
|
2010-10-23 21:21:51 +04:00
|
|
|
"""``sort(set[, [-]key...])``
|
|
|
|
Sort set by keys. The default sort order is ascending, specify a key
|
|
|
|
as ``-key`` to sort in descending order.
|
|
|
|
|
|
|
|
The keys can be:
|
|
|
|
|
|
|
|
- ``rev`` for the revision number,
|
|
|
|
- ``branch`` for the branch name,
|
|
|
|
- ``desc`` for the commit message (description),
|
|
|
|
- ``user`` for user name (``author`` can be used as an alias),
|
|
|
|
- ``date`` for the commit date
|
|
|
|
"""
|
2010-10-23 16:59:19 +04:00
|
|
|
# i18n: "sort" is a keyword
|
2010-10-16 20:50:53 +04:00
|
|
|
l = getargs(x, 1, 2, _("sort requires one or two arguments"))
|
2010-06-01 20:18:57 +04:00
|
|
|
keys = "rev"
|
|
|
|
if len(l) == 2:
|
2012-07-26 08:58:43 +04:00
|
|
|
# i18n: "sort" is a keyword
|
2010-06-18 23:31:19 +04:00
|
|
|
keys = getstring(l[1], _("sort spec must be a string"))
|
2010-06-01 20:18:57 +04:00
|
|
|
|
|
|
|
s = l[0]
|
|
|
|
keys = keys.split()
|
|
|
|
l = []
|
|
|
|
def invert(s):
|
|
|
|
return "".join(chr(255 - ord(c)) for c in s)
|
2014-03-14 04:15:21 +04:00
|
|
|
revs = getset(repo, subset, s)
|
|
|
|
if keys == ["rev"]:
|
|
|
|
revs.sort()
|
|
|
|
return revs
|
|
|
|
elif keys == ["-rev"]:
|
|
|
|
revs.sort(reverse=True)
|
|
|
|
return revs
|
|
|
|
for r in revs:
|
2010-06-01 20:18:57 +04:00
|
|
|
c = repo[r]
|
|
|
|
e = []
|
|
|
|
for k in keys:
|
|
|
|
if k == 'rev':
|
|
|
|
e.append(r)
|
|
|
|
elif k == '-rev':
|
|
|
|
e.append(-r)
|
|
|
|
elif k == 'branch':
|
|
|
|
e.append(c.branch())
|
|
|
|
elif k == '-branch':
|
|
|
|
e.append(invert(c.branch()))
|
|
|
|
elif k == 'desc':
|
|
|
|
e.append(c.description())
|
|
|
|
elif k == '-desc':
|
|
|
|
e.append(invert(c.description()))
|
|
|
|
elif k in 'user author':
|
|
|
|
e.append(c.user())
|
|
|
|
elif k in '-user -author':
|
|
|
|
e.append(invert(c.user()))
|
|
|
|
elif k == 'date':
|
|
|
|
e.append(c.date()[0])
|
|
|
|
elif k == '-date':
|
|
|
|
e.append(-c.date()[0])
|
|
|
|
else:
|
2010-06-18 23:31:19 +04:00
|
|
|
raise error.ParseError(_("unknown sort key %r") % k)
|
2010-06-01 20:18:57 +04:00
|
|
|
e.append(r)
|
|
|
|
l.append(e)
|
|
|
|
l.sort()
|
2014-01-21 23:39:26 +04:00
|
|
|
return baseset([e[-1] for e in l])
|
2010-06-01 20:18:57 +04:00
|
|
|
|
2012-05-31 02:13:33 +04:00
|
|
|
def _stringmatcher(pattern):
|
|
|
|
"""
|
|
|
|
accepts a string, possibly starting with 're:' or 'literal:' prefix.
|
|
|
|
returns the matcher name, pattern, and matcher function.
|
|
|
|
missing or unknown prefixes are treated as literal matches.
|
|
|
|
|
|
|
|
helper for tests:
|
|
|
|
>>> def test(pattern, *tests):
|
|
|
|
... kind, pattern, matcher = _stringmatcher(pattern)
|
|
|
|
... return (kind, pattern, [bool(matcher(t)) for t in tests])
|
|
|
|
|
|
|
|
exact matching (no prefix):
|
|
|
|
>>> test('abcdefg', 'abc', 'def', 'abcdefg')
|
|
|
|
('literal', 'abcdefg', [False, False, True])
|
|
|
|
|
|
|
|
regex matching ('re:' prefix)
|
|
|
|
>>> test('re:a.+b', 'nomatch', 'fooadef', 'fooadefbar')
|
|
|
|
('re', 'a.+b', [False, False, True])
|
|
|
|
|
|
|
|
force exact matches ('literal:' prefix)
|
|
|
|
>>> test('literal:re:foobar', 'foobar', 're:foobar')
|
|
|
|
('literal', 're:foobar', [False, True])
|
|
|
|
|
|
|
|
unknown prefixes are ignored and treated as literals
|
|
|
|
>>> test('foo:bar', 'foo', 'bar', 'foo:bar')
|
|
|
|
('literal', 'foo:bar', [False, False, True])
|
|
|
|
"""
|
|
|
|
if pattern.startswith('re:'):
|
|
|
|
pattern = pattern[3:]
|
|
|
|
try:
|
|
|
|
regex = re.compile(pattern)
|
|
|
|
except re.error, e:
|
|
|
|
raise error.ParseError(_('invalid regular expression: %s')
|
|
|
|
% e)
|
|
|
|
return 're', pattern, regex.search
|
|
|
|
elif pattern.startswith('literal:'):
|
|
|
|
pattern = pattern[8:]
|
|
|
|
return 'literal', pattern, pattern.__eq__
|
|
|
|
|
2012-05-31 02:13:58 +04:00
|
|
|
def _substringmatcher(pattern):
|
|
|
|
kind, pattern, matcher = _stringmatcher(pattern)
|
|
|
|
if kind == 'literal':
|
|
|
|
matcher = lambda s: pattern in s
|
|
|
|
return kind, pattern, matcher
|
2012-05-31 02:13:33 +04:00
|
|
|
|
2010-10-10 21:41:36 +04:00
|
|
|
def tag(repo, subset, x):
|
2011-05-18 11:31:19 +04:00
|
|
|
"""``tag([name])``
|
2010-10-23 21:21:51 +04:00
|
|
|
The specified tag by name, or all tagged revisions if no name is given.
|
2014-03-25 05:27:40 +04:00
|
|
|
|
|
|
|
If `name` starts with `re:`, the remainder of the name is treated as
|
|
|
|
a regular expression. To match a tag that actually starts with `re:`,
|
|
|
|
use the prefix `literal:`.
|
2010-10-23 21:21:51 +04:00
|
|
|
"""
|
2010-10-23 16:59:19 +04:00
|
|
|
# i18n: "tag" is a keyword
|
2010-10-10 21:41:36 +04:00
|
|
|
args = getargs(x, 0, 1, _("tag takes one or no arguments"))
|
2010-06-04 02:39:40 +04:00
|
|
|
cl = repo.changelog
|
2010-10-10 21:41:36 +04:00
|
|
|
if args:
|
2012-05-31 02:13:33 +04:00
|
|
|
pattern = getstring(args[0],
|
|
|
|
# i18n: "tag" is a keyword
|
|
|
|
_('the argument to tag must be a string'))
|
|
|
|
kind, pattern, matcher = _stringmatcher(pattern)
|
|
|
|
if kind == 'literal':
|
2012-06-02 00:13:05 +04:00
|
|
|
# avoid resolving all tags
|
|
|
|
tn = repo._tagscache.tags.get(pattern, None)
|
|
|
|
if tn is None:
|
2012-05-31 02:13:33 +04:00
|
|
|
raise util.Abort(_("tag '%s' does not exist") % pattern)
|
2012-06-02 00:13:05 +04:00
|
|
|
s = set([repo[tn].rev()])
|
2012-05-31 02:13:33 +04:00
|
|
|
else:
|
|
|
|
s = set([cl.rev(n) for t, n in repo.tagslist() if matcher(t)])
|
2010-10-10 21:41:36 +04:00
|
|
|
else:
|
|
|
|
s = set([cl.rev(n) for t, n in repo.tagslist() if t != 'tip'])
|
2014-01-25 04:57:44 +04:00
|
|
|
return subset & s
|
2010-06-04 02:39:40 +04:00
|
|
|
|
2010-10-23 21:21:51 +04:00
|
|
|
def tagged(repo, subset, x):
|
|
|
|
return tag(repo, subset, x)
|
|
|
|
|
2012-07-06 02:18:09 +04:00
|
|
|
def unstable(repo, subset, x):
|
|
|
|
"""``unstable()``
|
2012-07-30 17:48:04 +04:00
|
|
|
Non-obsolete changesets with obsolete ancestors.
|
|
|
|
"""
|
2012-07-26 08:58:43 +04:00
|
|
|
# i18n: "unstable" is a keyword
|
2012-07-26 08:58:43 +04:00
|
|
|
getargs(x, 0, 0, _("unstable takes no arguments"))
|
2012-10-19 02:28:13 +04:00
|
|
|
unstables = obsmod.getrevs(repo, 'unstable')
|
2014-01-25 04:57:44 +04:00
|
|
|
return subset & unstables
|
2012-07-06 02:18:09 +04:00
|
|
|
|
|
|
|
|
2011-04-08 18:47:58 +04:00
|
|
|
def user(repo, subset, x):
|
|
|
|
"""``user(string)``
|
2011-05-18 11:56:27 +04:00
|
|
|
User name contains string. The match is case-insensitive.
|
2012-05-31 02:13:58 +04:00
|
|
|
|
|
|
|
If `string` starts with `re:`, the remainder of the string is treated as
|
|
|
|
a regular expression. To match a user that actually contains `re:`, use
|
|
|
|
the prefix `literal:`.
|
2011-03-12 20:48:30 +03:00
|
|
|
"""
|
2011-04-08 18:47:58 +04:00
|
|
|
return author(repo, subset, x)
|
2011-03-12 20:48:30 +03:00
|
|
|
|
2012-01-16 11:21:22 +04:00
|
|
|
# for internal use
|
|
|
|
def _list(repo, subset, x):
|
|
|
|
s = getstring(x, "internal error")
|
|
|
|
if not s:
|
2014-01-21 23:39:26 +04:00
|
|
|
return baseset([])
|
2012-01-16 11:21:22 +04:00
|
|
|
ls = [repo[r].rev() for r in s.split('\0')]
|
2014-01-22 22:46:02 +04:00
|
|
|
s = subset.set()
|
|
|
|
return baseset([r for r in ls if r in s])
|
2012-01-16 11:21:22 +04:00
|
|
|
|
2014-02-27 00:36:36 +04:00
|
|
|
# for internal use
|
|
|
|
def _intlist(repo, subset, x):
|
|
|
|
s = getstring(x, "internal error")
|
|
|
|
if not s:
|
|
|
|
return baseset([])
|
|
|
|
ls = [int(r) for r in s.split('\0')]
|
|
|
|
s = subset.set()
|
|
|
|
return baseset([r for r in ls if r in s])
|
|
|
|
|
2014-02-27 05:15:55 +04:00
|
|
|
# for internal use
|
|
|
|
def _hexlist(repo, subset, x):
|
|
|
|
s = getstring(x, "internal error")
|
|
|
|
if not s:
|
|
|
|
return baseset([])
|
|
|
|
cl = repo.changelog
|
|
|
|
ls = [cl.rev(node.bin(r)) for r in s.split('\0')]
|
|
|
|
s = subset.set()
|
|
|
|
return baseset([r for r in ls if r in s])
|
|
|
|
|
2010-06-01 20:18:57 +04:00
|
|
|
symbols = {
|
2010-06-04 19:27:23 +04:00
|
|
|
"adds": adds,
|
|
|
|
"all": getall,
|
2010-06-01 20:18:57 +04:00
|
|
|
"ancestor": ancestor,
|
|
|
|
"ancestors": ancestors,
|
2012-04-11 13:25:34 +04:00
|
|
|
"_firstancestors": _firstancestors,
|
2010-06-01 20:18:57 +04:00
|
|
|
"author": author,
|
2013-11-16 20:57:08 +04:00
|
|
|
"only": only,
|
2011-09-19 00:54:11 +04:00
|
|
|
"bisect": bisect,
|
2011-03-12 20:48:30 +03:00
|
|
|
"bisected": bisected,
|
2011-02-10 22:46:28 +03:00
|
|
|
"bookmark": bookmark,
|
2010-06-04 19:27:23 +04:00
|
|
|
"branch": branch,
|
2012-08-13 23:50:45 +04:00
|
|
|
"branchpoint": branchpoint,
|
2012-10-19 02:39:06 +04:00
|
|
|
"bumped": bumped,
|
2012-11-01 03:23:23 +04:00
|
|
|
"bundle": bundle,
|
2010-06-04 19:27:23 +04:00
|
|
|
"children": children,
|
|
|
|
"closed": closed,
|
|
|
|
"contains": contains,
|
2012-05-13 09:12:26 +04:00
|
|
|
"converted": converted,
|
2010-06-01 20:18:57 +04:00
|
|
|
"date": date,
|
2011-06-17 00:47:34 +04:00
|
|
|
"desc": desc,
|
2010-06-04 19:27:23 +04:00
|
|
|
"descendants": descendants,
|
2012-04-11 13:25:34 +04:00
|
|
|
"_firstdescendants": _firstdescendants,
|
revset: add destination() predicate
This predicate is used to find csets that were created because of a graft,
transplant or rebase --keep. An optional revset can be supplied, in which case
the result will be limited to those copies which specified one of the revs as
the source for the command.
hg log -r destination() # csets copied from anywhere
hg log -r destination(branch(default)) # all csets copied from default
hg log -r origin(x) or destination(origin(x)) # all instances of x
This predicate will follow a cset through different types of copies. Given a
repo with a cset 'S' that is grafted to create G(S), which itself is
transplanted to become T(G(S)):
o-S
/
o-o-G(S)
\
o-T(G(S))
hg log -r destination( S ) # { G(S), T(G(S)) }
hg log -r destination( G(S) ) # { T(G(S)) }
The implementation differences between the three different copy commands (see
the origin() predicate) are not intentionally exposed, however if the
transplant was a graft instead:
hg log -r destination( G(S) ) # {}
because the 'extra' field in G(G(S)) is S, not G(S). The implementation cannot
correct this by following sources before G(S) and then select the csets that
reference those sources because the cset provided to the predicate would also
end up selected. If there were more than two copies, sources of the argument
would also get selected.
Note that the convert extension does not currently update the 'extra' map in its
destination csets, and therefore copies made prior to the convert will be
missing from the resulting set.
Instead of the loop over 'subset', the following almost works, but does not
select a transplant of a transplant. That is, 'destination(S)' will only
select T(S).
dests = set([r for r in subset if _getrevsource(repo, r) in args])
2012-07-07 08:47:55 +04:00
|
|
|
"destination": destination,
|
2012-12-12 06:12:55 +04:00
|
|
|
"divergent": divergent,
|
2012-01-06 13:04:20 +04:00
|
|
|
"draft": draft,
|
2012-07-06 21:34:09 +04:00
|
|
|
"extinct": extinct,
|
2012-05-12 12:20:57 +04:00
|
|
|
"extra": extra,
|
2010-06-04 19:27:23 +04:00
|
|
|
"file": hasfile,
|
2011-05-17 02:02:35 +04:00
|
|
|
"filelog": filelog,
|
2011-09-17 21:34:47 +04:00
|
|
|
"first": first,
|
2010-06-04 19:27:23 +04:00
|
|
|
"follow": follow,
|
2012-02-26 01:11:36 +04:00
|
|
|
"_followfirst": _followfirst,
|
2010-06-01 20:18:57 +04:00
|
|
|
"grep": grep,
|
2010-06-04 19:27:23 +04:00
|
|
|
"head": head,
|
|
|
|
"heads": heads,
|
2012-08-04 22:20:48 +04:00
|
|
|
"hidden": hidden,
|
2012-04-14 00:32:49 +04:00
|
|
|
"id": node_,
|
2010-06-04 19:27:23 +04:00
|
|
|
"keyword": keyword,
|
2011-04-30 19:56:43 +04:00
|
|
|
"last": last,
|
2010-06-04 19:27:23 +04:00
|
|
|
"limit": limit,
|
2012-02-23 21:05:20 +04:00
|
|
|
"_matchfiles": _matchfiles,
|
2010-06-04 19:27:23 +04:00
|
|
|
"max": maxrev,
|
|
|
|
"merge": merge,
|
2011-06-17 00:03:26 +04:00
|
|
|
"min": minrev,
|
2010-06-04 19:27:23 +04:00
|
|
|
"modifies": modifies,
|
2012-07-06 21:29:10 +04:00
|
|
|
"obsolete": obsolete,
|
2012-07-07 08:47:30 +04:00
|
|
|
"origin": origin,
|
2010-06-04 19:27:23 +04:00
|
|
|
"outgoing": outgoing,
|
2010-06-01 20:18:57 +04:00
|
|
|
"p1": p1,
|
|
|
|
"p2": p2,
|
|
|
|
"parents": parents,
|
2010-08-13 20:11:41 +04:00
|
|
|
"present": present,
|
2012-01-06 13:04:20 +04:00
|
|
|
"public": public,
|
2012-01-20 00:31:05 +04:00
|
|
|
"remote": remote,
|
2010-06-01 20:18:57 +04:00
|
|
|
"removes": removes,
|
2010-10-11 18:44:19 +04:00
|
|
|
"rev": rev,
|
2011-06-17 00:03:26 +04:00
|
|
|
"reverse": reverse,
|
2010-06-04 19:27:23 +04:00
|
|
|
"roots": roots,
|
|
|
|
"sort": sort,
|
2012-01-06 13:04:20 +04:00
|
|
|
"secret": secret,
|
revset: add "matching" keyword
This keyword can be used to find revisions that "match" one or more fields of a
given set of revisions.
A revision matches another if all the selected fields (description, author,
branch, date, files, phase, parents, substate, user, summary and/or metadata)
match the corresponding values of those fields on the source revision.
By default this keyword looks for revisions that whose metadata match
(description, author and date) making it ideal to look for duplicate revisions.
matching takes 2 arguments (the second being optional):
1.- rev: a revset represeting a _single_ revision (e.g. tip, ., p1(.), etc)
2.- [field(s) to match]: an optional string containing the field or fields
(separated by spaces) to match.
Valid fields are most regular context fields and some special fields:
* regular fields:
- description, author, branch, date, files, phase, parents,
substate, user.
Note that author and user are synonyms.
* special fields: summary, metadata.
- summary: matches the first line of the description.
- metatadata: It is equivalent to matching 'description user date'
(i.e. it matches the main metadata fields).
Examples:
1.- Look for revisions with the same metadata (author, description and date)
as the 11th revision:
hg log -r "matching(11)"
2.- Look for revisions with the same description as the 11th revision:
hg log -r "matching(11, description)"
3.- Look for revisions with the same 'summary' (i.e. same first line on their
description) as the 11th revision:
hg log -r "matching(11, summary)"
4.- Look for revisions with the same author as the current revision:
hg log -r "matching(., author)"
You could use 'user' rather than 'author' to get the same result.
5.- Look for revisions with the same description _AND_ author as the tip of the
repository:
hg log -r "matching(tip, 'author description')"
6.- Look for revisions touching the same files as the parent of the tip of the
repository
hg log -r "matching(p1(tip), files)"
7.- Look for revisions whose subrepos are on the same state as the tip of the
repository or its parent
hg log -r "matching(p1(tip):tip, substate)"
8.- Look for revisions whose author and subrepo states both match those of any
of the revisions on the stable branch:
hg log -r "matching(branch(stable), 'author substate')"
2012-04-01 16:12:14 +04:00
|
|
|
"matching": matching,
|
2010-10-10 21:41:36 +04:00
|
|
|
"tag": tag,
|
2010-10-23 21:21:51 +04:00
|
|
|
"tagged": tagged,
|
|
|
|
"user": user,
|
2012-07-06 02:18:09 +04:00
|
|
|
"unstable": unstable,
|
2012-01-16 11:21:22 +04:00
|
|
|
"_list": _list,
|
2014-02-27 00:36:36 +04:00
|
|
|
"_intlist": _intlist,
|
2014-02-27 05:15:55 +04:00
|
|
|
"_hexlist": _hexlist,
|
2010-06-01 20:18:57 +04:00
|
|
|
}
|
|
|
|
|
2013-09-06 13:30:56 +04:00
|
|
|
# symbols which can't be used for a DoS attack for any given input
|
|
|
|
# (e.g. those which accept regexes as plain strings shouldn't be included)
|
|
|
|
# functions that just return a lot of changesets (like all) don't count here
|
|
|
|
safesymbols = set([
|
|
|
|
"adds",
|
|
|
|
"all",
|
|
|
|
"ancestor",
|
|
|
|
"ancestors",
|
|
|
|
"_firstancestors",
|
|
|
|
"author",
|
|
|
|
"bisect",
|
|
|
|
"bisected",
|
|
|
|
"bookmark",
|
|
|
|
"branch",
|
|
|
|
"branchpoint",
|
|
|
|
"bumped",
|
|
|
|
"bundle",
|
|
|
|
"children",
|
|
|
|
"closed",
|
|
|
|
"converted",
|
|
|
|
"date",
|
|
|
|
"desc",
|
|
|
|
"descendants",
|
|
|
|
"_firstdescendants",
|
|
|
|
"destination",
|
|
|
|
"divergent",
|
|
|
|
"draft",
|
|
|
|
"extinct",
|
|
|
|
"extra",
|
|
|
|
"file",
|
|
|
|
"filelog",
|
|
|
|
"first",
|
|
|
|
"follow",
|
|
|
|
"_followfirst",
|
|
|
|
"head",
|
|
|
|
"heads",
|
|
|
|
"hidden",
|
|
|
|
"id",
|
|
|
|
"keyword",
|
|
|
|
"last",
|
|
|
|
"limit",
|
|
|
|
"_matchfiles",
|
|
|
|
"max",
|
|
|
|
"merge",
|
|
|
|
"min",
|
|
|
|
"modifies",
|
|
|
|
"obsolete",
|
|
|
|
"origin",
|
|
|
|
"outgoing",
|
|
|
|
"p1",
|
|
|
|
"p2",
|
|
|
|
"parents",
|
|
|
|
"present",
|
|
|
|
"public",
|
|
|
|
"remote",
|
|
|
|
"removes",
|
|
|
|
"rev",
|
|
|
|
"reverse",
|
|
|
|
"roots",
|
|
|
|
"sort",
|
|
|
|
"secret",
|
|
|
|
"matching",
|
|
|
|
"tag",
|
|
|
|
"tagged",
|
|
|
|
"user",
|
|
|
|
"unstable",
|
|
|
|
"_list",
|
2014-02-27 00:36:36 +04:00
|
|
|
"_intlist",
|
2014-02-27 05:15:55 +04:00
|
|
|
"_hexlist",
|
2013-09-06 13:30:56 +04:00
|
|
|
])
|
|
|
|
|
2010-06-01 20:18:57 +04:00
|
|
|
methods = {
|
|
|
|
"range": rangeset,
|
2012-06-02 02:50:22 +04:00
|
|
|
"dagrange": dagrange,
|
2010-06-01 20:18:57 +04:00
|
|
|
"string": stringset,
|
|
|
|
"symbol": symbolset,
|
|
|
|
"and": andset,
|
|
|
|
"or": orset,
|
|
|
|
"not": notset,
|
|
|
|
"list": listset,
|
|
|
|
"func": func,
|
2011-04-30 19:43:04 +04:00
|
|
|
"ancestor": ancestorspec,
|
|
|
|
"parent": parentspec,
|
|
|
|
"parentpost": p1,
|
2010-06-01 20:18:57 +04:00
|
|
|
}
|
|
|
|
|
2010-06-04 02:39:34 +04:00
|
|
|
def optimize(x, small):
|
2010-11-22 20:15:58 +03:00
|
|
|
if x is None:
|
2010-06-04 02:39:34 +04:00
|
|
|
return 0, x
|
|
|
|
|
2010-06-01 20:18:57 +04:00
|
|
|
smallbonus = 1
|
|
|
|
if small:
|
|
|
|
smallbonus = .5
|
|
|
|
|
|
|
|
op = x[0]
|
2010-06-04 19:26:55 +04:00
|
|
|
if op == 'minus':
|
2010-06-04 02:39:34 +04:00
|
|
|
return optimize(('and', x[1], ('not', x[2])), small)
|
|
|
|
elif op == 'dagrangepre':
|
|
|
|
return optimize(('func', ('symbol', 'ancestors'), x[1]), small)
|
|
|
|
elif op == 'dagrangepost':
|
|
|
|
return optimize(('func', ('symbol', 'descendants'), x[1]), small)
|
|
|
|
elif op == 'rangepre':
|
|
|
|
return optimize(('range', ('string', '0'), x[1]), small)
|
|
|
|
elif op == 'rangepost':
|
|
|
|
return optimize(('range', x[1], ('string', 'tip')), small)
|
2010-07-01 02:44:36 +04:00
|
|
|
elif op == 'negate':
|
|
|
|
return optimize(('string',
|
|
|
|
'-' + getstring(x[1], _("can't negate that"))), small)
|
2010-06-04 02:39:34 +04:00
|
|
|
elif op in 'string symbol negate':
|
|
|
|
return smallbonus, x # single revisions are small
|
2012-06-02 02:50:22 +04:00
|
|
|
elif op == 'and':
|
2010-06-04 02:39:34 +04:00
|
|
|
wa, ta = optimize(x[1], True)
|
|
|
|
wb, tb = optimize(x[2], True)
|
2014-02-14 02:04:47 +04:00
|
|
|
|
|
|
|
# (::x and not ::y)/(not ::y and ::x) have a fast path
|
2014-07-12 11:31:36 +04:00
|
|
|
def isonly(revs, bases):
|
2014-02-14 02:04:47 +04:00
|
|
|
return (
|
|
|
|
revs[0] == 'func'
|
|
|
|
and getstring(revs[1], _('not a symbol')) == 'ancestors'
|
|
|
|
and bases[0] == 'not'
|
|
|
|
and bases[1][0] == 'func'
|
|
|
|
and getstring(bases[1][1], _('not a symbol')) == 'ancestors')
|
|
|
|
|
2010-06-04 02:39:34 +04:00
|
|
|
w = min(wa, wb)
|
2014-07-12 11:31:36 +04:00
|
|
|
if isonly(ta, tb):
|
|
|
|
return w, ('func', ('symbol', 'only'), ('list', ta[2], tb[1][2]))
|
|
|
|
if isonly(tb, ta):
|
|
|
|
return w, ('func', ('symbol', 'only'), ('list', tb[2], ta[1][2]))
|
2014-02-14 02:04:47 +04:00
|
|
|
|
2010-06-04 02:39:34 +04:00
|
|
|
if wa > wb:
|
|
|
|
return w, (op, tb, ta)
|
|
|
|
return w, (op, ta, tb)
|
|
|
|
elif op == 'or':
|
|
|
|
wa, ta = optimize(x[1], False)
|
|
|
|
wb, tb = optimize(x[2], False)
|
|
|
|
if wb < wa:
|
|
|
|
wb, wa = wa, wb
|
|
|
|
return max(wa, wb), (op, ta, tb)
|
2010-06-01 20:18:57 +04:00
|
|
|
elif op == 'not':
|
2010-06-04 02:39:34 +04:00
|
|
|
o = optimize(x[1], not small)
|
|
|
|
return o[0], (op, o[1])
|
2011-04-30 19:43:04 +04:00
|
|
|
elif op == 'parentpost':
|
|
|
|
o = optimize(x[1], small)
|
|
|
|
return o[0], (op, o[1])
|
2010-06-01 20:18:57 +04:00
|
|
|
elif op == 'group':
|
2010-06-04 02:39:34 +04:00
|
|
|
return optimize(x[1], small)
|
2012-06-02 02:50:22 +04:00
|
|
|
elif op in 'dagrange range list parent ancestorspec':
|
2011-07-06 22:37:50 +04:00
|
|
|
if op == 'parent':
|
|
|
|
# x^:y means (x^) : y, not x ^ (:y)
|
|
|
|
post = ('parentpost', x[1])
|
|
|
|
if x[2][0] == 'dagrangepre':
|
|
|
|
return optimize(('dagrange', post, x[2][1]), small)
|
|
|
|
elif x[2][0] == 'rangepre':
|
|
|
|
return optimize(('range', post, x[2][1]), small)
|
|
|
|
|
2010-06-04 02:39:34 +04:00
|
|
|
wa, ta = optimize(x[1], small)
|
|
|
|
wb, tb = optimize(x[2], small)
|
|
|
|
return wa + wb, (op, ta, tb)
|
2010-06-01 20:18:57 +04:00
|
|
|
elif op == 'func':
|
2010-06-18 23:31:19 +04:00
|
|
|
f = getstring(x[1], _("not a symbol"))
|
2010-06-04 02:39:34 +04:00
|
|
|
wa, ta = optimize(x[2], small)
|
2011-06-17 00:47:34 +04:00
|
|
|
if f in ("author branch closed date desc file grep keyword "
|
|
|
|
"outgoing user"):
|
2010-06-04 02:39:34 +04:00
|
|
|
w = 10 # slow
|
2010-09-21 01:40:36 +04:00
|
|
|
elif f in "modifies adds removes":
|
2010-06-04 02:39:34 +04:00
|
|
|
w = 30 # slower
|
2010-06-01 20:18:57 +04:00
|
|
|
elif f == "contains":
|
2010-06-04 02:39:34 +04:00
|
|
|
w = 100 # very slow
|
2010-06-01 20:18:57 +04:00
|
|
|
elif f == "ancestor":
|
2010-06-04 02:39:34 +04:00
|
|
|
w = 1 * smallbonus
|
2014-09-13 01:21:18 +04:00
|
|
|
elif f in "reverse limit first _intlist":
|
2010-06-04 02:39:34 +04:00
|
|
|
w = 0
|
2010-06-01 20:18:57 +04:00
|
|
|
elif f in "sort":
|
2010-06-04 02:39:34 +04:00
|
|
|
w = 10 # assume most sorts look at changelog
|
2010-06-01 20:18:57 +04:00
|
|
|
else:
|
2010-06-04 02:39:34 +04:00
|
|
|
w = 1
|
|
|
|
return w + wa, (op, x[1], ta)
|
|
|
|
return 1, x
|
2010-06-01 20:18:57 +04:00
|
|
|
|
2012-05-19 19:18:29 +04:00
|
|
|
_aliasarg = ('func', ('symbol', '_aliasarg'))
|
|
|
|
def _getaliasarg(tree):
|
|
|
|
"""If tree matches ('func', ('symbol', '_aliasarg'), ('string', X))
|
|
|
|
return X, None otherwise.
|
|
|
|
"""
|
|
|
|
if (len(tree) == 3 and tree[:2] == _aliasarg
|
|
|
|
and tree[2][0] == 'string'):
|
|
|
|
return tree[2][1]
|
|
|
|
return None
|
|
|
|
|
|
|
|
def _checkaliasarg(tree, known=None):
|
|
|
|
"""Check tree contains no _aliasarg construct or only ones which
|
|
|
|
value is in known. Used to avoid alias placeholders injection.
|
|
|
|
"""
|
|
|
|
if isinstance(tree, tuple):
|
|
|
|
arg = _getaliasarg(tree)
|
|
|
|
if arg is not None and (not known or arg not in known):
|
|
|
|
raise error.ParseError(_("not a function: %s") % '_aliasarg')
|
|
|
|
for t in tree:
|
|
|
|
_checkaliasarg(t, known)
|
|
|
|
|
2011-04-30 20:30:14 +04:00
|
|
|
class revsetalias(object):
|
|
|
|
funcre = re.compile('^([^(]+)\(([^)]+)\)$')
|
2011-06-22 03:55:00 +04:00
|
|
|
args = None
|
2011-04-30 20:30:14 +04:00
|
|
|
|
2011-06-22 03:55:00 +04:00
|
|
|
def __init__(self, name, value):
|
2011-04-30 20:30:14 +04:00
|
|
|
'''Aliases like:
|
|
|
|
|
|
|
|
h = heads(default)
|
|
|
|
b($1) = ancestors($1) - ancestors(default)
|
|
|
|
'''
|
2012-02-10 00:03:07 +04:00
|
|
|
m = self.funcre.search(name)
|
|
|
|
if m:
|
|
|
|
self.name = m.group(1)
|
|
|
|
self.tree = ('func', ('symbol', m.group(1)))
|
|
|
|
self.args = [x.strip() for x in m.group(2).split(',')]
|
|
|
|
for arg in self.args:
|
2012-05-19 19:18:29 +04:00
|
|
|
# _aliasarg() is an unknown symbol only used separate
|
|
|
|
# alias argument placeholders from regular strings.
|
|
|
|
value = value.replace(arg, '_aliasarg(%r)' % (arg,))
|
2012-02-10 00:03:07 +04:00
|
|
|
else:
|
|
|
|
self.name = name
|
|
|
|
self.tree = ('symbol', name)
|
|
|
|
|
|
|
|
self.replacement, pos = parse(value)
|
|
|
|
if pos != len(value):
|
|
|
|
raise error.ParseError(_('invalid token'), pos)
|
2012-05-19 19:18:29 +04:00
|
|
|
# Check for placeholder injection
|
|
|
|
_checkaliasarg(self.replacement, self.args)
|
2012-02-10 00:03:07 +04:00
|
|
|
|
|
|
|
def _getalias(aliases, tree):
|
|
|
|
"""If tree looks like an unexpanded alias, return it. Return None
|
|
|
|
otherwise.
|
|
|
|
"""
|
|
|
|
if isinstance(tree, tuple) and tree:
|
|
|
|
if tree[0] == 'symbol' and len(tree) == 2:
|
|
|
|
name = tree[1]
|
|
|
|
alias = aliases.get(name)
|
|
|
|
if alias and alias.args is None and alias.tree == tree:
|
|
|
|
return alias
|
|
|
|
if tree[0] == 'func' and len(tree) > 1:
|
|
|
|
if tree[1][0] == 'symbol' and len(tree[1]) == 2:
|
|
|
|
name = tree[1][1]
|
|
|
|
alias = aliases.get(name)
|
|
|
|
if alias and alias.args is not None and alias.tree == tree[:2]:
|
|
|
|
return alias
|
|
|
|
return None
|
|
|
|
|
|
|
|
def _expandargs(tree, args):
|
2012-05-19 19:18:29 +04:00
|
|
|
"""Replace _aliasarg instances with the substitution value of the
|
|
|
|
same name in args, recursively.
|
2012-02-10 00:03:07 +04:00
|
|
|
"""
|
2012-05-19 19:18:29 +04:00
|
|
|
if not tree or not isinstance(tree, tuple):
|
2011-04-30 20:30:14 +04:00
|
|
|
return tree
|
2012-05-19 19:18:29 +04:00
|
|
|
arg = _getaliasarg(tree)
|
|
|
|
if arg is not None:
|
|
|
|
return args[arg]
|
2012-02-10 00:03:07 +04:00
|
|
|
return tuple(_expandargs(t, args) for t in tree)
|
|
|
|
|
2012-05-24 15:05:06 +04:00
|
|
|
def _expandaliases(aliases, tree, expanding, cache):
|
2012-02-10 00:03:07 +04:00
|
|
|
"""Expand aliases in tree, recursively.
|
|
|
|
|
|
|
|
'aliases' is a dictionary mapping user defined aliases to
|
|
|
|
revsetalias objects.
|
|
|
|
"""
|
|
|
|
if not isinstance(tree, tuple):
|
|
|
|
# Do not expand raw strings
|
|
|
|
return tree
|
|
|
|
alias = _getalias(aliases, tree)
|
|
|
|
if alias is not None:
|
|
|
|
if alias in expanding:
|
|
|
|
raise error.ParseError(_('infinite expansion of revset alias "%s" '
|
|
|
|
'detected') % alias.name)
|
|
|
|
expanding.append(alias)
|
2012-05-24 15:05:06 +04:00
|
|
|
if alias.name not in cache:
|
|
|
|
cache[alias.name] = _expandaliases(aliases, alias.replacement,
|
|
|
|
expanding, cache)
|
|
|
|
result = cache[alias.name]
|
2012-05-19 19:19:55 +04:00
|
|
|
expanding.pop()
|
2012-02-10 00:03:07 +04:00
|
|
|
if alias.args is not None:
|
|
|
|
l = getlist(tree[2])
|
|
|
|
if len(l) != len(alias.args):
|
|
|
|
raise error.ParseError(
|
|
|
|
_('invalid number of arguments: %s') % len(l))
|
2012-05-24 15:05:06 +04:00
|
|
|
l = [_expandaliases(aliases, a, [], cache) for a in l]
|
2012-02-10 00:03:07 +04:00
|
|
|
result = _expandargs(result, dict(zip(alias.args, l)))
|
|
|
|
else:
|
2012-05-24 15:05:06 +04:00
|
|
|
result = tuple(_expandaliases(aliases, t, expanding, cache)
|
2012-02-10 00:03:07 +04:00
|
|
|
for t in tree)
|
|
|
|
return result
|
2011-04-30 20:30:14 +04:00
|
|
|
|
|
|
|
def findaliases(ui, tree):
|
2012-05-19 19:18:29 +04:00
|
|
|
_checkaliasarg(tree)
|
2012-02-10 00:03:07 +04:00
|
|
|
aliases = {}
|
2011-04-30 20:30:14 +04:00
|
|
|
for k, v in ui.configitems('revsetalias'):
|
|
|
|
alias = revsetalias(k, v)
|
2012-02-10 00:03:07 +04:00
|
|
|
aliases[alias.name] = alias
|
2012-05-24 15:05:06 +04:00
|
|
|
return _expandaliases(aliases, tree, [], {})
|
2011-04-30 20:30:14 +04:00
|
|
|
|
2014-03-19 02:19:44 +04:00
|
|
|
def parse(spec, lookup=None):
|
2013-12-21 07:44:19 +04:00
|
|
|
p = parser.parser(tokenize, elements)
|
2014-03-19 02:19:44 +04:00
|
|
|
return p.parse(spec, lookup=lookup)
|
2010-06-01 20:18:57 +04:00
|
|
|
|
2014-03-19 02:19:44 +04:00
|
|
|
def match(ui, spec, repo=None):
|
2010-06-19 02:34:13 +04:00
|
|
|
if not spec:
|
|
|
|
raise error.ParseError(_("empty query"))
|
2014-03-19 02:19:44 +04:00
|
|
|
lookup = None
|
|
|
|
if repo:
|
|
|
|
lookup = repo.__contains__
|
|
|
|
tree, pos = parse(spec, lookup)
|
2011-03-17 01:09:14 +03:00
|
|
|
if (pos != len(spec)):
|
2011-06-21 02:17:52 +04:00
|
|
|
raise error.ParseError(_("invalid token"), pos)
|
2011-07-21 23:04:57 +04:00
|
|
|
if ui:
|
|
|
|
tree = findaliases(ui, tree)
|
2010-06-04 02:39:34 +04:00
|
|
|
weight, tree = optimize(tree, True)
|
2010-06-01 20:18:57 +04:00
|
|
|
def mfunc(repo, subset):
|
2014-02-19 03:54:46 +04:00
|
|
|
if util.safehasattr(subset, 'set'):
|
|
|
|
return getset(repo, subset, tree)
|
|
|
|
return getset(repo, baseset(subset), tree)
|
2010-06-01 20:18:57 +04:00
|
|
|
return mfunc
|
2010-10-23 21:21:51 +04:00
|
|
|
|
2011-07-21 23:05:45 +04:00
|
|
|
def formatspec(expr, *args):
|
|
|
|
'''
|
|
|
|
This is a convenience function for using revsets internally, and
|
|
|
|
escapes arguments appropriately. Aliases are intentionally ignored
|
|
|
|
so that intended expression behavior isn't accidentally subverted.
|
|
|
|
|
|
|
|
Supported arguments:
|
|
|
|
|
2011-10-15 19:20:08 +04:00
|
|
|
%r = revset expression, parenthesized
|
2011-07-21 23:05:45 +04:00
|
|
|
%d = int(arg), no quoting
|
|
|
|
%s = string(arg), escaped and single-quoted
|
|
|
|
%b = arg.branch(), escaped and single-quoted
|
|
|
|
%n = hex(arg), single-quoted
|
|
|
|
%% = a literal '%'
|
|
|
|
|
2011-10-15 19:20:08 +04:00
|
|
|
Prefixing the type with 'l' specifies a parenthesized list of that type.
|
2011-09-20 01:28:44 +04:00
|
|
|
|
2011-10-15 21:52:43 +04:00
|
|
|
>>> formatspec('%r:: and %lr', '10 or 11', ("this()", "that()"))
|
|
|
|
'(10 or 11):: and ((this()) or (that()))'
|
2011-07-21 23:05:45 +04:00
|
|
|
>>> formatspec('%d:: and not %d::', 10, 20)
|
|
|
|
'10:: and not 20::'
|
2011-10-21 21:12:21 +04:00
|
|
|
>>> formatspec('%ld or %ld', [], [1])
|
2012-01-16 11:21:22 +04:00
|
|
|
"_list('') or 1"
|
2011-07-21 23:05:45 +04:00
|
|
|
>>> formatspec('keyword(%s)', 'foo\\xe9')
|
|
|
|
"keyword('foo\\\\xe9')"
|
|
|
|
>>> b = lambda: 'default'
|
|
|
|
>>> b.branch = b
|
|
|
|
>>> formatspec('branch(%b)', b)
|
|
|
|
"branch('default')"
|
2011-09-20 01:28:44 +04:00
|
|
|
>>> formatspec('root(%ls)', ['a', 'b', 'c', 'd'])
|
2012-01-16 11:21:22 +04:00
|
|
|
"root(_list('a\\x00b\\x00c\\x00d'))"
|
2011-07-21 23:05:45 +04:00
|
|
|
'''
|
|
|
|
|
|
|
|
def quote(s):
|
|
|
|
return repr(str(s))
|
|
|
|
|
2011-09-20 01:28:44 +04:00
|
|
|
def argtype(c, arg):
|
|
|
|
if c == 'd':
|
|
|
|
return str(int(arg))
|
|
|
|
elif c == 's':
|
|
|
|
return quote(arg)
|
2011-10-15 19:20:08 +04:00
|
|
|
elif c == 'r':
|
|
|
|
parse(arg) # make sure syntax errors are confined
|
|
|
|
return '(%s)' % arg
|
2011-09-20 01:28:44 +04:00
|
|
|
elif c == 'n':
|
2012-04-14 00:32:49 +04:00
|
|
|
return quote(node.hex(arg))
|
2011-09-20 01:28:44 +04:00
|
|
|
elif c == 'b':
|
|
|
|
return quote(arg.branch())
|
|
|
|
|
2011-12-01 08:43:24 +04:00
|
|
|
def listexp(s, t):
|
|
|
|
l = len(s)
|
|
|
|
if l == 0:
|
2012-01-16 11:21:22 +04:00
|
|
|
return "_list('')"
|
|
|
|
elif l == 1:
|
2011-12-01 08:43:24 +04:00
|
|
|
return argtype(t, s[0])
|
2012-01-16 11:21:22 +04:00
|
|
|
elif t == 'd':
|
2014-02-27 00:36:36 +04:00
|
|
|
return "_intlist('%s')" % "\0".join(str(int(a)) for a in s)
|
2012-01-16 11:21:22 +04:00
|
|
|
elif t == 's':
|
|
|
|
return "_list('%s')" % "\0".join(s)
|
|
|
|
elif t == 'n':
|
2014-02-27 05:15:55 +04:00
|
|
|
return "_hexlist('%s')" % "\0".join(node.hex(a) for a in s)
|
2012-01-16 11:21:22 +04:00
|
|
|
elif t == 'b':
|
|
|
|
return "_list('%s')" % "\0".join(a.branch() for a in s)
|
|
|
|
|
2012-01-08 21:15:54 +04:00
|
|
|
m = l // 2
|
2011-12-01 08:43:24 +04:00
|
|
|
return '(%s or %s)' % (listexp(s[:m], t), listexp(s[m:], t))
|
|
|
|
|
2011-07-21 23:05:45 +04:00
|
|
|
ret = ''
|
|
|
|
pos = 0
|
|
|
|
arg = 0
|
|
|
|
while pos < len(expr):
|
|
|
|
c = expr[pos]
|
|
|
|
if c == '%':
|
|
|
|
pos += 1
|
|
|
|
d = expr[pos]
|
|
|
|
if d == '%':
|
|
|
|
ret += d
|
2011-10-15 21:52:43 +04:00
|
|
|
elif d in 'dsnbr':
|
2011-09-20 01:28:44 +04:00
|
|
|
ret += argtype(d, args[arg])
|
2011-07-21 23:05:45 +04:00
|
|
|
arg += 1
|
2011-09-20 01:28:44 +04:00
|
|
|
elif d == 'l':
|
|
|
|
# a list of some type
|
|
|
|
pos += 1
|
|
|
|
d = expr[pos]
|
2011-12-01 11:42:03 +04:00
|
|
|
ret += listexp(list(args[arg]), d)
|
2011-07-21 23:05:45 +04:00
|
|
|
arg += 1
|
|
|
|
else:
|
|
|
|
raise util.Abort('unexpected revspec format character %s' % d)
|
|
|
|
else:
|
|
|
|
ret += c
|
|
|
|
pos += 1
|
|
|
|
|
|
|
|
return ret
|
|
|
|
|
debugrevspec: pretty print output
Before:
('func', ('symbol', 'reverse'), ('func', ('symbol', 'sort'), ('list', ('or',
('symbol', '2'), ('symbol', '3')), ('symbol', 'date'))))
After:
(func
('symbol', 'reverse')
(func
('symbol', 'sort')
(list
(or
('symbol', '2')
('symbol', '3'))
('symbol', 'date'))))
v2:
- Rebased on stable to avoid having to merge tests output
2012-02-24 14:02:21 +04:00
|
|
|
def prettyformat(tree):
|
|
|
|
def _prettyformat(tree, level, lines):
|
|
|
|
if not isinstance(tree, tuple) or tree[0] in ('string', 'symbol'):
|
|
|
|
lines.append((level, str(tree)))
|
|
|
|
else:
|
|
|
|
lines.append((level, '(%s' % tree[0]))
|
|
|
|
for s in tree[1:]:
|
|
|
|
_prettyformat(s, level + 1, lines)
|
|
|
|
lines[-1:] = [(lines[-1][0], lines[-1][1] + ')')]
|
|
|
|
|
|
|
|
lines = []
|
|
|
|
_prettyformat(tree, 0, lines)
|
|
|
|
output = '\n'.join((' '*l + s) for l, s in lines)
|
|
|
|
return output
|
|
|
|
|
2013-08-09 22:52:58 +04:00
|
|
|
def depth(tree):
|
|
|
|
if isinstance(tree, tuple):
|
|
|
|
return max(map(depth, tree)) + 1
|
|
|
|
else:
|
|
|
|
return 0
|
|
|
|
|
2013-08-07 01:21:31 +04:00
|
|
|
def funcsused(tree):
|
|
|
|
if not isinstance(tree, tuple) or tree[0] in ('string', 'symbol'):
|
|
|
|
return set()
|
|
|
|
else:
|
|
|
|
funcs = set()
|
|
|
|
for s in tree[1:]:
|
|
|
|
funcs |= funcsused(s)
|
|
|
|
if tree[0] == 'func':
|
|
|
|
funcs.add(tree[1][1])
|
|
|
|
return funcs
|
|
|
|
|
2014-01-21 23:39:26 +04:00
|
|
|
class baseset(list):
|
2014-02-06 23:33:36 +04:00
|
|
|
"""Basic data structure that represents a revset and contains the basic
|
|
|
|
operation that it should be able to perform.
|
2014-03-14 20:18:14 +04:00
|
|
|
|
|
|
|
Every method in this class should be implemented by any smartset class.
|
2014-02-06 23:33:36 +04:00
|
|
|
"""
|
2014-03-14 22:41:26 +04:00
|
|
|
def __init__(self, data=()):
|
2014-01-22 22:46:02 +04:00
|
|
|
super(baseset, self).__init__(data)
|
|
|
|
self._set = None
|
|
|
|
|
2014-02-07 20:44:18 +04:00
|
|
|
def ascending(self):
|
2014-03-14 20:18:14 +04:00
|
|
|
"""Sorts the set in ascending order (in place).
|
|
|
|
|
|
|
|
This is part of the mandatory API for smartset."""
|
2014-02-07 20:44:18 +04:00
|
|
|
self.sort()
|
|
|
|
|
|
|
|
def descending(self):
|
2014-03-14 20:18:14 +04:00
|
|
|
"""Sorts the set in descending order (in place).
|
|
|
|
|
|
|
|
This is part of the mandatory API for smartset."""
|
2014-02-07 20:44:18 +04:00
|
|
|
self.sort(reverse=True)
|
|
|
|
|
2014-02-19 21:28:17 +04:00
|
|
|
def min(self):
|
|
|
|
return min(self)
|
|
|
|
|
|
|
|
def max(self):
|
|
|
|
return max(self)
|
|
|
|
|
2014-01-22 22:46:02 +04:00
|
|
|
def set(self):
|
2014-03-14 20:18:14 +04:00
|
|
|
"""Returns a set or a smartset containing all the elements.
|
|
|
|
|
|
|
|
The returned structure should be the fastest option for membership
|
|
|
|
testing.
|
|
|
|
|
|
|
|
This is part of the mandatory API for smartset."""
|
2014-01-22 22:46:02 +04:00
|
|
|
if not self._set:
|
|
|
|
self._set = set(self)
|
|
|
|
return self._set
|
2014-01-21 23:39:26 +04:00
|
|
|
|
2014-09-17 10:59:29 +04:00
|
|
|
@util.propertycache
|
|
|
|
def __contains__(self):
|
|
|
|
return self.set().__contains__
|
|
|
|
|
2014-03-14 21:10:18 +04:00
|
|
|
def __sub__(self, other):
|
2014-03-14 20:18:14 +04:00
|
|
|
"""Returns a new object with the substraction of the two collections.
|
|
|
|
|
|
|
|
This is part of the mandatory API for smartset."""
|
2014-07-24 23:12:12 +04:00
|
|
|
# If we are operating on 2 baseset, do the computation now since all
|
|
|
|
# data is available. The alternative is to involve a lazyset, which
|
|
|
|
# may be slow.
|
|
|
|
if isinstance(other, baseset):
|
|
|
|
other = other.set()
|
|
|
|
return baseset([x for x in self if x not in other])
|
|
|
|
|
2014-07-15 02:55:31 +04:00
|
|
|
return self.filter(lambda x: x not in other)
|
2014-01-24 02:20:58 +04:00
|
|
|
|
2014-03-14 21:10:18 +04:00
|
|
|
def __and__(self, other):
|
2014-03-14 20:18:14 +04:00
|
|
|
"""Returns a new object with the intersection of the two collections.
|
2014-01-25 04:57:44 +04:00
|
|
|
|
2014-03-14 20:18:14 +04:00
|
|
|
This is part of the mandatory API for smartset."""
|
2014-03-14 21:10:18 +04:00
|
|
|
if isinstance(other, baseset):
|
|
|
|
other = other.set()
|
|
|
|
return baseset([y for y in self if y in other])
|
2014-03-14 20:18:14 +04:00
|
|
|
|
2014-03-14 21:10:18 +04:00
|
|
|
def __add__(self, other):
|
2014-03-14 20:18:14 +04:00
|
|
|
"""Returns a new object with the union of the two collections.
|
|
|
|
|
|
|
|
This is part of the mandatory API for smartset."""
|
2014-02-06 23:37:16 +04:00
|
|
|
s = self.set()
|
2014-03-14 21:10:18 +04:00
|
|
|
l = [r for r in other if r not in s]
|
2014-02-06 23:37:16 +04:00
|
|
|
return baseset(list(self) + l)
|
|
|
|
|
2014-03-12 04:09:23 +04:00
|
|
|
def isascending(self):
|
2014-03-14 20:18:14 +04:00
|
|
|
"""Returns True if the collection is ascending order, False if not.
|
|
|
|
|
|
|
|
This is part of the mandatory API for smartset."""
|
2014-03-12 04:09:23 +04:00
|
|
|
return False
|
|
|
|
|
|
|
|
def isdescending(self):
|
2014-03-14 20:18:14 +04:00
|
|
|
"""Returns True if the collection is descending order, False if not.
|
|
|
|
|
|
|
|
This is part of the mandatory API for smartset."""
|
2014-03-12 04:09:23 +04:00
|
|
|
return False
|
|
|
|
|
2014-03-14 21:10:18 +04:00
|
|
|
def filter(self, condition):
|
2014-03-14 20:18:14 +04:00
|
|
|
"""Returns this smartset filtered by condition as a new smartset.
|
|
|
|
|
|
|
|
`condition` is a callable which takes a revision number and returns a
|
|
|
|
boolean.
|
|
|
|
|
|
|
|
This is part of the mandatory API for smartset."""
|
2014-03-14 21:10:18 +04:00
|
|
|
return lazyset(self, condition)
|
2014-02-07 05:18:11 +04:00
|
|
|
|
2014-03-13 03:40:18 +04:00
|
|
|
class _orderedsetmixin(object):
|
|
|
|
"""Mixin class with utility methods for smartsets
|
|
|
|
|
|
|
|
This should be extended by smartsets which have the isascending(),
|
|
|
|
isdescending() and reverse() methods"""
|
|
|
|
|
|
|
|
def _first(self):
|
|
|
|
"""return the first revision in the set"""
|
|
|
|
for r in self:
|
|
|
|
return r
|
2014-03-29 04:00:13 +04:00
|
|
|
raise ValueError('arg is an empty sequence')
|
2014-03-13 03:40:18 +04:00
|
|
|
|
|
|
|
def _last(self):
|
|
|
|
"""return the last revision in the set"""
|
|
|
|
self.reverse()
|
|
|
|
m = self._first()
|
|
|
|
self.reverse()
|
|
|
|
return m
|
|
|
|
|
|
|
|
def min(self):
|
|
|
|
"""return the smallest element in the set"""
|
|
|
|
if self.isascending():
|
|
|
|
return self._first()
|
|
|
|
return self._last()
|
|
|
|
|
|
|
|
def max(self):
|
|
|
|
"""return the largest element in the set"""
|
|
|
|
if self.isascending():
|
|
|
|
return self._last()
|
|
|
|
return self._first()
|
|
|
|
|
2014-02-07 02:19:40 +04:00
|
|
|
class lazyset(object):
|
|
|
|
"""Duck type for baseset class which iterates lazily over the revisions in
|
|
|
|
the subset and contains a function which tests for membership in the
|
|
|
|
revset
|
|
|
|
"""
|
2014-02-13 21:00:25 +04:00
|
|
|
def __init__(self, subset, condition=lambda x: True):
|
2014-03-14 21:55:03 +04:00
|
|
|
"""
|
|
|
|
condition: a function that decide whether a revision in the subset
|
|
|
|
belongs to the revset or not.
|
|
|
|
"""
|
2014-02-07 02:19:40 +04:00
|
|
|
self._subset = subset
|
|
|
|
self._condition = condition
|
2014-02-05 03:31:57 +04:00
|
|
|
self._cache = {}
|
2014-02-07 02:19:40 +04:00
|
|
|
|
2014-02-07 20:44:18 +04:00
|
|
|
def ascending(self):
|
|
|
|
self._subset.sort()
|
|
|
|
|
|
|
|
def descending(self):
|
|
|
|
self._subset.sort(reverse=True)
|
|
|
|
|
2014-02-19 21:28:17 +04:00
|
|
|
def min(self):
|
|
|
|
return min(self)
|
|
|
|
|
|
|
|
def max(self):
|
|
|
|
return max(self)
|
|
|
|
|
2014-02-07 02:19:40 +04:00
|
|
|
def __contains__(self, x):
|
2014-02-05 03:31:57 +04:00
|
|
|
c = self._cache
|
|
|
|
if x not in c:
|
2014-04-26 01:51:24 +04:00
|
|
|
v = c[x] = x in self._subset and self._condition(x)
|
|
|
|
return v
|
2014-02-05 03:31:57 +04:00
|
|
|
return c[x]
|
2014-02-07 02:19:40 +04:00
|
|
|
|
|
|
|
def __iter__(self):
|
|
|
|
cond = self._condition
|
|
|
|
for x in self._subset:
|
|
|
|
if cond(x):
|
|
|
|
yield x
|
|
|
|
|
2014-02-07 02:25:37 +04:00
|
|
|
def __and__(self, x):
|
2014-05-01 23:15:00 +04:00
|
|
|
return lazyset(self, x.__contains__)
|
2014-02-07 02:25:37 +04:00
|
|
|
|
|
|
|
def __sub__(self, x):
|
|
|
|
return lazyset(self, lambda r: r not in x)
|
|
|
|
|
|
|
|
def __add__(self, x):
|
2014-03-12 04:25:53 +04:00
|
|
|
return _addset(self, x)
|
2014-02-07 02:25:37 +04:00
|
|
|
|
2014-02-20 22:15:38 +04:00
|
|
|
def __nonzero__(self):
|
|
|
|
for r in self:
|
|
|
|
return True
|
|
|
|
return False
|
|
|
|
|
2014-02-07 02:29:37 +04:00
|
|
|
def __len__(self):
|
|
|
|
# Basic implementation to be changed in future patches.
|
|
|
|
l = baseset([r for r in self])
|
|
|
|
return len(l)
|
|
|
|
|
|
|
|
def __getitem__(self, x):
|
|
|
|
# Basic implementation to be changed in future patches.
|
|
|
|
l = baseset([r for r in self])
|
|
|
|
return l[x]
|
|
|
|
|
|
|
|
def sort(self, reverse=False):
|
2014-03-06 21:41:47 +04:00
|
|
|
if not util.safehasattr(self._subset, 'sort'):
|
|
|
|
self._subset = baseset(self._subset)
|
2014-02-07 02:29:37 +04:00
|
|
|
self._subset.sort(reverse=reverse)
|
|
|
|
|
|
|
|
def reverse(self):
|
|
|
|
self._subset.reverse()
|
|
|
|
|
|
|
|
def set(self):
|
|
|
|
return set([r for r in self])
|
|
|
|
|
2014-03-12 04:09:23 +04:00
|
|
|
def isascending(self):
|
|
|
|
return False
|
|
|
|
|
|
|
|
def isdescending(self):
|
|
|
|
return False
|
|
|
|
|
2014-02-07 05:18:11 +04:00
|
|
|
def filter(self, l):
|
|
|
|
return lazyset(self, l)
|
|
|
|
|
2014-03-13 22:36:45 +04:00
|
|
|
class orderedlazyset(_orderedsetmixin, lazyset):
|
2014-02-06 03:24:08 +04:00
|
|
|
"""Subclass of lazyset which subset can be ordered either ascending or
|
|
|
|
descendingly
|
|
|
|
"""
|
|
|
|
def __init__(self, subset, condition, ascending=True):
|
|
|
|
super(orderedlazyset, self).__init__(subset, condition)
|
|
|
|
self._ascending = ascending
|
|
|
|
|
2014-02-07 05:18:11 +04:00
|
|
|
def filter(self, l):
|
|
|
|
return orderedlazyset(self, l, ascending=self._ascending)
|
|
|
|
|
2014-02-07 20:44:18 +04:00
|
|
|
def ascending(self):
|
|
|
|
if not self._ascending:
|
|
|
|
self.reverse()
|
|
|
|
|
|
|
|
def descending(self):
|
|
|
|
if self._ascending:
|
|
|
|
self.reverse()
|
|
|
|
|
2014-02-07 05:42:08 +04:00
|
|
|
def __and__(self, x):
|
2014-05-01 23:15:28 +04:00
|
|
|
return orderedlazyset(self, x.__contains__,
|
2014-02-07 05:42:08 +04:00
|
|
|
ascending=self._ascending)
|
|
|
|
|
|
|
|
def __sub__(self, x):
|
|
|
|
return orderedlazyset(self, lambda r: r not in x,
|
|
|
|
ascending=self._ascending)
|
|
|
|
|
2014-03-12 04:25:53 +04:00
|
|
|
def __add__(self, x):
|
|
|
|
kwargs = {}
|
|
|
|
if self.isascending() and x.isascending():
|
|
|
|
kwargs['ascending'] = True
|
|
|
|
if self.isdescending() and x.isdescending():
|
|
|
|
kwargs['ascending'] = False
|
|
|
|
return _addset(self, x, **kwargs)
|
|
|
|
|
2014-02-25 22:36:23 +04:00
|
|
|
def sort(self, reverse=False):
|
|
|
|
if reverse:
|
|
|
|
if self._ascending:
|
|
|
|
self._subset.sort(reverse=reverse)
|
|
|
|
else:
|
|
|
|
if not self._ascending:
|
|
|
|
self._subset.sort(reverse=reverse)
|
|
|
|
self._ascending = not reverse
|
|
|
|
|
2014-03-12 04:09:23 +04:00
|
|
|
def isascending(self):
|
|
|
|
return self._ascending
|
|
|
|
|
|
|
|
def isdescending(self):
|
|
|
|
return not self._ascending
|
|
|
|
|
2014-02-07 20:44:18 +04:00
|
|
|
def reverse(self):
|
|
|
|
self._subset.reverse()
|
|
|
|
self._ascending = not self._ascending
|
|
|
|
|
2014-03-15 01:43:44 +04:00
|
|
|
class _addset(_orderedsetmixin):
|
2014-03-13 04:20:26 +04:00
|
|
|
"""Represent the addition of two sets
|
|
|
|
|
|
|
|
Wrapper structure for lazily adding two structures without losing much
|
2014-03-08 01:48:31 +04:00
|
|
|
performance on the __contains__ method
|
2014-03-13 04:20:26 +04:00
|
|
|
|
2014-03-12 03:52:15 +04:00
|
|
|
If the ascending attribute is set, that means the two structures are
|
|
|
|
ordered in either an ascending or descending way. Therefore, we can add
|
2014-04-13 21:01:00 +04:00
|
|
|
them maintaining the order by iterating over both at the same time
|
2014-03-12 03:52:15 +04:00
|
|
|
|
2014-03-13 04:20:26 +04:00
|
|
|
This class does not duck-type baseset and it's only supposed to be used
|
|
|
|
internally
|
2014-03-08 01:48:31 +04:00
|
|
|
"""
|
2014-03-12 03:52:15 +04:00
|
|
|
def __init__(self, revs1, revs2, ascending=None):
|
2014-03-08 01:48:31 +04:00
|
|
|
self._r1 = revs1
|
|
|
|
self._r2 = revs2
|
|
|
|
self._iter = None
|
2014-03-12 03:52:15 +04:00
|
|
|
self._ascending = ascending
|
2014-03-12 03:59:42 +04:00
|
|
|
self._genlist = None
|
|
|
|
|
2014-03-21 05:55:28 +04:00
|
|
|
def __len__(self):
|
|
|
|
return len(self._list)
|
|
|
|
|
2014-03-12 03:59:42 +04:00
|
|
|
@util.propertycache
|
|
|
|
def _list(self):
|
|
|
|
if not self._genlist:
|
|
|
|
self._genlist = baseset(self._iterator())
|
|
|
|
return self._genlist
|
2014-03-08 01:48:31 +04:00
|
|
|
|
2014-03-14 06:12:36 +04:00
|
|
|
def filter(self, condition):
|
|
|
|
if self._ascending is not None:
|
|
|
|
return orderedlazyset(self, condition, ascending=self._ascending)
|
|
|
|
return lazyset(self, condition)
|
|
|
|
|
2014-03-14 21:21:56 +04:00
|
|
|
def ascending(self):
|
|
|
|
if self._ascending is None:
|
|
|
|
self.sort()
|
|
|
|
self._ascending = True
|
|
|
|
else:
|
|
|
|
if not self._ascending:
|
|
|
|
self.reverse()
|
|
|
|
|
|
|
|
def descending(self):
|
|
|
|
if self._ascending is None:
|
|
|
|
self.sort(reverse=True)
|
|
|
|
self._ascending = False
|
|
|
|
else:
|
|
|
|
if self._ascending:
|
|
|
|
self.reverse()
|
|
|
|
|
2014-03-14 21:22:29 +04:00
|
|
|
def __and__(self, other):
|
|
|
|
filterfunc = other.__contains__
|
|
|
|
if self._ascending is not None:
|
2014-03-14 21:22:51 +04:00
|
|
|
return orderedlazyset(self, filterfunc, ascending=self._ascending)
|
|
|
|
return lazyset(self, filterfunc)
|
|
|
|
|
|
|
|
def __sub__(self, other):
|
|
|
|
filterfunc = lambda r: r not in other
|
|
|
|
if self._ascending is not None:
|
2014-03-14 21:22:29 +04:00
|
|
|
return orderedlazyset(self, filterfunc, ascending=self._ascending)
|
|
|
|
return lazyset(self, filterfunc)
|
|
|
|
|
2014-03-14 21:23:54 +04:00
|
|
|
def __add__(self, other):
|
|
|
|
"""When both collections are ascending or descending, preserve the order
|
|
|
|
"""
|
|
|
|
kwargs = {}
|
|
|
|
if self._ascending is not None:
|
|
|
|
if self.isascending() and other.isascending():
|
|
|
|
kwargs['ascending'] = True
|
|
|
|
if self.isdescending() and other.isdescending():
|
|
|
|
kwargs['ascending'] = False
|
|
|
|
return _addset(self, other, **kwargs)
|
|
|
|
|
2014-03-08 01:48:31 +04:00
|
|
|
def _iterator(self):
|
2014-03-14 00:29:04 +04:00
|
|
|
"""Iterate over both collections without repeating elements
|
|
|
|
|
|
|
|
If the ascending attribute is not set, iterate over the first one and
|
|
|
|
then over the second one checking for membership on the first one so we
|
|
|
|
dont yield any duplicates.
|
|
|
|
|
|
|
|
If the ascending attribute is set, iterate over both collections at the
|
|
|
|
same time, yielding only one value at a time in the given order.
|
|
|
|
"""
|
2014-03-08 01:48:31 +04:00
|
|
|
if not self._iter:
|
|
|
|
def gen():
|
2014-03-14 00:29:04 +04:00
|
|
|
if self._ascending is None:
|
|
|
|
for r in self._r1:
|
2014-03-08 01:48:31 +04:00
|
|
|
yield r
|
2014-03-14 00:29:04 +04:00
|
|
|
s = self._r1.set()
|
|
|
|
for r in self._r2:
|
|
|
|
if r not in s:
|
|
|
|
yield r
|
|
|
|
else:
|
|
|
|
iter1 = iter(self._r1)
|
|
|
|
iter2 = iter(self._r2)
|
|
|
|
|
|
|
|
val1 = None
|
|
|
|
val2 = None
|
|
|
|
|
|
|
|
choice = max
|
|
|
|
if self._ascending:
|
|
|
|
choice = min
|
|
|
|
try:
|
|
|
|
# Consume both iterators in an ordered way until one is
|
|
|
|
# empty
|
|
|
|
while True:
|
|
|
|
if val1 is None:
|
|
|
|
val1 = iter1.next()
|
|
|
|
if val2 is None:
|
|
|
|
val2 = iter2.next()
|
|
|
|
next = choice(val1, val2)
|
|
|
|
yield next
|
|
|
|
if val1 == next:
|
|
|
|
val1 = None
|
|
|
|
if val2 == next:
|
|
|
|
val2 = None
|
|
|
|
except StopIteration:
|
|
|
|
# Flush any remaining values and consume the other one
|
|
|
|
it = iter2
|
|
|
|
if val1 is not None:
|
|
|
|
yield val1
|
|
|
|
it = iter1
|
|
|
|
elif val2 is not None:
|
|
|
|
# might have been equality and both are empty
|
|
|
|
yield val2
|
|
|
|
for val in it:
|
|
|
|
yield val
|
|
|
|
|
2014-03-13 04:07:38 +04:00
|
|
|
self._iter = _generatorset(gen())
|
2014-03-08 01:48:31 +04:00
|
|
|
|
|
|
|
return self._iter
|
|
|
|
|
|
|
|
def __iter__(self):
|
2014-03-14 01:51:04 +04:00
|
|
|
if self._genlist:
|
|
|
|
return iter(self._genlist)
|
|
|
|
return iter(self._iterator())
|
2014-03-08 01:48:31 +04:00
|
|
|
|
|
|
|
def __contains__(self, x):
|
|
|
|
return x in self._r1 or x in self._r2
|
|
|
|
|
2014-03-10 21:49:04 +04:00
|
|
|
def set(self):
|
|
|
|
return self
|
|
|
|
|
2014-03-12 04:03:43 +04:00
|
|
|
def sort(self, reverse=False):
|
|
|
|
"""Sort the added set
|
|
|
|
|
|
|
|
For this we use the cached list with all the generated values and if we
|
|
|
|
know they are ascending or descending we can sort them in a smart way.
|
|
|
|
"""
|
|
|
|
if self._ascending is None:
|
|
|
|
self._list.sort(reverse=reverse)
|
|
|
|
self._ascending = not reverse
|
|
|
|
else:
|
|
|
|
if bool(self._ascending) == bool(reverse):
|
|
|
|
self.reverse()
|
|
|
|
|
2014-03-14 21:24:09 +04:00
|
|
|
def isascending(self):
|
|
|
|
return self._ascending is not None and self._ascending
|
|
|
|
|
|
|
|
def isdescending(self):
|
|
|
|
return self._ascending is not None and not self._ascending
|
|
|
|
|
2014-03-14 05:57:30 +04:00
|
|
|
def reverse(self):
|
|
|
|
self._list.reverse()
|
|
|
|
if self._ascending is not None:
|
|
|
|
self._ascending = not self._ascending
|
|
|
|
|
2014-03-13 04:07:38 +04:00
|
|
|
class _generatorset(object):
|
|
|
|
"""Wrap a generator for lazy iteration
|
|
|
|
|
|
|
|
Wrapper structure for generators that provides lazy membership and can
|
2014-02-13 06:45:49 +04:00
|
|
|
be iterated more than once.
|
|
|
|
When asked for membership it generates values until either it finds the
|
|
|
|
requested one or has gone through all the elements in the generator
|
2014-03-13 04:07:38 +04:00
|
|
|
|
|
|
|
This class does not duck-type baseset and it's only supposed to be used
|
|
|
|
internally
|
2014-02-13 06:45:49 +04:00
|
|
|
"""
|
2014-02-06 03:23:11 +04:00
|
|
|
def __init__(self, gen):
|
2014-03-14 21:57:04 +04:00
|
|
|
"""
|
|
|
|
gen: a generator producing the values for the generatorset.
|
|
|
|
"""
|
2014-02-06 03:23:11 +04:00
|
|
|
self._gen = gen
|
|
|
|
self._cache = {}
|
2014-02-13 06:45:49 +04:00
|
|
|
self._genlist = baseset([])
|
2014-02-25 04:36:17 +04:00
|
|
|
self._finished = False
|
2014-02-13 06:45:49 +04:00
|
|
|
|
2014-02-06 03:23:11 +04:00
|
|
|
def __contains__(self, x):
|
|
|
|
if x in self._cache:
|
|
|
|
return self._cache[x]
|
|
|
|
|
2014-03-25 07:00:18 +04:00
|
|
|
# Use new values only, as existing values would be cached.
|
|
|
|
for l in self._consumegen():
|
2014-03-04 00:54:46 +04:00
|
|
|
if l == x:
|
|
|
|
return True
|
2014-02-06 03:23:11 +04:00
|
|
|
|
|
|
|
self._cache[x] = False
|
|
|
|
return False
|
|
|
|
|
|
|
|
def __iter__(self):
|
2014-03-26 03:10:07 +04:00
|
|
|
if self._finished:
|
|
|
|
for x in self._genlist:
|
|
|
|
yield x
|
|
|
|
return
|
2014-03-25 07:00:18 +04:00
|
|
|
|
2014-09-17 10:42:41 +04:00
|
|
|
# We have to use this complex iteration strategy to allow multiple
|
|
|
|
# iterations at the same time. We need to be able to catch revision
|
|
|
|
# removed from `consumegen` and added to genlist in another instance.
|
|
|
|
#
|
|
|
|
# Getting rid of it would provide an about 15% speed up on this
|
|
|
|
# iteration.
|
2014-03-26 03:10:07 +04:00
|
|
|
i = 0
|
|
|
|
genlist = self._genlist
|
2014-09-19 02:52:45 +04:00
|
|
|
nextrev = self._consumegen().next
|
|
|
|
_len = len # cache global lookup
|
2014-03-26 03:10:07 +04:00
|
|
|
while True:
|
2014-09-19 02:52:45 +04:00
|
|
|
if i < _len(genlist):
|
2014-03-26 03:10:07 +04:00
|
|
|
yield genlist[i]
|
|
|
|
else:
|
2014-09-19 02:52:45 +04:00
|
|
|
yield nextrev()
|
2014-03-26 03:10:07 +04:00
|
|
|
i += 1
|
2014-03-25 07:00:18 +04:00
|
|
|
|
|
|
|
def _consumegen(self):
|
2014-05-01 03:56:48 +04:00
|
|
|
cache = self._cache
|
|
|
|
genlist = self._genlist.append
|
2014-03-04 00:54:46 +04:00
|
|
|
for item in self._gen:
|
2014-05-01 03:56:48 +04:00
|
|
|
cache[item] = True
|
|
|
|
genlist(item)
|
2014-03-04 00:54:46 +04:00
|
|
|
yield item
|
2014-02-25 04:36:17 +04:00
|
|
|
self._finished = True
|
|
|
|
|
2014-02-06 03:23:11 +04:00
|
|
|
def set(self):
|
|
|
|
return self
|
|
|
|
|
2014-02-25 04:36:17 +04:00
|
|
|
def sort(self, reverse=False):
|
|
|
|
if not self._finished:
|
|
|
|
for i in self:
|
|
|
|
continue
|
|
|
|
self._genlist.sort(reverse=reverse)
|
|
|
|
|
2014-03-13 04:18:54 +04:00
|
|
|
class _ascgeneratorset(_generatorset):
|
|
|
|
"""Wrap a generator of ascending elements for lazy iteration
|
|
|
|
|
|
|
|
Same structure as _generatorset but stops iterating after it goes past
|
2014-02-28 05:27:03 +04:00
|
|
|
the value when asked for membership and the element is not contained
|
2014-03-13 04:18:54 +04:00
|
|
|
|
|
|
|
This class does not duck-type baseset and it's only supposed to be used
|
|
|
|
internally
|
2014-02-28 05:27:03 +04:00
|
|
|
"""
|
|
|
|
def __contains__(self, x):
|
|
|
|
if x in self._cache:
|
|
|
|
return self._cache[x]
|
|
|
|
|
2014-03-25 07:00:18 +04:00
|
|
|
# Use new values only, as existing values would be cached.
|
|
|
|
for l in self._consumegen():
|
2014-02-28 05:27:03 +04:00
|
|
|
if l == x:
|
|
|
|
return True
|
|
|
|
if l > x:
|
|
|
|
break
|
|
|
|
|
|
|
|
self._cache[x] = False
|
|
|
|
return False
|
|
|
|
|
2014-03-13 04:19:46 +04:00
|
|
|
class _descgeneratorset(_generatorset):
|
|
|
|
"""Wrap a generator of descending elements for lazy iteration
|
|
|
|
|
|
|
|
Same structure as _generatorset but stops iterating after it goes past
|
2014-02-28 05:27:03 +04:00
|
|
|
the value when asked for membership and the element is not contained
|
2014-03-13 04:19:46 +04:00
|
|
|
|
|
|
|
This class does not duck-type baseset and it's only supposed to be used
|
|
|
|
internally
|
2014-02-28 05:27:03 +04:00
|
|
|
"""
|
|
|
|
def __contains__(self, x):
|
|
|
|
if x in self._cache:
|
|
|
|
return self._cache[x]
|
|
|
|
|
2014-03-25 07:00:18 +04:00
|
|
|
# Use new values only, as existing values would be cached.
|
|
|
|
for l in self._consumegen():
|
2014-02-28 05:27:03 +04:00
|
|
|
if l == x:
|
|
|
|
return True
|
|
|
|
if l < x:
|
|
|
|
break
|
|
|
|
|
|
|
|
self._cache[x] = False
|
|
|
|
return False
|
|
|
|
|
2014-09-19 00:04:02 +04:00
|
|
|
def spanset(repo, start=None, end=None):
|
|
|
|
"""factory function to dispatch between fullreposet and actual spanset
|
|
|
|
|
|
|
|
Feel free to update all spanset call sites and kill this function at some
|
|
|
|
point.
|
|
|
|
"""
|
|
|
|
if start is None and end is None:
|
|
|
|
return fullreposet(repo)
|
|
|
|
return _spanset(repo, start, end)
|
|
|
|
|
|
|
|
|
|
|
|
class _spanset(_orderedsetmixin):
|
2014-02-12 22:16:21 +04:00
|
|
|
"""Duck type for baseset class which represents a range of revisions and
|
|
|
|
can work lazily and without having all the range in memory
|
2014-03-14 21:59:51 +04:00
|
|
|
|
|
|
|
Note that spanset(x, y) behave almost like xrange(x, y) except for two
|
|
|
|
notable points:
|
|
|
|
- when x < y it will be automatically descending,
|
|
|
|
- revision filtered with this repoview will be skipped.
|
|
|
|
|
2014-02-12 22:16:21 +04:00
|
|
|
"""
|
2014-02-18 23:38:03 +04:00
|
|
|
def __init__(self, repo, start=0, end=None):
|
2014-03-14 21:59:51 +04:00
|
|
|
"""
|
|
|
|
start: first revision included the set
|
|
|
|
(default to 0)
|
|
|
|
end: first revision excluded (last+1)
|
|
|
|
(default to len(repo)
|
|
|
|
|
|
|
|
Spanset will be descending if `end` < `start`.
|
|
|
|
"""
|
2014-02-12 22:16:21 +04:00
|
|
|
self._start = start
|
2014-02-18 23:38:03 +04:00
|
|
|
if end is not None:
|
|
|
|
self._end = end
|
|
|
|
else:
|
|
|
|
self._end = len(repo)
|
|
|
|
self._hiddenrevs = repo.changelog.filteredrevs
|
2014-02-11 05:38:43 +04:00
|
|
|
|
2014-02-07 20:44:18 +04:00
|
|
|
def ascending(self):
|
2014-09-17 10:37:03 +04:00
|
|
|
if not self.isascending():
|
2014-02-07 20:44:18 +04:00
|
|
|
self.reverse()
|
|
|
|
|
|
|
|
def descending(self):
|
2014-09-17 10:37:03 +04:00
|
|
|
if not self.isdescending():
|
2014-02-07 20:44:18 +04:00
|
|
|
self.reverse()
|
|
|
|
|
2014-02-12 22:16:21 +04:00
|
|
|
def __iter__(self):
|
2014-09-17 10:37:03 +04:00
|
|
|
if self.isascending():
|
2014-02-11 05:38:43 +04:00
|
|
|
iterrange = xrange(self._start, self._end)
|
2014-02-12 22:16:21 +04:00
|
|
|
else:
|
2014-02-11 05:38:43 +04:00
|
|
|
iterrange = xrange(self._start, self._end, -1)
|
|
|
|
|
|
|
|
if self._hiddenrevs:
|
|
|
|
s = self._hiddenrevs
|
|
|
|
for r in iterrange:
|
|
|
|
if r not in s:
|
|
|
|
yield r
|
|
|
|
else:
|
|
|
|
for r in iterrange:
|
2014-02-12 22:16:21 +04:00
|
|
|
yield r
|
|
|
|
|
2014-04-29 03:28:52 +04:00
|
|
|
def __contains__(self, rev):
|
2014-04-26 04:53:58 +04:00
|
|
|
start = self._start
|
|
|
|
end = self._end
|
|
|
|
hidden = self._hiddenrevs
|
|
|
|
return (((end < rev <= start) or (start <= rev and rev < end))
|
|
|
|
and not (hidden and rev in hidden))
|
2014-02-12 22:16:21 +04:00
|
|
|
|
2014-03-14 20:07:59 +04:00
|
|
|
def __nonzero__(self):
|
|
|
|
for r in self:
|
|
|
|
return True
|
|
|
|
return False
|
|
|
|
|
2014-02-12 22:22:43 +04:00
|
|
|
def __and__(self, x):
|
2014-02-22 01:16:17 +04:00
|
|
|
if isinstance(x, baseset):
|
|
|
|
x = x.set()
|
2014-09-17 10:47:34 +04:00
|
|
|
return orderedlazyset(self, x.__contains__,
|
|
|
|
ascending=self.isascending())
|
2014-02-12 22:22:43 +04:00
|
|
|
|
|
|
|
def __sub__(self, x):
|
2014-02-22 01:16:17 +04:00
|
|
|
if isinstance(x, baseset):
|
|
|
|
x = x.set()
|
2014-09-17 10:47:34 +04:00
|
|
|
return orderedlazyset(self, lambda r: r not in x,
|
|
|
|
ascending=self.isascending())
|
2014-02-12 22:22:43 +04:00
|
|
|
|
|
|
|
def __add__(self, x):
|
2014-03-12 04:25:53 +04:00
|
|
|
kwargs = {}
|
|
|
|
if self.isascending() and x.isascending():
|
|
|
|
kwargs['ascending'] = True
|
|
|
|
if self.isdescending() and x.isdescending():
|
|
|
|
kwargs['ascending'] = False
|
|
|
|
return _addset(self, x, **kwargs)
|
2014-02-12 22:22:43 +04:00
|
|
|
|
2014-02-07 03:56:25 +04:00
|
|
|
def __len__(self):
|
2014-02-11 05:38:43 +04:00
|
|
|
if not self._hiddenrevs:
|
|
|
|
return abs(self._end - self._start)
|
|
|
|
else:
|
|
|
|
count = 0
|
2014-04-26 05:00:07 +04:00
|
|
|
start = self._start
|
|
|
|
end = self._end
|
2014-02-11 05:38:43 +04:00
|
|
|
for rev in self._hiddenrevs:
|
2014-04-29 02:14:11 +04:00
|
|
|
if (end < rev <= start) or (start <= rev < end):
|
2014-02-11 05:38:43 +04:00
|
|
|
count += 1
|
|
|
|
return abs(self._end - self._start) - count
|
2014-02-07 03:56:25 +04:00
|
|
|
|
|
|
|
def __getitem__(self, x):
|
|
|
|
# Basic implementation to be changed in future patches.
|
|
|
|
l = baseset([r for r in self])
|
|
|
|
return l[x]
|
|
|
|
|
|
|
|
def sort(self, reverse=False):
|
2014-03-14 04:16:58 +04:00
|
|
|
if bool(reverse) != (self._start > self._end):
|
2014-02-07 03:56:25 +04:00
|
|
|
self.reverse()
|
|
|
|
|
|
|
|
def reverse(self):
|
2014-03-14 21:59:51 +04:00
|
|
|
# Just switch the _start and _end parameters
|
2014-09-17 10:37:03 +04:00
|
|
|
if self.isascending():
|
2014-02-07 03:56:25 +04:00
|
|
|
self._start, self._end = self._end - 1, self._start - 1
|
|
|
|
else:
|
|
|
|
self._start, self._end = self._end + 1, self._start + 1
|
|
|
|
|
|
|
|
def set(self):
|
|
|
|
return self
|
|
|
|
|
2014-03-12 04:09:23 +04:00
|
|
|
def isascending(self):
|
2014-09-17 10:34:18 +04:00
|
|
|
return self._start <= self._end
|
2014-03-12 04:09:23 +04:00
|
|
|
|
|
|
|
def isdescending(self):
|
2014-09-17 10:34:18 +04:00
|
|
|
return self._start >= self._end
|
2014-03-12 04:09:23 +04:00
|
|
|
|
2014-02-07 05:18:11 +04:00
|
|
|
def filter(self, l):
|
2014-09-17 10:47:34 +04:00
|
|
|
return orderedlazyset(self, l, ascending=self.isascending())
|
2014-02-07 05:18:11 +04:00
|
|
|
|
2014-09-19 00:04:02 +04:00
|
|
|
class fullreposet(_spanset):
|
2014-04-30 06:06:15 +04:00
|
|
|
"""a set containing all revisions in the repo
|
|
|
|
|
|
|
|
This class exists to host special optimisation.
|
|
|
|
"""
|
|
|
|
|
|
|
|
def __init__(self, repo):
|
|
|
|
super(fullreposet, self).__init__(repo)
|
|
|
|
|
2014-09-25 07:11:36 +04:00
|
|
|
def __and__(self, other):
|
|
|
|
"""fullrepo & other -> other
|
|
|
|
|
|
|
|
As self contains the whole repo, all of the other set should also be in
|
|
|
|
self. Therefor `self & other = other`.
|
|
|
|
|
|
|
|
This boldly assumes the other contains valid revs only.
|
|
|
|
"""
|
|
|
|
# other not a smartset, make is so
|
|
|
|
if not util.safehasattr(other, 'set'):
|
|
|
|
# filter out hidden revision
|
|
|
|
# (this boldly assumes all smartset are pure)
|
|
|
|
#
|
|
|
|
# `other` was used with "&", let's assume this is a set like
|
|
|
|
# object.
|
|
|
|
other = baseset(other - self._hiddenrevs)
|
|
|
|
elif not util.safehasattr(other, 'ascending'):
|
|
|
|
# "other" is _generatorset not a real smart set
|
|
|
|
# we fallback to the old way (sad kitten)
|
|
|
|
return super(fullreposet, self).__and__(other)
|
|
|
|
|
|
|
|
# preserve order:
|
|
|
|
#
|
|
|
|
# this is probably useless and harmful in multiple cases but matches
|
|
|
|
# the current behavior.
|
|
|
|
if self.isascending():
|
|
|
|
other.ascending()
|
|
|
|
else:
|
|
|
|
other.descending()
|
|
|
|
return other
|
|
|
|
|
2010-10-24 14:52:37 +04:00
|
|
|
# tell hggettext to extract docstrings from these functions:
|
|
|
|
i18nfunctions = symbols.values()
|