sapling/mercurial/context.py

2372 lines
84 KiB
Python
Raw Normal View History

# context.py - changeset and file context objects for mercurial
#
# Copyright 2006, 2007 Matt Mackall <mpm@selenic.com>
#
# This software may be used and distributed according to the terms of the
2010-01-20 07:20:08 +03:00
# GNU General Public License version 2 or any later version.
2015-12-22 08:51:31 +03:00
from __future__ import absolute_import
import errno
import os
import re
2015-12-22 08:51:31 +03:00
import stat
2015-12-22 08:51:31 +03:00
from .i18n import _
from .node import (
addednodeid,
2015-12-22 08:51:31 +03:00
bin,
hex,
modifiednodeid,
2015-12-22 08:51:31 +03:00
nullid,
nullrev,
short,
wdirid,
wdirnodes,
wdirrev,
2015-12-22 08:51:31 +03:00
)
from . import (
encoding,
error,
fileset,
match as matchmod,
mdiff,
obsolete as obsmod,
patch,
pathutil,
2015-12-22 08:51:31 +03:00
phases,
pycompat,
2015-12-22 08:51:31 +03:00
repoview,
revlog,
scmutil,
sparse,
2015-12-22 08:51:31 +03:00
subrepo,
util,
)
propertycache = util.propertycache
nonascii = re.compile(r'[^\x21-\x7f]').search
class basectx(object):
"""A basectx object represents the common logic for its children:
changectx: read-only context that is already present in the repo,
workingctx: a context that represents the working directory and can
be committed,
memctx: a context that represents changes in-memory and can also
be committed."""
def __new__(cls, repo, changeid='', *args, **kwargs):
if isinstance(changeid, basectx):
return changeid
o = super(basectx, cls).__new__(cls)
o._repo = repo
o._rev = nullrev
o._node = nullid
return o
def __bytes__(self):
2013-08-03 01:46:23 +04:00
return short(self.node())
__str__ = encoding.strmethod(__bytes__)
2013-08-03 01:52:13 +04:00
def __int__(self):
return self.rev()
def __repr__(self):
return r"<%s %s>" % (type(self).__name__, str(self))
def __eq__(self, other):
try:
return type(self) == type(other) and self._rev == other._rev
except AttributeError:
return False
2013-08-06 02:00:32 +04:00
def __ne__(self, other):
return not (self == other)
def __contains__(self, key):
return key in self._manifest
def __getitem__(self, key):
return self.filectx(key)
2013-08-06 02:22:18 +04:00
def __iter__(self):
return iter(self._manifest)
2013-08-06 02:22:18 +04:00
def _buildstatusmanifest(self, status):
"""Builds a manifest that includes the given status results, if this is
a working copy context. For non-working copy contexts, it just returns
the normal manifest."""
return self.manifest()
def _matchstatus(self, other, match):
"""This internal method provides a way for child objects to override the
match operator.
"""
return match
def _buildstatus(self, other, s, match, listignored, listclean,
listunknown):
"""build a status with respect to another context"""
# Load earliest manifest first for caching reasons. More specifically,
# if you have revisions 1000 and 1001, 1001 is probably stored as a
# delta against 1000. Thus, if you read 1000 first, we'll reconstruct
# 1000 and cache it so that when you read 1001, we just need to apply a
# delta to what's in the cache. So that's one full reconstruction + one
# delta application.
mf2 = None
if self.rev() is not None and self.rev() < other.rev():
mf2 = self._buildstatusmanifest(s)
mf1 = other._buildstatusmanifest(s)
if mf2 is None:
mf2 = self._buildstatusmanifest(s)
modified, added = [], []
removed = []
clean = []
deleted, unknown, ignored = s.deleted, s.unknown, s.ignored
deletedset = set(deleted)
d = mf1.diff(mf2, match=match, clean=listclean)
for fn, value in d.iteritems():
if fn in deletedset:
continue
if value is None:
clean.append(fn)
continue
(node1, flag1), (node2, flag2) = value
if node1 is None:
added.append(fn)
elif node2 is None:
removed.append(fn)
elif flag1 != flag2:
modified.append(fn)
elif node2 not in wdirnodes:
# When comparing files between two commits, we save time by
# not comparing the file contents when the nodeids differ.
# Note that this means we incorrectly report a reverted change
# to a file as a modification.
modified.append(fn)
elif self[fn].cmp(other[fn]):
modified.append(fn)
else:
clean.append(fn)
if removed:
# need to filter files if they are already reported as removed
unknown = [fn for fn in unknown if fn not in mf1 and
(not match or match(fn))]
ignored = [fn for fn in ignored if fn not in mf1 and
(not match or match(fn))]
# if they're deleted, don't report them as removed
removed = [fn for fn in removed if fn not in deletedset]
return scmutil.status(modified, added, removed, deleted, unknown,
ignored, clean)
2013-08-06 02:21:23 +04:00
@propertycache
def substate(self):
return subrepo.state(self, self._repo.ui)
def subrev(self, subpath):
return self.substate[subpath][1]
2013-08-03 04:09:06 +04:00
def rev(self):
return self._rev
2013-08-03 01:48:19 +04:00
def node(self):
return self._node
2013-08-03 01:49:01 +04:00
def hex(self):
return hex(self.node())
2013-08-06 02:22:49 +04:00
def manifest(self):
return self._manifest
def manifestctx(self):
return self._manifestctx
def repo(self):
return self._repo
2013-08-06 03:19:04 +04:00
def phasestr(self):
return phases.phasenames[self.phase()]
2013-08-06 03:19:19 +04:00
def mutable(self):
return self.phase() > phases.public
2013-08-03 04:09:06 +04:00
def getfileset(self, expr):
return fileset.getfileset(self, expr)
def obsolete(self):
"""True if the changeset is obsolete"""
return self.rev() in obsmod.getrevs(self._repo, 'obsolete')
def extinct(self):
"""True if the changeset is extinct"""
return self.rev() in obsmod.getrevs(self._repo, 'extinct')
def unstable(self):
msg = ("'context.unstable' is deprecated, "
"use 'context.orphan'")
self._repo.ui.deprecwarn(msg, '4.4')
return self.orphan()
def orphan(self):
"""True if the changeset is not obsolete but it's ancestor are"""
return self.rev() in obsmod.getrevs(self._repo, 'orphan')
def bumped(self):
msg = ("'context.bumped' is deprecated, "
"use 'context.phasedivergent'")
self._repo.ui.deprecwarn(msg, '4.4')
return self.phasedivergent()
def phasedivergent(self):
"""True if the changeset try to be a successor of a public changeset
Only non-public and non-obsolete changesets may be bumped.
"""
return self.rev() in obsmod.getrevs(self._repo, 'phasedivergent')
def divergent(self):
msg = ("'context.divergent' is deprecated, "
"use 'context.contentdivergent'")
self._repo.ui.deprecwarn(msg, '4.4')
return self.contentdivergent()
def contentdivergent(self):
"""Is a successors of a changeset with multiple possible successors set
Only non-public and non-obsolete changesets may be divergent.
"""
return self.rev() in obsmod.getrevs(self._repo, 'contentdivergent')
def troubled(self):
msg = ("'context.troubled' is deprecated, "
"use 'context.isunstable'")
self._repo.ui.deprecwarn(msg, '4.4')
return self.isunstable()
def isunstable(self):
"""True if the changeset is either unstable, bumped or divergent"""
return self.orphan() or self.phasedivergent() or self.contentdivergent()
def troubles(self):
"""Keep the old version around in order to avoid breaking extensions
about different return values.
"""
msg = ("'context.troubles' is deprecated, "
"use 'context.instabilities'")
self._repo.ui.deprecwarn(msg, '4.4')
troubles = []
if self.orphan():
troubles.append('orphan')
if self.phasedivergent():
troubles.append('bumped')
if self.contentdivergent():
troubles.append('divergent')
return troubles
def instabilities(self):
"""return the list of instabilities affecting this changeset.
Instabilities are returned as strings. possible values are:
- orphan,
- phase-divergent,
- content-divergent.
"""
instabilities = []
if self.orphan():
instabilities.append('orphan')
if self.phasedivergent():
instabilities.append('phase-divergent')
if self.contentdivergent():
instabilities.append('content-divergent')
return instabilities
2013-08-06 03:19:38 +04:00
def parents(self):
"""return contexts for each parent changeset"""
return self._parents
2013-08-06 03:26:15 +04:00
def p1(self):
return self._parents[0]
2013-08-06 03:26:54 +04:00
def p2(self):
parents = self._parents
if len(parents) == 2:
return parents[1]
return changectx(self._repo, nullrev)
2013-08-06 03:26:54 +04:00
2013-08-06 03:28:23 +04:00
def _fileinfo(self, path):
if r'_manifest' in self.__dict__:
2013-08-06 03:28:23 +04:00
try:
return self._manifest[path], self._manifest.flags(path)
except KeyError:
raise error.ManifestLookupError(self._node, path,
_('not found in manifest'))
if r'_manifestdelta' in self.__dict__ or path in self.files():
2013-08-06 03:28:23 +04:00
if path in self._manifestdelta:
return (self._manifestdelta[path],
self._manifestdelta.flags(path))
mfl = self._repo.manifestlog
try:
node, flag = mfl[self._changeset.manifest].find(path)
except KeyError:
2013-08-06 03:28:23 +04:00
raise error.ManifestLookupError(self._node, path,
_('not found in manifest'))
return node, flag
2013-08-06 03:28:40 +04:00
def filenode(self, path):
return self._fileinfo(path)[0]
2013-08-06 03:28:54 +04:00
def flags(self, path):
try:
return self._fileinfo(path)[1]
except error.LookupError:
return ''
verify: don't init subrepo when missing one is referenced (issue5128) (API) Initializing a subrepo when one doesn't exist is the right thing to do when the parent is being updated, but in few other cases. Unfortunately, there isn't enough context in the subrepo module to distinguish this case. This same issue can be caused with other subrepo aware commands, so there is a general issue here beyond the scope of this fix. A simpler attempt I tried was to add an '_updating' boolean to localrepo, and set/clear it around the call to mergemod.update() in hg.updaterepo(). That mostly worked, but doesn't handle the case where archive will clone the subrepo if it is missing. (I vaguely recall that there may be other commands that will clone if needed like this, but certainly not all do. It seems both handy, and a bit surprising for what should be a read only operation. It might be nice if all commands did this consistently, but we probably need Angel's subrepo caching first, to not make a mess of the working directory.) I originally handled 'Exception' in order to pick up the Aborts raised in subrepo.state(), but this turns out to be unnecessary because that is called once and cached by ctx.sub() when iterating the subrepos. It was suggested in the bug discussion to skip looking at the subrepo links unless -S is specified. I don't really like that idea because missing a subrepo or (less likely, but worse) a corrupt .hgsubstate is a problem of the parent repo when checking out a revision. The -S option seems like a better fit for functionality that would recurse into each subrepo and do a full verification. Ultimately, the default value for 'allowcreate' should probably be flipped, but since the default behavior was to allow creation, this is less risky for now.
2016-04-28 05:45:52 +03:00
def sub(self, path, allowcreate=True):
'''return a subrepo for the stored revision of path, never wdir()'''
verify: don't init subrepo when missing one is referenced (issue5128) (API) Initializing a subrepo when one doesn't exist is the right thing to do when the parent is being updated, but in few other cases. Unfortunately, there isn't enough context in the subrepo module to distinguish this case. This same issue can be caused with other subrepo aware commands, so there is a general issue here beyond the scope of this fix. A simpler attempt I tried was to add an '_updating' boolean to localrepo, and set/clear it around the call to mergemod.update() in hg.updaterepo(). That mostly worked, but doesn't handle the case where archive will clone the subrepo if it is missing. (I vaguely recall that there may be other commands that will clone if needed like this, but certainly not all do. It seems both handy, and a bit surprising for what should be a read only operation. It might be nice if all commands did this consistently, but we probably need Angel's subrepo caching first, to not make a mess of the working directory.) I originally handled 'Exception' in order to pick up the Aborts raised in subrepo.state(), but this turns out to be unnecessary because that is called once and cached by ctx.sub() when iterating the subrepos. It was suggested in the bug discussion to skip looking at the subrepo links unless -S is specified. I don't really like that idea because missing a subrepo or (less likely, but worse) a corrupt .hgsubstate is a problem of the parent repo when checking out a revision. The -S option seems like a better fit for functionality that would recurse into each subrepo and do a full verification. Ultimately, the default value for 'allowcreate' should probably be flipped, but since the default behavior was to allow creation, this is less risky for now.
2016-04-28 05:45:52 +03:00
return subrepo.subrepo(self, path, allowcreate=allowcreate)
2013-08-06 03:40:36 +04:00
def nullsub(self, path, pctx):
return subrepo.nullsubrepo(self, path, pctx)
def workingsub(self, path):
'''return a subrepo for the stored revision, or wdir if this is a wdir
context.
'''
return subrepo.subrepo(self, path, allowwdir=True)
def match(self, pats=None, include=None, exclude=None, default='glob',
listsubrepos=False, badfn=None):
2013-08-06 03:41:00 +04:00
r = self._repo
return matchmod.match(r.root, r.getcwd(), pats,
2013-08-06 03:41:00 +04:00
include, exclude, default,
auditor=r.nofsauditor, ctx=self,
listsubrepos=listsubrepos, badfn=badfn)
2013-08-06 03:41:00 +04:00
2013-08-06 03:41:12 +04:00
def diff(self, ctx2=None, match=None, **opts):
"""Returns a diff generator for the given contexts and matcher"""
if ctx2 is None:
ctx2 = self.p1()
if ctx2 is not None:
2013-08-06 03:41:12 +04:00
ctx2 = self._repo[ctx2]
diffopts = patch.diffopts(self._repo.ui, opts)
return patch.diff(self._repo, ctx2, self, match=match, opts=diffopts)
2013-08-06 03:41:12 +04:00
2013-08-06 03:41:56 +04:00
def dirs(self):
return self._manifest.dirs()
def hasdir(self, dir):
return self._manifest.hasdir(dir)
2013-08-06 03:41:56 +04:00
def status(self, other=None, match=None, listignored=False,
listclean=False, listunknown=False, listsubrepos=False):
"""return status of files between two nodes or node and working
directory.
If other is None, compare this node with working directory.
returns (modified, added, removed, deleted, unknown, ignored, clean)
"""
ctx1 = self
ctx2 = self._repo[other]
# This next code block is, admittedly, fragile logic that tests for
# reversing the contexts and wouldn't need to exist if it weren't for
# the fast (and common) code path of comparing the working directory
# with its first parent.
#
# What we're aiming for here is the ability to call:
#
# workingctx.status(parentctx)
#
# If we always built the manifest for each context and compared those,
# then we'd be done. But the special case of the above call means we
# just copy the manifest of the parent.
reversed = False
if (not isinstance(ctx1, changectx)
and isinstance(ctx2, changectx)):
reversed = True
ctx1, ctx2 = ctx2, ctx1
match = match or matchmod.always(self._repo.root, self._repo.getcwd())
match = ctx2._matchstatus(ctx1, match)
r = scmutil.status([], [], [], [], [], [], [])
r = ctx2._buildstatus(ctx1, r, match, listignored, listclean,
listunknown)
if reversed:
# Reverse added and removed. Clear deleted, unknown and ignored as
# these make no sense to reverse.
r = scmutil.status(r.modified, r.removed, r.added, [], [], [],
r.clean)
if listsubrepos:
for subpath, sub in scmutil.itersubrepos(ctx1, ctx2):
try:
rev2 = ctx2.subrev(subpath)
except KeyError:
# A subrepo that existed in node1 was deleted between
# node1 and node2 (inclusive). Thus, ctx2's substate
# won't contain that subpath. The best we can do ignore it.
rev2 = None
submatch = matchmod.subdirmatcher(subpath, match)
s = sub.status(rev2, match=submatch, ignored=listignored,
clean=listclean, unknown=listunknown,
listsubrepos=True)
for rfiles, sfiles in zip(r, s):
rfiles.extend("%s/%s" % (subpath, f) for f in sfiles)
for l in r:
l.sort()
return r
def _filterederror(repo, changeid):
"""build an exception to be raised about a filtered changeid
This is extracted in a function to help extensions (eg: evolve) to
experiment with various message variants."""
if repo.filtername.startswith('visible'):
msg = _("hidden revision '%s'") % changeid
hint = _('use --hidden to access hidden revisions')
return error.FilteredRepoLookupError(msg, hint=hint)
msg = _("filtered revision '%s' (not in '%s' subset)")
msg %= (changeid, repo.filtername)
return error.FilteredRepoLookupError(msg)
class changectx(basectx):
"""A changecontext object makes access to data related to a particular
2013-10-23 21:49:56 +04:00
changeset convenient. It represents a read-only context already present in
the repo."""
2008-06-26 22:46:31 +04:00
def __init__(self, repo, changeid=''):
"""changeid is a revision number, node, or tag"""
# since basectx.__new__ already took care of copying the object, we
# don't need to do anything in __init__, so we just exit here
if isinstance(changeid, basectx):
return
2008-06-26 22:46:31 +04:00
if changeid == '':
changeid = '.'
self._repo = repo
try:
if isinstance(changeid, int):
self._node = repo.changelog.node(changeid)
self._rev = changeid
return
if not pycompat.ispy3 and isinstance(changeid, long):
changeid = str(changeid)
if changeid == 'null':
self._node = nullid
self._rev = nullrev
return
if changeid == 'tip':
self._node = repo.changelog.tip()
self._rev = repo.changelog.rev(self._node)
return
if changeid == '.' or changeid == repo.dirstate.p1():
# this is a hack to delay/avoid loading obsmarkers
# when we know that '.' won't be hidden
self._node = repo.dirstate.p1()
self._rev = repo.unfiltered().changelog.rev(self._node)
return
if len(changeid) == 20:
try:
self._node = changeid
self._rev = repo.changelog.rev(changeid)
return
except error.FilteredRepoLookupError:
raise
except LookupError:
pass
try:
r = int(changeid)
if '%d' % r != changeid:
raise ValueError
l = len(repo.changelog)
if r < 0:
r += l
if r < 0 or r >= l and r != wdirrev:
raise ValueError
self._rev = r
self._node = repo.changelog.node(r)
return
except error.FilteredIndexError:
raise
except (ValueError, OverflowError, IndexError):
pass
if len(changeid) == 40:
try:
self._node = bin(changeid)
self._rev = repo.changelog.rev(self._node)
return
except error.FilteredLookupError:
raise
except (TypeError, LookupError):
pass
# lookup bookmarks through the name interface
try:
self._node = repo.names.singlenode(repo, changeid)
self._rev = repo.changelog.rev(self._node)
return
except KeyError:
pass
except error.FilteredRepoLookupError:
raise
except error.RepoLookupError:
pass
self._node = repo.unfiltered().changelog._partialmatch(changeid)
if self._node is not None:
self._rev = repo.changelog.rev(self._node)
return
# lookup failed
# check if it might have come from damaged dirstate
#
# XXX we could avoid the unfiltered if we had a recognizable
# exception for filtered changeset access
if changeid in repo.unfiltered().dirstate.parents():
msg = _("working directory has unknown parent '%s'!")
raise error.Abort(msg % short(changeid))
try:
if len(changeid) == 20 and nonascii(changeid):
changeid = hex(changeid)
except TypeError:
pass
except (error.FilteredIndexError, error.FilteredLookupError,
error.FilteredRepoLookupError):
raise _filterederror(repo, changeid)
except IndexError:
pass
raise error.RepoLookupError(
_("unknown revision '%s'") % changeid)
def __hash__(self):
try:
return hash(self._rev)
except AttributeError:
return id(self)
2006-09-30 02:47:51 +04:00
def __nonzero__(self):
return self._rev != nullrev
2006-09-30 02:47:51 +04:00
__bool__ = __nonzero__
@propertycache
def _changeset(self):
return self._repo.changelog.changelogrevision(self.rev())
@propertycache
def _manifest(self):
return self._manifestctx.read()
@property
def _manifestctx(self):
return self._repo.manifestlog[self._changeset.manifest]
@propertycache
def _manifestdelta(self):
return self._manifestctx.readdelta()
@propertycache
def _parents(self):
repo = self._repo
p1, p2 = repo.changelog.parentrevs(self._rev)
if p2 == nullrev:
return [changectx(repo, p1)]
return [changectx(repo, p1), changectx(repo, p2)]
2010-01-25 09:05:27 +03:00
def changeset(self):
c = self._changeset
return (
c.manifest,
c.user,
c.date,
c.files,
c.description,
c.extra,
)
2010-01-25 09:05:27 +03:00
def manifestnode(self):
return self._changeset.manifest
2010-01-25 09:05:27 +03:00
def user(self):
return self._changeset.user
2010-01-25 09:05:27 +03:00
def date(self):
return self._changeset.date
2010-01-25 09:05:27 +03:00
def files(self):
return self._changeset.files
2010-01-25 09:05:27 +03:00
def description(self):
return self._changeset.description
2010-01-25 09:05:27 +03:00
def branch(self):
return encoding.tolocal(self._changeset.extra.get("branch"))
def closesbranch(self):
return 'close' in self._changeset.extra
2010-01-25 09:05:27 +03:00
def extra(self):
return self._changeset.extra
2010-01-25 09:05:27 +03:00
def tags(self):
return self._repo.nodetags(self._node)
def bookmarks(self):
return self._repo.nodebookmarks(self._node)
2011-10-18 20:25:53 +04:00
def phase(self):
return self._repo._phasecache.phase(self._repo, self._rev)
def hidden(self):
return self._rev in repoview.filterrevs(self._repo, 'visible')
def children(self):
"""return contexts for each child changeset"""
c = self._repo.changelog.children(self._node)
2006-11-17 10:06:54 +03:00
return [changectx(self._repo, x) for x in c]
def ancestors(self):
for a in self._repo.changelog.ancestors([self._rev]):
yield changectx(self._repo, a)
def descendants(self):
for d in self._repo.changelog.descendants([self._rev]):
yield changectx(self._repo, d)
def filectx(self, path, fileid=None, filelog=None):
"""get a file context from this changeset"""
if fileid is None:
fileid = self.filenode(path)
return filectx(self._repo, path, fileid=fileid,
changectx=self, filelog=filelog)
def ancestor(self, c2, warn=False):
"""return the "best" ancestor context of self and c2
If there are multiple candidates, it will show a message and check
merge.preferancestor configuration before falling back to the
revlog ancestor."""
# deal with workingctxs
n2 = c2._node
if n2 is None:
n2 = c2._parents[0]._node
cahs = self._repo.changelog.commonancestorsheads(self._node, n2)
if not cahs:
anc = nullid
elif len(cahs) == 1:
anc = cahs[0]
else:
# experimental config: merge.preferancestor
for r in self._repo.ui.configlist('merge', 'preferancestor', ['*']):
try:
ctx = changectx(self._repo, r)
except error.RepoLookupError:
continue
anc = ctx.node()
if anc in cahs:
break
else:
anc = self._repo.changelog.ancestor(self._node, n2)
if warn:
self._repo.ui.status(
(_("note: using %s as ancestor of %s and %s\n") %
(short(anc), short(self._node), short(n2))) +
''.join(_(" alternatively, use --config "
"merge.preferancestor=%s\n") %
short(n) for n in sorted(cahs) if n != anc))
return changectx(self._repo, anc)
2006-09-18 07:59:33 +04:00
def descendant(self, other):
"""True if other is descendant of this changeset"""
return self._repo.changelog.descendant(self._rev, other._rev)
2008-06-28 04:25:48 +04:00
def walk(self, match):
'''Generates matching file names.'''
# Wrap match.bad method to have message with nodeid
def bad(fn, msg):
# The manifest doesn't know about subrepos, so don't complain about
# paths into valid subrepos.
2015-05-19 15:17:57 +03:00
if any(fn == s or fn.startswith(s + '/')
for s in self.substate):
return
match.bad(fn, _('no such file in rev %s') % self)
m = matchmod.badmatch(match, bad)
return self._manifest.walk(m)
2008-06-28 04:25:48 +04:00
def matches(self, match):
return self.walk(match)
class basefilectx(object):
"""A filecontext object represents the common logic for its children:
filectx: read-only access to a filerevision that is already present
in the repo,
workingfilectx: a filecontext that represents files from the working
directory,
memfilectx: a filecontext that represents files in-memory,
overlayfilectx: duplicate another filecontext with some fields overridden.
"""
@propertycache
def _filelog(self):
return self._repo.file(self._path)
@propertycache
def _changeid(self):
if r'_changeid' in self.__dict__:
return self._changeid
elif r'_changectx' in self.__dict__:
return self._changectx.rev()
elif r'_descendantrev' in self.__dict__:
# this file context was created from a revision with a known
# descendant, we can (lazily) correct for linkrev aliases
return self._adjustlinkrev(self._descendantrev)
else:
return self._filelog.linkrev(self._filerev)
@propertycache
def _filenode(self):
if r'_fileid' in self.__dict__:
return self._filelog.lookup(self._fileid)
else:
return self._changectx.filenode(self._path)
@propertycache
def _filerev(self):
return self._filelog.rev(self._filenode)
@propertycache
def _repopath(self):
return self._path
def __nonzero__(self):
try:
self._filenode
return True
except error.LookupError:
# file is missing
return False
__bool__ = __nonzero__
def __bytes__(self):
try:
return "%s@%s" % (self.path(), self._changectx)
except error.LookupError:
return "%s@???" % self.path()
__str__ = encoding.strmethod(__bytes__)
def __repr__(self):
return "<%s %s>" % (type(self).__name__, str(self))
def __hash__(self):
try:
return hash((self._path, self._filenode))
except AttributeError:
return id(self)
def __eq__(self, other):
try:
return (type(self) == type(other) and self._path == other._path
and self._filenode == other._filenode)
except AttributeError:
return False
2013-08-12 07:49:40 +04:00
def __ne__(self, other):
return not (self == other)
2013-08-12 07:50:15 +04:00
def filerev(self):
return self._filerev
def filenode(self):
return self._filenode
@propertycache
def _flags(self):
2013-08-12 07:51:04 +04:00
return self._changectx.flags(self._path)
def flags(self):
return self._flags
2013-08-12 07:51:18 +04:00
def filelog(self):
return self._filelog
2013-08-12 07:51:30 +04:00
def rev(self):
return self._changeid
2013-08-12 07:51:41 +04:00
def linkrev(self):
return self._filelog.linkrev(self._filerev)
2013-08-12 07:51:53 +04:00
def node(self):
return self._changectx.node()
def hex(self):
return self._changectx.hex()
2013-08-12 07:53:47 +04:00
def user(self):
return self._changectx.user()
2013-08-12 07:53:56 +04:00
def date(self):
return self._changectx.date()
2013-08-12 07:54:12 +04:00
def files(self):
return self._changectx.files()
def description(self):
return self._changectx.description()
2013-08-12 07:54:31 +04:00
def branch(self):
return self._changectx.branch()
2013-08-12 07:54:39 +04:00
def extra(self):
return self._changectx.extra()
2013-08-12 07:54:48 +04:00
def phase(self):
return self._changectx.phase()
def phasestr(self):
return self._changectx.phasestr()
def manifest(self):
return self._changectx.manifest()
def changectx(self):
return self._changectx
def renamed(self):
return self._copied
def repo(self):
return self._repo
def size(self):
return len(self.data())
2013-08-12 07:50:15 +04:00
2013-08-12 07:56:18 +04:00
def path(self):
return self._path
def isbinary(self):
try:
return util.binary(self.data())
except IOError:
return False
def isexec(self):
return 'x' in self.flags()
def islink(self):
return 'l' in self.flags()
def isabsent(self):
"""whether this filectx represents a file not in self._changectx
This is mainly for merge code to detect change/delete conflicts. This is
expected to be True for all subclasses of basectx."""
return False
_customcmp = False
2013-08-12 07:56:53 +04:00
def cmp(self, fctx):
"""compare with other file context
returns True if different than fctx.
"""
if fctx._customcmp:
return fctx.cmp(self)
if (fctx._filenode is None
2013-08-12 07:56:53 +04:00
and (self._repo._encodefilterpats
# if file data starts with '\1\n', empty metadata block is
# prepended, which adds 4 bytes to filelog.size().
or self.size() - 4 == fctx.size())
or self.size() == fctx.size()):
return self._filelog.cmp(self._filenode, fctx.data())
return True
def _adjustlinkrev(self, srcrev, inclusive=False):
"""return the first ancestor of <srcrev> introducing <fnode>
If the linkrev of the file revision does not point to an ancestor of
srcrev, we'll walk down the ancestors until we find one introducing
this file revision.
:srcrev: the changeset revision we search ancestors from
:inclusive: if true, the src revision will also be checked
"""
repo = self._repo
cl = repo.unfiltered().changelog
mfl = repo.manifestlog
# fetch the linkrev
lkr = self.linkrev()
# hack to reuse ancestor computation when searching for renames
memberanc = getattr(self, '_ancestrycontext', None)
iteranc = None
if srcrev is None:
# wctx case, used by workingfilectx during mergecopy
revs = [p.rev() for p in self._repo[None].parents()]
inclusive = True # we skipped the real (revless) source
else:
revs = [srcrev]
if memberanc is None:
memberanc = iteranc = cl.ancestors(revs, lkr,
inclusive=inclusive)
# check if this linkrev is an ancestor of srcrev
if lkr not in memberanc:
if iteranc is None:
iteranc = cl.ancestors(revs, lkr, inclusive=inclusive)
fnode = self._filenode
path = self._path
for a in iteranc:
ac = cl.read(a) # get changeset data (we avoid object creation)
if path in ac[3]: # checking the 'files' field.
# The file has been touched, check if the content is
# similar to the one we search for.
if fnode == mfl[ac[0]].readfast().get(path):
return a
# In theory, we should never get out of that loop without a result.
# But if manifest uses a buggy file revision (not children of the
# one it replaces) we could. Such a buggy situation will likely
# result is crash somewhere else at to some point.
return lkr
def introrev(self):
"""return the rev of the changeset which introduced this file revision
This method is different from linkrev because it take into account the
changeset the filectx was created from. It ensures the returned
revision is one of its ancestors. This prevents bugs from
'linkrev-shadowing' when a file revision is used by multiple
changesets.
"""
lkr = self.linkrev()
attrs = vars(self)
noctx = not ('_changeid' in attrs or '_changectx' in attrs)
if noctx or self.rev() == lkr:
return self.linkrev()
return self._adjustlinkrev(self.rev(), inclusive=True)
def _parentfilectx(self, path, fileid, filelog):
"""create parent filectx keeping ancestry info for _adjustlinkrev()"""
fctx = filectx(self._repo, path, fileid=fileid, filelog=filelog)
if '_changeid' in vars(self) or '_changectx' in vars(self):
# If self is associated with a changeset (probably explicitly
# fed), ensure the created filectx is associated with a
# changeset that is an ancestor of self.changectx.
# This lets us later use _adjustlinkrev to get a correct link.
fctx._descendantrev = self.rev()
fctx._ancestrycontext = getattr(self, '_ancestrycontext', None)
elif '_descendantrev' in vars(self):
# Otherwise propagate _descendantrev if we have one associated.
fctx._descendantrev = self._descendantrev
fctx._ancestrycontext = getattr(self, '_ancestrycontext', None)
return fctx
2013-08-12 07:57:21 +04:00
def parents(self):
_path = self._path
2013-08-12 07:57:21 +04:00
fl = self._filelog
parents = self._filelog.parents(self._filenode)
pl = [(_path, node, fl) for node in parents if node != nullid]
2013-08-12 07:57:21 +04:00
filectx.parents: enforce changeid of parent to be in own changectx ancestors Because of the way filenodes are computed, you can have multiple changesets "introducing" the same file revision. For example, in the changeset graph below, changeset 2 and 3 both change a file -to- and -from- the same content. o 3: content = new | | o 2: content = new |/ o 1: content = old In such cases, the file revision is create once, when 2 is added, and just reused for 3. So the file change in '3' (from "old" to "new)" has no linkrev pointing to it). We'll call this situation "linkrev-shadowing". As the linkrev is used for optimization purposes when walking a file history, the linkrev-shadowing results in an unexpected jump to another branch during such a walk.. This leads to multiple bugs with log, annotate and rename detection. One element to fix such bugs is to ensure that walking the file history sticks on the same topology as the changeset's history. For this purpose, we extend the logic in 'basefilectx.parents' so that it always defines the proper changeset to associate the parent file revision with. This "proper" changeset has to be an ancestor of the changeset associated with the child file revision. This logic is performed in the '_adjustlinkrev' function. This function is given the starting changeset and all the information regarding the parent file revision. If the linkrev for the file revision is an ancestor of the starting changeset, the linkrev is valid and will be used. If it is not, we detected a topological jump caused by linkrev shadowing, we are going to walk the ancestors of the starting changeset until we find one setting the file to the revision we are trying to create. The performance impact appears acceptable: - We are walking the changelog once for each filelog traversal (as there should be no overlap between searches), - changelog traversal itself is fairly cheap, compared to what is likely going to be perform on the result on the filelog traversal, - We only touch the manifest for ancestors touching the file, And such changesets are likely to be the one introducing the file. (except in pathological cases involving merge), - We use manifest diff instead of full manifest unpacking to check manifest content, so it does not involve applying multiple diffs in most case. - linkrev shadowing is not the common case. Tests for fixed issues in log, annotate and rename detection have been added. But this changeset does not solve all problems. It fixes -ancestry- computation, but if the linkrev-shadowed changesets is the starting one, we'll still get things wrong. We'll have to fix the bootstrapping of such operations in a later changeset. Also, the usage of `hg log FILE` without --follow still has issues with linkrev pointing to hidden changesets, because it relies on the `filelog` revset which implement its own traversal logic that is still to be fixed. Thanks goes to: - Matt Mackall: for nudging me in the right direction - Julien Cristau and Rémi Cardona: for keep telling me linkrev bug were an evolution show stopper for 3 years. - Durham Goode: for finding a new linkrev issue every few weeks - Mads Kiilerich: for that last rename bug who raise this topic over my anoyance limit.
2014-12-24 02:30:38 +03:00
r = fl.renamed(self._filenode)
2013-08-12 07:57:21 +04:00
if r:
# - In the simple rename case, both parent are nullid, pl is empty.
# - In case of merge, only one of the parent is null id and should
# be replaced with the rename information. This parent is -always-
# the first one.
#
# As null id have always been filtered out in the previous list
# comprehension, inserting to 0 will always result in "replacing
# first nullid parent with rename information.
pl.insert(0, (r[0], r[1], self._repo.file(r[0])))
2013-08-12 07:57:21 +04:00
return [self._parentfilectx(path, fnode, l) for path, fnode, l in pl]
2013-08-12 07:57:21 +04:00
2013-08-12 07:59:10 +04:00
def p1(self):
return self.parents()[0]
2013-08-12 08:00:11 +04:00
def p2(self):
p = self.parents()
if len(p) == 2:
return p[1]
return filectx(self._repo, self._path, fileid=-1, filelog=self._filelog)
def annotate(self, follow=False, linenumber=False, skiprevs=None,
diffopts=None):
'''returns a list of tuples of ((ctx, number), line) for each line
2006-09-27 20:10:21 +04:00
in the file, where ctx is the filectx of the node where
that line was last changed; if linenumber parameter is true, number is
the line number at the first appearance in the managed file, otherwise,
number has a fixed value of False.
'''
def lines(text):
if text.endswith("\n"):
return text.count("\n")
return text.count("\n") + int(bool(text))
if linenumber:
def decorate(text, rev):
return ([(rev, i) for i in xrange(1, lines(text) + 1)], text)
else:
def decorate(text, rev):
return ([(rev, False)] * lines(text), text)
getlog = util.lrucachefunc(lambda x: self._repo.file(x))
2006-09-27 20:10:21 +04:00
def parents(f):
# Cut _descendantrev here to mitigate the penalty of lazy linkrev
# adjustment. Otherwise, p._adjustlinkrev() would walk changelog
# from the topmost introrev (= srcrev) down to p.linkrev() if it
# isn't an ancestor of the srcrev.
f._changeid
pl = f.parents()
# Don't return renamed parents if we aren't following.
if not follow:
pl = [p for p in pl if p.path() == f.path()]
# renamed filectx won't have a filelog yet, so set it
# from the cache to save time
for p in pl:
if not '_filelog' in p.__dict__:
p._filelog = getlog(p.path())
return pl
2006-10-01 21:26:33 +04:00
# use linkrev to find the first changeset where self appeared
base = self
introrev = self.introrev()
if self.rev() != introrev:
base = self.filectx(self.filenode(), changeid=introrev)
if getattr(base, '_ancestrycontext', None) is None:
cl = self._repo.changelog
if introrev is None:
# wctx is not inclusive, but works because _ancestrycontext
# is used to test filelog revisions
ac = cl.ancestors([p.rev() for p in base.parents()],
inclusive=True)
else:
ac = cl.ancestors([introrev], inclusive=True)
base._ancestrycontext = ac
# This algorithm would prefer to be recursive, but Python is a
# bit recursion-hostile. Instead we do an iterative
# depth-first search.
annotate: pre-calculate the "needed" dictionary (issue5360) The "needed" dict is used as a reference counter to free items in the giant "hist" dict. However, currently it is not very accurate and can lead to dropping "hist" items unnecessarily, for example, with the following DAG, -3- / \ 0--1--2--4-- The current algorithm will visit and calculate rev 1 twice, undesired. And it tries to be smart by clearing rev 1's parents: "pcache[1] = []" at the time hist[1] being accessed (note: hist[1] needs to be used twice, by rev 2 and rev 3). It can result in incorrect results if p1 of rev 4 deletes chunks belonging to rev 0. However, simply removing "needed" is not okay, because it will consume 10x memory: # without any change % HGRCPATH= lrun ./hg annotate mercurial/commands.py -r d130a38 3>&2 [1] MEMORY 49074176 CPUTIME 9.213 REALTIME 9.270 # with "needed" removed MEMORY 637673472 CPUTIME 8.164 REALTIME 8.249 This patch moves "needed" (and "pcache") calculation to a separate DFS to address the issue. It improves perf and fixes issue5360 by correctly reusing hist, while maintaining low memory usage. Some additional attempt has been made to further reduce memory usage, like changing "pcache[f] = []" to "del pcache[f]". Therefore the result can be both faster and lower memory usage: # with this patch applied MEMORY 47575040 CPUTIME 7.870 REALTIME 7.926 [1]: lrun is a lightweight sandbox built on Linux cgroup and namespace. It's used to measure CPU and memory usage here. Source code is available at github.com/quark-zju/lrun.
2016-09-02 17:20:59 +03:00
# 1st DFS pre-calculates pcache and needed
visit = [base]
pcache = {}
needed = {base: 1}
annotate: pre-calculate the "needed" dictionary (issue5360) The "needed" dict is used as a reference counter to free items in the giant "hist" dict. However, currently it is not very accurate and can lead to dropping "hist" items unnecessarily, for example, with the following DAG, -3- / \ 0--1--2--4-- The current algorithm will visit and calculate rev 1 twice, undesired. And it tries to be smart by clearing rev 1's parents: "pcache[1] = []" at the time hist[1] being accessed (note: hist[1] needs to be used twice, by rev 2 and rev 3). It can result in incorrect results if p1 of rev 4 deletes chunks belonging to rev 0. However, simply removing "needed" is not okay, because it will consume 10x memory: # without any change % HGRCPATH= lrun ./hg annotate mercurial/commands.py -r d130a38 3>&2 [1] MEMORY 49074176 CPUTIME 9.213 REALTIME 9.270 # with "needed" removed MEMORY 637673472 CPUTIME 8.164 REALTIME 8.249 This patch moves "needed" (and "pcache") calculation to a separate DFS to address the issue. It improves perf and fixes issue5360 by correctly reusing hist, while maintaining low memory usage. Some additional attempt has been made to further reduce memory usage, like changing "pcache[f] = []" to "del pcache[f]". Therefore the result can be both faster and lower memory usage: # with this patch applied MEMORY 47575040 CPUTIME 7.870 REALTIME 7.926 [1]: lrun is a lightweight sandbox built on Linux cgroup and namespace. It's used to measure CPU and memory usage here. Source code is available at github.com/quark-zju/lrun.
2016-09-02 17:20:59 +03:00
while visit:
f = visit.pop()
if f in pcache:
continue
pl = parents(f)
pcache[f] = pl
for p in pl:
needed[p] = needed.get(p, 0) + 1
if p not in pcache:
visit.append(p)
# 2nd DFS does the actual annotate
visit[:] = [base]
hist = {}
2006-09-27 20:10:21 +04:00
while visit:
f = visit[-1]
annotate: pre-calculate the "needed" dictionary (issue5360) The "needed" dict is used as a reference counter to free items in the giant "hist" dict. However, currently it is not very accurate and can lead to dropping "hist" items unnecessarily, for example, with the following DAG, -3- / \ 0--1--2--4-- The current algorithm will visit and calculate rev 1 twice, undesired. And it tries to be smart by clearing rev 1's parents: "pcache[1] = []" at the time hist[1] being accessed (note: hist[1] needs to be used twice, by rev 2 and rev 3). It can result in incorrect results if p1 of rev 4 deletes chunks belonging to rev 0. However, simply removing "needed" is not okay, because it will consume 10x memory: # without any change % HGRCPATH= lrun ./hg annotate mercurial/commands.py -r d130a38 3>&2 [1] MEMORY 49074176 CPUTIME 9.213 REALTIME 9.270 # with "needed" removed MEMORY 637673472 CPUTIME 8.164 REALTIME 8.249 This patch moves "needed" (and "pcache") calculation to a separate DFS to address the issue. It improves perf and fixes issue5360 by correctly reusing hist, while maintaining low memory usage. Some additional attempt has been made to further reduce memory usage, like changing "pcache[f] = []" to "del pcache[f]". Therefore the result can be both faster and lower memory usage: # with this patch applied MEMORY 47575040 CPUTIME 7.870 REALTIME 7.926 [1]: lrun is a lightweight sandbox built on Linux cgroup and namespace. It's used to measure CPU and memory usage here. Source code is available at github.com/quark-zju/lrun.
2016-09-02 17:20:59 +03:00
if f in hist:
visit.pop()
continue
ready = True
pl = pcache[f]
for p in pl:
if p not in hist:
ready = False
2006-09-27 20:10:21 +04:00
visit.append(p)
if ready:
visit.pop()
annotate: pre-calculate the "needed" dictionary (issue5360) The "needed" dict is used as a reference counter to free items in the giant "hist" dict. However, currently it is not very accurate and can lead to dropping "hist" items unnecessarily, for example, with the following DAG, -3- / \ 0--1--2--4-- The current algorithm will visit and calculate rev 1 twice, undesired. And it tries to be smart by clearing rev 1's parents: "pcache[1] = []" at the time hist[1] being accessed (note: hist[1] needs to be used twice, by rev 2 and rev 3). It can result in incorrect results if p1 of rev 4 deletes chunks belonging to rev 0. However, simply removing "needed" is not okay, because it will consume 10x memory: # without any change % HGRCPATH= lrun ./hg annotate mercurial/commands.py -r d130a38 3>&2 [1] MEMORY 49074176 CPUTIME 9.213 REALTIME 9.270 # with "needed" removed MEMORY 637673472 CPUTIME 8.164 REALTIME 8.249 This patch moves "needed" (and "pcache") calculation to a separate DFS to address the issue. It improves perf and fixes issue5360 by correctly reusing hist, while maintaining low memory usage. Some additional attempt has been made to further reduce memory usage, like changing "pcache[f] = []" to "del pcache[f]". Therefore the result can be both faster and lower memory usage: # with this patch applied MEMORY 47575040 CPUTIME 7.870 REALTIME 7.926 [1]: lrun is a lightweight sandbox built on Linux cgroup and namespace. It's used to measure CPU and memory usage here. Source code is available at github.com/quark-zju/lrun.
2016-09-02 17:20:59 +03:00
curr = decorate(f.data(), f)
skipchild = False
if skiprevs is not None:
skipchild = f._changeid in skiprevs
curr = _annotatepair([hist[p] for p in pl], f, curr, skipchild,
diffopts)
for p in pl:
if needed[p] == 1:
del hist[p]
del needed[p]
else:
needed[p] -= 1
hist[f] = curr
annotate: pre-calculate the "needed" dictionary (issue5360) The "needed" dict is used as a reference counter to free items in the giant "hist" dict. However, currently it is not very accurate and can lead to dropping "hist" items unnecessarily, for example, with the following DAG, -3- / \ 0--1--2--4-- The current algorithm will visit and calculate rev 1 twice, undesired. And it tries to be smart by clearing rev 1's parents: "pcache[1] = []" at the time hist[1] being accessed (note: hist[1] needs to be used twice, by rev 2 and rev 3). It can result in incorrect results if p1 of rev 4 deletes chunks belonging to rev 0. However, simply removing "needed" is not okay, because it will consume 10x memory: # without any change % HGRCPATH= lrun ./hg annotate mercurial/commands.py -r d130a38 3>&2 [1] MEMORY 49074176 CPUTIME 9.213 REALTIME 9.270 # with "needed" removed MEMORY 637673472 CPUTIME 8.164 REALTIME 8.249 This patch moves "needed" (and "pcache") calculation to a separate DFS to address the issue. It improves perf and fixes issue5360 by correctly reusing hist, while maintaining low memory usage. Some additional attempt has been made to further reduce memory usage, like changing "pcache[f] = []" to "del pcache[f]". Therefore the result can be both faster and lower memory usage: # with this patch applied MEMORY 47575040 CPUTIME 7.870 REALTIME 7.926 [1]: lrun is a lightweight sandbox built on Linux cgroup and namespace. It's used to measure CPU and memory usage here. Source code is available at github.com/quark-zju/lrun.
2016-09-02 17:20:59 +03:00
del pcache[f]
return zip(hist[base][0], hist[base][1].splitlines(True))
def ancestors(self, followfirst=False):
visit = {}
c = self
if followfirst:
cut = 1
else:
cut = None
while True:
for parent in c.parents()[:cut]:
visit[(parent.linkrev(), parent.filenode())] = parent
if not visit:
break
c = visit.pop(max(visit))
yield c
def decodeddata(self):
"""Returns `data()` after running repository decoding filters.
This is often equivalent to how the data would be expressed on disk.
"""
return self._repo.wwritedata(self.path(), self.data())
def _annotatepair(parents, childfctx, child, skipchild, diffopts):
r'''
Given parent and child fctxes and annotate data for parents, for all lines
in either parent that match the child, annotate the child with the parent's
data.
Additionally, if `skipchild` is True, replace all other lines with parent
annotate data as well such that child is never blamed for any lines.
>>> oldfctx = 'old'
>>> p1fctx, p2fctx, childfctx = 'p1', 'p2', 'c'
>>> olddata = 'a\nb\n'
>>> p1data = 'a\nb\nc\n'
>>> p2data = 'a\nc\nd\n'
>>> childdata = 'a\nb2\nc\nc2\nd\n'
>>> diffopts = mdiff.diffopts()
>>> def decorate(text, rev):
... return ([(rev, i) for i in xrange(1, text.count('\n') + 1)], text)
Basic usage:
>>> oldann = decorate(olddata, oldfctx)
>>> p1ann = decorate(p1data, p1fctx)
>>> p1ann = _annotatepair([oldann], p1fctx, p1ann, False, diffopts)
>>> p1ann[0]
[('old', 1), ('old', 2), ('p1', 3)]
>>> p2ann = decorate(p2data, p2fctx)
>>> p2ann = _annotatepair([oldann], p2fctx, p2ann, False, diffopts)
>>> p2ann[0]
[('old', 1), ('p2', 2), ('p2', 3)]
Test with multiple parents (note the difference caused by ordering):
>>> childann = decorate(childdata, childfctx)
>>> childann = _annotatepair([p1ann, p2ann], childfctx, childann, False,
... diffopts)
>>> childann[0]
[('old', 1), ('c', 2), ('p2', 2), ('c', 4), ('p2', 3)]
>>> childann = decorate(childdata, childfctx)
>>> childann = _annotatepair([p2ann, p1ann], childfctx, childann, False,
... diffopts)
>>> childann[0]
[('old', 1), ('c', 2), ('p1', 3), ('c', 4), ('p2', 3)]
Test with skipchild (note the difference caused by ordering):
>>> childann = decorate(childdata, childfctx)
>>> childann = _annotatepair([p1ann, p2ann], childfctx, childann, True,
... diffopts)
>>> childann[0]
[('old', 1), ('old', 2), ('p2', 2), ('p2', 2), ('p2', 3)]
>>> childann = decorate(childdata, childfctx)
>>> childann = _annotatepair([p2ann, p1ann], childfctx, childann, True,
... diffopts)
>>> childann[0]
[('old', 1), ('old', 2), ('p1', 3), ('p1', 3), ('p2', 3)]
'''
pblocks = [(parent, mdiff.allblocks(parent[1], child[1], opts=diffopts))
for parent in parents]
if skipchild:
# Need to iterate over the blocks twice -- make it a list
pblocks = [(p, list(blocks)) for (p, blocks) in pblocks]
# Mercurial currently prefers p2 over p1 for annotate.
# TODO: change this?
for parent, blocks in pblocks:
for (a1, a2, b1, b2), t in blocks:
# Changed blocks ('!') or blocks made only of blank lines ('~')
# belong to the child.
if t == '=':
child[0][b1:b2] = parent[0][a1:a2]
if skipchild:
# Now try and match up anything that couldn't be matched,
# Reversing pblocks maintains bias towards p2, matching above
# behavior.
pblocks.reverse()
# The heuristics are:
# * Work on blocks of changed lines (effectively diff hunks with -U0).
# This could potentially be smarter but works well enough.
# * For a non-matching section, do a best-effort fit. Match lines in
# diff hunks 1:1, dropping lines as necessary.
# * Repeat the last line as a last resort.
# First, replace as much as possible without repeating the last line.
remaining = [(parent, []) for parent, _blocks in pblocks]
for idx, (parent, blocks) in enumerate(pblocks):
for (a1, a2, b1, b2), _t in blocks:
if a2 - a1 >= b2 - b1:
for bk in xrange(b1, b2):
if child[0][bk][0] == childfctx:
ak = min(a1 + (bk - b1), a2 - 1)
child[0][bk] = parent[0][ak]
else:
remaining[idx][1].append((a1, a2, b1, b2))
# Then, look at anything left, which might involve repeating the last
# line.
for parent, blocks in remaining:
for a1, a2, b1, b2 in blocks:
for bk in xrange(b1, b2):
if child[0][bk][0] == childfctx:
ak = min(a1 + (bk - b1), a2 - 1)
child[0][bk] = parent[0][ak]
return child
class filectx(basefilectx):
"""A filecontext object makes access to data related to a particular
filerevision convenient."""
def __init__(self, repo, path, changeid=None, fileid=None,
filelog=None, changectx=None):
"""changeid can be a changeset revision, node, or tag.
fileid can be a file revision or node."""
self._repo = repo
self._path = path
assert (changeid is not None
or fileid is not None
or changectx is not None), \
("bad args: changeid=%r, fileid=%r, changectx=%r"
% (changeid, fileid, changectx))
if filelog is not None:
self._filelog = filelog
if changeid is not None:
self._changeid = changeid
if changectx is not None:
self._changectx = changectx
if fileid is not None:
self._fileid = fileid
@propertycache
def _changectx(self):
try:
return changectx(self._repo, self._changeid)
except error.FilteredRepoLookupError:
# Linkrev may point to any revision in the repository. When the
# repository is filtered this may lead to `filectx` trying to build
# `changectx` for filtered revision. In such case we fallback to
# creating `changectx` on the unfiltered version of the reposition.
# This fallback should not be an issue because `changectx` from
# `filectx` are not used in complex operations that care about
# filtering.
#
# This fallback is a cheap and dirty fix that prevent several
# crashes. It does not ensure the behavior is correct. However the
# behavior was not correct before filtering either and "incorrect
# behavior" is seen as better as "crash"
#
# Linkrevs have several serious troubles with filtering that are
# complicated to solve. Proper handling of the issue here should be
# considered when solving linkrev issue are on the table.
return changectx(self._repo.unfiltered(), self._changeid)
def filectx(self, fileid, changeid=None):
'''opens an arbitrary revision of the file without
opening a new filelog'''
return filectx(self._repo, self._path, fileid=fileid,
filelog=self._filelog, changeid=changeid)
def rawdata(self):
return self._filelog.revision(self._filenode, raw=True)
def rawflags(self):
"""low-level revlog flags"""
return self._filelog.flags(self._filerev)
def data(self):
try:
return self._filelog.read(self._filenode)
except error.CensoredNodeError:
codemod: register core configitems using a script This is done by a script [2] using RedBaron [1], a tool designed for doing code refactoring. All "default" values are decided by the script and are strongly consistent with the existing code. There are 2 changes done manually to fix tests: [warn] mercurial/exchange.py: experimental.bundle2-output-capture: default needs manual removal [warn] mercurial/localrepo.py: experimental.hook-track-tags: default needs manual removal Since RedBaron is not confident about how to indent things [2]. [1]: https://github.com/PyCQA/redbaron [2]: https://github.com/PyCQA/redbaron/issues/100 [3]: #!/usr/bin/env python # codemod_configitems.py - codemod tool to fill configitems # # Copyright 2017 Facebook, Inc. # # This software may be used and distributed according to the terms of the # GNU General Public License version 2 or any later version. from __future__ import absolute_import, print_function import os import sys import redbaron def readpath(path): with open(path) as f: return f.read() def writepath(path, content): with open(path, 'w') as f: f.write(content) _configmethods = {'config', 'configbool', 'configint', 'configbytes', 'configlist', 'configdate'} def extractstring(rnode): """get the string from a RedBaron string or call_argument node""" while rnode.type != 'string': rnode = rnode.value return rnode.value[1:-1] # unquote, "'str'" -> "str" def uiconfigitems(red): """match *.ui.config* pattern, yield (node, method, args, section, name)""" for node in red.find_all('atomtrailers'): entry = None try: obj = node[-3].value method = node[-2].value args = node[-1] section = args[0].value name = args[1].value if (obj in ('ui', 'self') and method in _configmethods and section.type == 'string' and name.type == 'string'): entry = (node, method, args, extractstring(section), extractstring(name)) except Exception: pass else: if entry: yield entry def coreconfigitems(red): """match coreconfigitem(...) pattern, yield (node, args, section, name)""" for node in red.find_all('atomtrailers'): entry = None try: args = node[1] section = args[0].value name = args[1].value if (node[0].value == 'coreconfigitem' and section.type == 'string' and name.type == 'string'): entry = (node, args, extractstring(section), extractstring(name)) except Exception: pass else: if entry: yield entry def registercoreconfig(cfgred, section, name, defaultrepr): """insert coreconfigitem to cfgred AST section and name are plain string, defaultrepr is a string """ # find a place to insert the "coreconfigitem" item entries = list(coreconfigitems(cfgred)) for node, args, nodesection, nodename in reversed(entries): if (nodesection, nodename) < (section, name): # insert after this entry node.insert_after( 'coreconfigitem(%r, %r,\n' ' default=%s,\n' ')' % (section, name, defaultrepr)) return def main(argv): if not argv: print('Usage: codemod_configitems.py FILES\n' 'For example, FILES could be "{hgext,mercurial}/*/**.py"') dirname = os.path.dirname reporoot = dirname(dirname(dirname(os.path.abspath(__file__)))) # register configitems to this destination cfgpath = os.path.join(reporoot, 'mercurial', 'configitems.py') cfgred = redbaron.RedBaron(readpath(cfgpath)) # state about what to do registered = set((s, n) for n, a, s, n in coreconfigitems(cfgred)) toregister = {} # {(section, name): defaultrepr} coreconfigs = set() # {(section, name)}, whether it's used in core # first loop: scan all files before taking any action for i, path in enumerate(argv): print('(%d/%d) scanning %s' % (i + 1, len(argv), path)) iscore = ('mercurial' in path) and ('hgext' not in path) red = redbaron.RedBaron(readpath(path)) # find all repo.ui.config* and ui.config* calls, and collect their # section, name and default value information. for node, method, args, section, name in uiconfigitems(red): if section == 'web': # [web] section has some weirdness, ignore them for now continue defaultrepr = None key = (section, name) if len(args) == 2: if key in registered: continue if method == 'configlist': defaultrepr = 'list' elif method == 'configbool': defaultrepr = 'False' else: defaultrepr = 'None' elif len(args) >= 3 and (args[2].target is None or args[2].target.value == 'default'): # try to understand the "default" value dnode = args[2].value if dnode.type == 'name': if dnode.value in {'None', 'True', 'False'}: defaultrepr = dnode.value elif dnode.type == 'string': defaultrepr = repr(dnode.value[1:-1]) elif dnode.type in ('int', 'float'): defaultrepr = dnode.value # inconsistent default if key in toregister and toregister[key] != defaultrepr: defaultrepr = None # interesting to rewrite if key not in registered: if defaultrepr is None: print('[note] %s: %s.%s: unsupported default' % (path, section, name)) registered.add(key) # skip checking it again else: toregister[key] = defaultrepr if iscore: coreconfigs.add(key) # second loop: rewrite files given "toregister" result for path in argv: # reconstruct redbaron - trade CPU for memory red = redbaron.RedBaron(readpath(path)) changed = False for node, method, args, section, name in uiconfigitems(red): key = (section, name) defaultrepr = toregister.get(key) if defaultrepr is None or key not in coreconfigs: continue if len(args) >= 3 and (args[2].target is None or args[2].target.value == 'default'): try: del args[2] changed = True except Exception: # redbaron fails to do the rewrite due to indentation # see https://github.com/PyCQA/redbaron/issues/100 print('[warn] %s: %s.%s: default needs manual removal' % (path, section, name)) if key not in registered: print('registering %s.%s' % (section, name)) registercoreconfig(cfgred, section, name, defaultrepr) registered.add(key) if changed: print('updating %s' % path) writepath(path, red.dumps()) if toregister: print('updating configitems.py') writepath(cfgpath, cfgred.dumps()) if __name__ == "__main__": sys.exit(main(sys.argv[1:]))
2017-07-15 00:22:40 +03:00
if self._repo.ui.config("censor", "policy") == "ignore":
return ""
raise error.Abort(_("censored node: %s") % short(self._filenode),
hint=_("set censor.policy to ignore errors"))
def size(self):
return self._filelog.size(self._filerev)
@propertycache
def _copied(self):
"""check if file was actually renamed in this changeset revision
If rename logged in file revision, we report copy for changeset only
if file revisions linkrev points back to the changeset in question
or both changeset parents contain different file revisions.
"""
renamed = self._filelog.renamed(self._filenode)
if not renamed:
return renamed
if self.rev() == self.linkrev():
return renamed
name = self.path()
fnode = self._filenode
for p in self._changectx.parents():
try:
if fnode == p.filenode(name):
return None
except error.LookupError:
pass
return renamed
def children(self):
# hard for renames
c = self._filelog.children(self._filenode)
return [filectx(self._repo, self._path, fileid=x,
filelog=self._filelog) for x in c]
class committablectx(basectx):
"""A committablectx object provides common functionality for a context that
wants the ability to commit, e.g. workingctx or memctx."""
def __init__(self, repo, text="", user=None, date=None, extra=None,
changes=None):
self._repo = repo
self._rev = None
self._node = None
self._text = text
if date:
self._date = util.parsedate(date)
2008-07-22 22:00:22 +04:00
if user:
self._user = user
if changes:
self._status = changes
self._extra = {}
if extra:
self._extra = extra.copy()
if 'branch' not in self._extra:
try:
branch = encoding.fromlocal(self._repo.dirstate.branch())
except UnicodeDecodeError:
raise error.Abort(_('branch name not in UTF-8!'))
self._extra['branch'] = branch
if self._extra['branch'] == '':
self._extra['branch'] = 'default'
def __bytes__(self):
return bytes(self._parents[0]) + "+"
__str__ = encoding.strmethod(__bytes__)
def __nonzero__(self):
return True
__bool__ = __nonzero__
def _buildflagfunc(self):
# Create a fallback function for getting file flags when the
# filesystem doesn't support them
copiesget = self._repo.dirstate.copies().get
parents = self.parents()
if len(parents) < 2:
# when we have one parent, it's easy: copy from parent
man = parents[0].manifest()
def func(f):
f = copiesget(f, f)
return man.flags(f)
else:
# merges are tricky: we try to reconstruct the unstored
# result from the merge (issue1802)
p1, p2 = parents
pa = p1.ancestor(p2)
m1, m2, ma = p1.manifest(), p2.manifest(), pa.manifest()
def func(f):
f = copiesget(f, f) # may be wrong for merges with copies
fl1, fl2, fla = m1.flags(f), m2.flags(f), ma.flags(f)
if fl1 == fl2:
return fl1
if fl1 == fla:
return fl2
if fl2 == fla:
return fl1
return '' # punt for conflicts
return func
@propertycache
def _flagfunc(self):
return self._repo.dirstate.flagfunc(self._buildflagfunc)
@propertycache
def _status(self):
return self._repo.status()
@propertycache
def _user(self):
return self._repo.ui.username()
@propertycache
def _date(self):
ui = self._repo.ui
date = ui.configdate('devel', 'default-date')
if date is None:
date = util.makedate()
return date
def subrev(self, subpath):
return None
def manifestnode(self):
return None
def user(self):
return self._user or self._repo.ui.username()
def date(self):
return self._date
def description(self):
return self._text
def files(self):
return sorted(self._status.modified + self._status.added +
self._status.removed)
def modified(self):
return self._status.modified
def added(self):
return self._status.added
def removed(self):
return self._status.removed
def deleted(self):
return self._status.deleted
def branch(self):
return encoding.tolocal(self._extra['branch'])
def closesbranch(self):
return 'close' in self._extra
def extra(self):
return self._extra
def tags(self):
return []
def bookmarks(self):
b = []
for p in self.parents():
b.extend(p.bookmarks())
return b
def phase(self):
phase = phases.draft # default phase to draft
for p in self.parents():
phase = max(phase, p.phase())
return phase
def hidden(self):
return False
def children(self):
return []
def flags(self, path):
if r'_manifest' in self.__dict__:
try:
return self._manifest.flags(path)
except KeyError:
return ''
try:
return self._flagfunc(path)
except OSError:
return ''
def ancestor(self, c2):
"""return the "best" ancestor context of self and c2"""
return self._parents[0].ancestor(c2) # punt on two parents for now
def walk(self, match):
'''Generates matching file names.'''
return sorted(self._repo.dirstate.walk(match, sorted(self.substate),
True, False))
def matches(self, match):
return sorted(self._repo.dirstate.matches(match))
def ancestors(self):
for p in self._parents:
yield p
for a in self._repo.changelog.ancestors(
[p.rev() for p in self._parents]):
yield changectx(self._repo, a)
def markcommitted(self, node):
"""Perform post-commit cleanup necessary after committing this ctx
Specifically, this updates backing stores this working context
wraps to reflect the fact that the changes reflected by this
workingctx have been committed. For example, it marks
modified and added files as normal in the dirstate.
"""
with self._repo.dirstate.parentchange():
for f in self.modified() + self.added():
self._repo.dirstate.normal(f)
for f in self.removed():
self._repo.dirstate.drop(f)
self._repo.dirstate.setparents(node)
context: write dirstate out explicitly at the end of markcommitted To detect change of a file without redundant comparison of file content, dirstate recognizes a file as certainly clean, if: (1) it is already known as "normal", (2) dirstate entry for it has valid (= not "-1") timestamp, and (3) mode, size and timestamp of it on the filesystem are as same as ones expected in dirstate This works as expected in many cases, but doesn't in the corner case that changing a file keeps mode, size and timestamp of it on the filesystem. The timetable below shows steps in one of typical such situations: ---- ----------------------------------- ---------------- timestamp of "f" ---------------- dirstate file- time action mem file system ---- ----------------------------------- ---- ----- ----- * *** *** - 'hg transplant REV1 REV2 ...' - transplanting REV1 .... N - change "f", but keep size N (via 'patch.patch()') - 'dirstate.normal("f")' N *** (via 'repo.commit()') - transplanting REV2 - change "f", but keep size N (via 'patch.patch()') - aborted while patching N+1 - release wlock - 'dirstate.write()' N N N - 'hg status' shows "r1" as "clean" N N N ---- ----------------------------------- ---- ----- ----- The most important point is that 'dirstate.write()' is executed at N+1 or later. This causes writing dirstate timestamp N of "f" out successfully. If it is executed at N, 'parsers.pack_dirstate()' replaces timestamp N with "-1" before actual writing dirstate out. This issue can occur when 'hg transplant' satisfies conditions below: - multiple revisions to be transplanted change the same file - those revisions don't change mode and size of the file, and - the 2nd or later revision of them fails after changing the file The root cause of this issue is that files are changed without flushing in-memory dirstate changes via 'repo.commit()' (even though omitting 'dirstate.normallookup()' on files changed by 'patch.patch()' for efficiency also causes this issue). To detect changes of files correctly, this patch writes in-memory dirstate changes out explicitly after marking files as clean in 'committablectx.markcommitted()', which is invoked via 'repo.commit()'. After this change, timetable is changed as below: ---- ----------------------------------- ---------------- timestamp of "f" ---------------- dirstate file- time action mem file system ---- ----------------------------------- ---- ----- ----- * *** *** - 'hg transplant REV1 REV2 ...' - transplanting REV1 .... N - change "f", but keep size N (via 'patch.patch()') - 'dirstate.normal("f")' N *** (via 'repo.commit()') ----------------------------------- ---- ----- ----- - 'dirsttate.write()' -1 -1 ----------------------------------- ---- ----- ----- - transplanting REV2 - change "f", but keep size N (via 'patch.patch()') - aborted while patching N+1 - release wlock - 'dirstate.write()' -1 -1 N - 'hg status' shows "r1" as "clean" -1 -1 N ---- ----------------------------------- ---- ----- ----- To reproduce this issue in tests certainly, this patch emulates some timing critical actions as below: - change "f" at N 'patch.patch()' with 'fakepatchtime.py' explicitly changes mtime of patched files to "2000-01-01 00:00" (= N). - 'dirstate.write()' via 'repo.commit()' at N 'fakedirstatewritetime.py' forces 'pack_dirstate()' to use "2000-01-01 00:00" as "now", only if 'pack_dirstate()' is invoked via 'committablectx.markcommitted()'. - 'dirstate.write()' via releasing wlock at N+1 (or "not at N") 'pack_dirstate()' via releasing wlock uses actual timestamp at runtime as "now", and it should be different from the "2000-01-01 00:00" of "f". BTW, this patch doesn't test cases below, even though 'patch.patch()' is used similarly in these cases: 1. failure of 'hg import' or 'hg qpush' 2. success of 'hg import', 'hg qpush' or 'hg transplant' Case (1) above doesn't cause this kind of issue, because: - if patching is aborted by conflicts, changed files are committed changed files are marked as CLEAN, even though they are partially patched. - otherwise, dirstate are fully restored by 'dirstateguard' For example in timetable above, timestamp of "f" in .hg/dirstate is restored to -1 (or less than N), and subsequent 'hg status' can detect changes correctly. Case (2) always causes 'repo.status()' invocation via 'repo.commit()' just after changing files inside same wlock scope. ---- ----------------------------------- ---------------- timestamp of "f" ---------------- dirstate file- time action mem file system ---- ----------------------------------- ---- ----- ----- N *** *** - make file "f" clean N - execute 'hg foobar' .... - 'dirstate.normal("f")' N *** (e.g. via dirty check or previous 'repo.commit()') - change "f", but keep size N - 'repo.status()' (*1) (via 'repo.commit()') ---- ----------------------------------- ---- ----- ----- At a glance, 'repo.status()' at (*1) seems to cause similar issue (= "changed files are treated as clean"), but actually doesn't. 'dirstate._lastnormaltime' should be N at (*1) above, because 'dirstate.normal()' via dirty check is finished at N. Therefore, "f" changed at N (= 'dirstate._lastnormaltime') is forcibly treated as "unsure" at (*1), and changes are detected as expected (see 'dirstate.status()' for detail). If 'hg import' is executed with '--no-commit', 'repo.status()' isn't invoked just after changing files inside same wlock scope. But preceding 'dirstate.normal()' is invoked inside another wlock scope via 'cmdutil.bailifchanged()', and in-memory changes should be flushed at the end of that scope. Therefore, timestamp N of clean "f" should be replaced by -1, if 'dirstate.write()' is invoked at N. It means that condition of this issue isn't satisfied.
2015-07-08 11:01:09 +03:00
# write changes out explicitly, because nesting wlock at
# runtime may prevent 'wlock.release()' in 'repo.commit()'
# from immediately doing so for subsequent changing files
self._repo.dirstate.write(self._repo.currenttransaction())
context: write dirstate out explicitly at the end of markcommitted To detect change of a file without redundant comparison of file content, dirstate recognizes a file as certainly clean, if: (1) it is already known as "normal", (2) dirstate entry for it has valid (= not "-1") timestamp, and (3) mode, size and timestamp of it on the filesystem are as same as ones expected in dirstate This works as expected in many cases, but doesn't in the corner case that changing a file keeps mode, size and timestamp of it on the filesystem. The timetable below shows steps in one of typical such situations: ---- ----------------------------------- ---------------- timestamp of "f" ---------------- dirstate file- time action mem file system ---- ----------------------------------- ---- ----- ----- * *** *** - 'hg transplant REV1 REV2 ...' - transplanting REV1 .... N - change "f", but keep size N (via 'patch.patch()') - 'dirstate.normal("f")' N *** (via 'repo.commit()') - transplanting REV2 - change "f", but keep size N (via 'patch.patch()') - aborted while patching N+1 - release wlock - 'dirstate.write()' N N N - 'hg status' shows "r1" as "clean" N N N ---- ----------------------------------- ---- ----- ----- The most important point is that 'dirstate.write()' is executed at N+1 or later. This causes writing dirstate timestamp N of "f" out successfully. If it is executed at N, 'parsers.pack_dirstate()' replaces timestamp N with "-1" before actual writing dirstate out. This issue can occur when 'hg transplant' satisfies conditions below: - multiple revisions to be transplanted change the same file - those revisions don't change mode and size of the file, and - the 2nd or later revision of them fails after changing the file The root cause of this issue is that files are changed without flushing in-memory dirstate changes via 'repo.commit()' (even though omitting 'dirstate.normallookup()' on files changed by 'patch.patch()' for efficiency also causes this issue). To detect changes of files correctly, this patch writes in-memory dirstate changes out explicitly after marking files as clean in 'committablectx.markcommitted()', which is invoked via 'repo.commit()'. After this change, timetable is changed as below: ---- ----------------------------------- ---------------- timestamp of "f" ---------------- dirstate file- time action mem file system ---- ----------------------------------- ---- ----- ----- * *** *** - 'hg transplant REV1 REV2 ...' - transplanting REV1 .... N - change "f", but keep size N (via 'patch.patch()') - 'dirstate.normal("f")' N *** (via 'repo.commit()') ----------------------------------- ---- ----- ----- - 'dirsttate.write()' -1 -1 ----------------------------------- ---- ----- ----- - transplanting REV2 - change "f", but keep size N (via 'patch.patch()') - aborted while patching N+1 - release wlock - 'dirstate.write()' -1 -1 N - 'hg status' shows "r1" as "clean" -1 -1 N ---- ----------------------------------- ---- ----- ----- To reproduce this issue in tests certainly, this patch emulates some timing critical actions as below: - change "f" at N 'patch.patch()' with 'fakepatchtime.py' explicitly changes mtime of patched files to "2000-01-01 00:00" (= N). - 'dirstate.write()' via 'repo.commit()' at N 'fakedirstatewritetime.py' forces 'pack_dirstate()' to use "2000-01-01 00:00" as "now", only if 'pack_dirstate()' is invoked via 'committablectx.markcommitted()'. - 'dirstate.write()' via releasing wlock at N+1 (or "not at N") 'pack_dirstate()' via releasing wlock uses actual timestamp at runtime as "now", and it should be different from the "2000-01-01 00:00" of "f". BTW, this patch doesn't test cases below, even though 'patch.patch()' is used similarly in these cases: 1. failure of 'hg import' or 'hg qpush' 2. success of 'hg import', 'hg qpush' or 'hg transplant' Case (1) above doesn't cause this kind of issue, because: - if patching is aborted by conflicts, changed files are committed changed files are marked as CLEAN, even though they are partially patched. - otherwise, dirstate are fully restored by 'dirstateguard' For example in timetable above, timestamp of "f" in .hg/dirstate is restored to -1 (or less than N), and subsequent 'hg status' can detect changes correctly. Case (2) always causes 'repo.status()' invocation via 'repo.commit()' just after changing files inside same wlock scope. ---- ----------------------------------- ---------------- timestamp of "f" ---------------- dirstate file- time action mem file system ---- ----------------------------------- ---- ----- ----- N *** *** - make file "f" clean N - execute 'hg foobar' .... - 'dirstate.normal("f")' N *** (e.g. via dirty check or previous 'repo.commit()') - change "f", but keep size N - 'repo.status()' (*1) (via 'repo.commit()') ---- ----------------------------------- ---- ----- ----- At a glance, 'repo.status()' at (*1) seems to cause similar issue (= "changed files are treated as clean"), but actually doesn't. 'dirstate._lastnormaltime' should be N at (*1) above, because 'dirstate.normal()' via dirty check is finished at N. Therefore, "f" changed at N (= 'dirstate._lastnormaltime') is forcibly treated as "unsure" at (*1), and changes are detected as expected (see 'dirstate.status()' for detail). If 'hg import' is executed with '--no-commit', 'repo.status()' isn't invoked just after changing files inside same wlock scope. But preceding 'dirstate.normal()' is invoked inside another wlock scope via 'cmdutil.bailifchanged()', and in-memory changes should be flushed at the end of that scope. Therefore, timestamp N of clean "f" should be replaced by -1, if 'dirstate.write()' is invoked at N. It means that condition of this issue isn't satisfied.
2015-07-08 11:01:09 +03:00
def dirty(self, missing=False, merge=True, branch=True):
return False
class workingctx(committablectx):
"""A workingctx object makes access to data related to
the current working directory convenient.
date - any valid date string or (unixtime, offset), or None.
user - username string, or None.
extra - a dictionary of extra values, or None.
changes - a list of file lists as returned by localrepo.status()
or None to use the repository status.
"""
def __init__(self, repo, text="", user=None, date=None, extra=None,
changes=None):
super(workingctx, self).__init__(repo, text, user, date, extra, changes)
def __iter__(self):
d = self._repo.dirstate
for f in d:
if d[f] != 'r':
yield f
def __contains__(self, key):
return self._repo.dirstate[key] not in "?r"
def hex(self):
2015-06-22 16:05:10 +03:00
return hex(wdirid)
@propertycache
def _parents(self):
p = self._repo.dirstate.parents()
if p[1] == nullid:
p = p[:-1]
2012-08-02 19:48:58 +04:00
return [changectx(self._repo, x) for x in p]
def filectx(self, path, filelog=None):
"""get a file context from the working directory"""
return workingfilectx(self._repo, path, workingctx=self,
filelog=filelog)
def dirty(self, missing=False, merge=True, branch=True):
"check whether a working directory is modified"
# check subrepos first
for s in sorted(self.substate):
if self.sub(s).dirty(missing=missing):
return True
# check current working dir
return ((merge and self.p2()) or
(branch and self.branch() != self.p1().branch()) or
self.modified() or self.added() or self.removed() or
(missing and self.deleted()))
def add(self, list, prefix=""):
with self._repo.wlock():
ui, ds = self._repo.ui, self._repo.dirstate
uipath = lambda f: ds.pathto(pathutil.join(prefix, f))
rejected = []
lstat = self._repo.wvfs.lstat
for f in list:
# ds.pathto() returns an absolute file when this is invoked from
# the keyword extension. That gets flagged as non-portable on
# Windows, since it contains the drive letter and colon.
scmutil.checkportable(ui, os.path.join(prefix, f))
try:
st = lstat(f)
2011-04-23 01:51:25 +04:00
except OSError:
ui.warn(_("%s does not exist!\n") % uipath(f))
rejected.append(f)
continue
if st.st_size > 10000000:
ui.warn(_("%s: up to %d MB of RAM may be required "
"to manage this file\n"
"(use 'hg revert %s' to cancel the "
"pending addition)\n")
% (f, 3 * st.st_size // 1000000, uipath(f)))
if not (stat.S_ISREG(st.st_mode) or stat.S_ISLNK(st.st_mode)):
ui.warn(_("%s not added: only files and symlinks "
"supported currently\n") % uipath(f))
rejected.append(f)
elif ds[f] in 'amn':
ui.warn(_("%s already tracked!\n") % uipath(f))
elif ds[f] == 'r':
ds.normallookup(f)
else:
ds.add(f)
return rejected
def forget(self, files, prefix=""):
with self._repo.wlock():
ds = self._repo.dirstate
uipath = lambda f: ds.pathto(pathutil.join(prefix, f))
rejected = []
for f in files:
if f not in self._repo.dirstate:
self._repo.ui.warn(_("%s not tracked!\n") % uipath(f))
rejected.append(f)
elif self._repo.dirstate[f] != 'a':
self._repo.dirstate.remove(f)
else:
self._repo.dirstate.drop(f)
return rejected
def undelete(self, list):
pctxs = self.parents()
with self._repo.wlock():
ds = self._repo.dirstate
for f in list:
if self._repo.dirstate[f] != 'r':
self._repo.ui.warn(_("%s not removed!\n") % ds.pathto(f))
else:
fctx = f in pctxs[0] and pctxs[0][f] or pctxs[1][f]
t = fctx.data()
self._repo.wwrite(f, t, fctx.flags())
self._repo.dirstate.normal(f)
def copy(self, source, dest):
try:
st = self._repo.wvfs.lstat(dest)
except OSError as err:
if err.errno != errno.ENOENT:
raise
self._repo.ui.warn(_("%s does not exist!\n")
% self._repo.dirstate.pathto(dest))
return
if not (stat.S_ISREG(st.st_mode) or stat.S_ISLNK(st.st_mode)):
self._repo.ui.warn(_("copy failed: %s is not a file or a "
"symbolic link\n")
% self._repo.dirstate.pathto(dest))
else:
with self._repo.wlock():
if self._repo.dirstate[dest] in '?':
self._repo.dirstate.add(dest)
elif self._repo.dirstate[dest] in 'r':
self._repo.dirstate.normallookup(dest)
self._repo.dirstate.copy(source, dest)
def match(self, pats=None, include=None, exclude=None, default='glob',
listsubrepos=False, badfn=None):
match: add a subclass for dirstate normalizing of the matched patterns This class is only needed on case insensitive filesystems, and only for wdir context matches. It allows the user to not match the case of the items in the filesystem- especially for naming directories, which dirstate doesn't handle[1]. Making dirstate handle mismatched directory cases is too expensive[2]. Since dirstate doesn't apply to committed csets, this is only created by overriding basectx.match() in workingctx, and only on icasefs. The default arguments have been dropped, because the ctx must be passed to the matcher in order to function. For operations that can apply to both wdir and some other context, this ends up normalizing the filename to the case as it exists in the filesystem, and using that case for the lookup in the other context. See the diff example in the test. Previously, given a directory with an inexact case: - add worked as expected - diff, forget and status would silently ignore the request - files would exit with 1 - commit, revert and remove would fail (even when the commands leading up to them worked): $ hg ci -m "AbCDef" capsdir1/capsdir abort: CapsDir1/CapsDir: no match under directory! $ hg revert -r '.^' capsdir1/capsdir capsdir1\capsdir: no such file in rev 64dae27060b7 $ hg remove capsdir1/capsdir not removing capsdir1\capsdir: no tracked files [1] Globs are normalized, so that the -I and -X don't need to be specified with a case match. Without that, the second last remove (with -X) removes the files, leaving nothing for the last remove. However, specifying the files as 'glob:**.Txt' does not work. Perhaps this requires 're.IGNORECASE'? There are only a handful of places that create matchers directly, instead of being routed through the context.match() method. Some may benefit from changing over to using ctx.match() as a factory function: revset.checkstatus() revset.contains() revset.filelog() revset._matchfiles() localrepository._loadfilter() ignore.ignore() fileset.subrepo() filemerge._picktool() overrides.addlargefiles() lfcommands.lfconvert() kwtemplate.__init__() eolfile.__init__() eolfile.checkrev() acl.buildmatch() Currently, a toplevel subrepo can be named with an inexact case. However, the path auditor gets in the way of naming _anything_ in the subrepo if the top level case doesn't match. That is trickier to handle, because there's the user provided case, the case in the filesystem, and the case stored in .hgsub. This can be fixed next cycle. --- a/tests/test-subrepo-deep-nested-change.t +++ b/tests/test-subrepo-deep-nested-change.t @@ -170,8 +170,15 @@ R sub1/sub2/test.txt $ hg update -Cq $ touch sub1/sub2/folder/bar +#if icasefs + $ hg addremove Sub1/sub2 + abort: path 'Sub1\sub2' is inside nested repo 'Sub1' + [255] + $ hg -q addremove sub1/sub2 +#else $ hg addremove sub1/sub2 adding sub1/sub2/folder/bar (glob) +#endif $ hg status -S A sub1/sub2/folder/bar ? foo/bar/abc The narrowmatcher class may need to be tweaked when that is fixed. [1] http://www.selenic.com/pipermail/mercurial-devel/2015-April/068183.html [2] http://www.selenic.com/pipermail/mercurial-devel/2015-April/068191.html
2015-04-12 08:39:21 +03:00
r = self._repo
# Only a case insensitive filesystem needs magic to translate user input
# to actual case in the filesystem.
icasefs = not util.fscasesensitive(r.root)
return matchmod.match(r.root, r.getcwd(), pats, include, exclude,
default, auditor=r.auditor, ctx=self,
listsubrepos=listsubrepos, badfn=badfn,
icasefs=icasefs)
match: add a subclass for dirstate normalizing of the matched patterns This class is only needed on case insensitive filesystems, and only for wdir context matches. It allows the user to not match the case of the items in the filesystem- especially for naming directories, which dirstate doesn't handle[1]. Making dirstate handle mismatched directory cases is too expensive[2]. Since dirstate doesn't apply to committed csets, this is only created by overriding basectx.match() in workingctx, and only on icasefs. The default arguments have been dropped, because the ctx must be passed to the matcher in order to function. For operations that can apply to both wdir and some other context, this ends up normalizing the filename to the case as it exists in the filesystem, and using that case for the lookup in the other context. See the diff example in the test. Previously, given a directory with an inexact case: - add worked as expected - diff, forget and status would silently ignore the request - files would exit with 1 - commit, revert and remove would fail (even when the commands leading up to them worked): $ hg ci -m "AbCDef" capsdir1/capsdir abort: CapsDir1/CapsDir: no match under directory! $ hg revert -r '.^' capsdir1/capsdir capsdir1\capsdir: no such file in rev 64dae27060b7 $ hg remove capsdir1/capsdir not removing capsdir1\capsdir: no tracked files [1] Globs are normalized, so that the -I and -X don't need to be specified with a case match. Without that, the second last remove (with -X) removes the files, leaving nothing for the last remove. However, specifying the files as 'glob:**.Txt' does not work. Perhaps this requires 're.IGNORECASE'? There are only a handful of places that create matchers directly, instead of being routed through the context.match() method. Some may benefit from changing over to using ctx.match() as a factory function: revset.checkstatus() revset.contains() revset.filelog() revset._matchfiles() localrepository._loadfilter() ignore.ignore() fileset.subrepo() filemerge._picktool() overrides.addlargefiles() lfcommands.lfconvert() kwtemplate.__init__() eolfile.__init__() eolfile.checkrev() acl.buildmatch() Currently, a toplevel subrepo can be named with an inexact case. However, the path auditor gets in the way of naming _anything_ in the subrepo if the top level case doesn't match. That is trickier to handle, because there's the user provided case, the case in the filesystem, and the case stored in .hgsub. This can be fixed next cycle. --- a/tests/test-subrepo-deep-nested-change.t +++ b/tests/test-subrepo-deep-nested-change.t @@ -170,8 +170,15 @@ R sub1/sub2/test.txt $ hg update -Cq $ touch sub1/sub2/folder/bar +#if icasefs + $ hg addremove Sub1/sub2 + abort: path 'Sub1\sub2' is inside nested repo 'Sub1' + [255] + $ hg -q addremove sub1/sub2 +#else $ hg addremove sub1/sub2 adding sub1/sub2/folder/bar (glob) +#endif $ hg status -S A sub1/sub2/folder/bar ? foo/bar/abc The narrowmatcher class may need to be tweaked when that is fixed. [1] http://www.selenic.com/pipermail/mercurial-devel/2015-April/068183.html [2] http://www.selenic.com/pipermail/mercurial-devel/2015-April/068191.html
2015-04-12 08:39:21 +03:00
def _filtersuspectsymlink(self, files):
if not files or self._repo.dirstate._checklink:
return files
# Symlink placeholders may get non-symlink-like contents
# via user error or dereferencing by NFS or Samba servers,
# so we filter out any placeholders that don't look like a
# symlink
sane = []
for f in files:
if self.flags(f) == 'l':
d = self[f].data()
if d == '' or len(d) >= 1024 or '\n' in d or util.binary(d):
self._repo.ui.debug('ignoring suspect symlink placeholder'
' "%s"\n' % f)
continue
sane.append(f)
return sane
def _checklookup(self, files):
# check for any possibly clean files
if not files:
return [], [], []
modified = []
deleted = []
fixup = []
pctx = self._parents[0]
# do a full compare of any files that might have changed
for f in sorted(files):
try:
# This will return True for a file that got replaced by a
# directory in the interim, but fixing that is pretty hard.
if (f not in pctx or self.flags(f) != pctx.flags(f)
or pctx[f].cmp(self[f])):
modified.append(f)
else:
fixup.append(f)
except (IOError, OSError):
# A file become inaccessible in between? Mark it as deleted,
# matching dirstate behavior (issue5584).
# The dirstate has more complex behavior around whether a
# missing file matches a directory, etc, but we don't need to
# bother with that: if f has made it to this point, we're sure
# it's in the dirstate.
deleted.append(f)
return modified, deleted, fixup
def _poststatusfixup(self, status, fixup):
"""update dirstate for files that are actually clean"""
poststatus = self._repo.postdsstatus()
if fixup or poststatus:
try:
oldid = self._repo.dirstate.identity()
# updating the dirstate is optional
# so we don't wait on the lock
# wlock can invalidate the dirstate, so cache normal _after_
# taking the lock
with self._repo.wlock(False):
if self._repo.dirstate.identity() == oldid:
if fixup:
normal = self._repo.dirstate.normal
for f in fixup:
normal(f)
# write changes out explicitly, because nesting
# wlock at runtime may prevent 'wlock.release()'
# after this block from doing so for subsequent
# changing files
tr = self._repo.currenttransaction()
self._repo.dirstate.write(tr)
if poststatus:
for ps in poststatus:
ps(self, status)
else:
# in this case, writing changes out breaks
# consistency, because .hg/dirstate was
# already changed simultaneously after last
# caching (see also issue5584 for detail)
self._repo.ui.debug('skip updating dirstate: '
'identity mismatch\n')
except error.LockError:
pass
finally:
# Even if the wlock couldn't be grabbed, clear out the list.
self._repo.clearpostdsstatus()
def _dirstatestatus(self, match, ignored=False, clean=False, unknown=False):
'''Gets the status from the dirstate -- internal use only.'''
listignored, listclean, listunknown = ignored, clean, unknown
subrepos = []
if '.hgsub' in self:
subrepos = sorted(self.substate)
cmp, s = self._repo.dirstate.status(match, subrepos, listignored,
listclean, listunknown)
# check for any possibly clean files
fixup = []
if cmp:
modified2, deleted2, fixup = self._checklookup(cmp)
s.modified.extend(modified2)
s.deleted.extend(deleted2)
if fixup and listclean:
s.clean.extend(fixup)
self._poststatusfixup(s, fixup)
if match.always():
# cache for performance
if s.unknown or s.ignored or s.clean:
# "_status" is cached with list*=False in the normal route
self._status = scmutil.status(s.modified, s.added, s.removed,
s.deleted, [], [], [])
else:
self._status = s
return s
@propertycache
def _manifest(self):
"""generate a manifest corresponding to the values in self._status
This reuse the file nodeid from parent, but we use special node
identifiers for added and modified files. This is used by manifests
merge to see that files are different and by update logic to avoid
deleting newly added files.
"""
return self._buildstatusmanifest(self._status)
def _buildstatusmanifest(self, status):
"""Builds a manifest that includes the given status results."""
parents = self.parents()
man = parents[0].manifest().copy()
ff = self._flagfunc
for i, l in ((addednodeid, status.added),
(modifiednodeid, status.modified)):
for f in l:
man[f] = i
try:
man.setflag(f, ff(f))
except OSError:
pass
for f in status.deleted + status.removed:
if f in man:
del man[f]
return man
def _buildstatus(self, other, s, match, listignored, listclean,
listunknown):
"""build a status with respect to another context
This includes logic for maintaining the fast path of status when
comparing the working directory against its parent, which is to skip
building a new manifest if self (working directory) is not comparing
against its parent (repo['.']).
"""
s = self._dirstatestatus(match, listignored, listclean, listunknown)
# Filter out symlinks that, in the case of FAT32 and NTFS filesystems,
# might have accidentally ended up with the entire contents of the file
# they are supposed to be linking to.
s.modified[:] = self._filtersuspectsymlink(s.modified)
if other != self._repo['.']:
s = super(workingctx, self)._buildstatus(other, s, match,
listignored, listclean,
listunknown)
return s
def _matchstatus(self, other, match):
"""override the match method with a filter for directory patterns
We use inheritance to customize the match.bad method only in cases of
workingctx since it belongs only to the working directory when
comparing against the parent changeset.
If we aren't comparing against the working directory's parent, then we
just use the default match object sent to us.
"""
if other != self._repo['.']:
def bad(f, msg):
# 'f' may be a directory pattern from 'match.files()',
# so 'f not in ctx1' is not enough
if f not in other and not other.hasdir(f):
self._repo.ui.warn('%s: %s\n' %
(self._repo.dirstate.pathto(f), msg))
match.bad = bad
return match
def markcommitted(self, node):
super(workingctx, self).markcommitted(node)
sparse.aftercommit(self._repo, node)
class committablefilectx(basefilectx):
"""A committablefilectx provides common functionality for a file context
that wants the ability to commit, e.g. workingfilectx or memfilectx."""
def __init__(self, repo, path, filelog=None, ctx=None):
self._repo = repo
self._path = path
self._changeid = None
self._filerev = self._filenode = None
if filelog is not None:
self._filelog = filelog
if ctx:
self._changectx = ctx
def __nonzero__(self):
return True
__bool__ = __nonzero__
def linkrev(self):
# linked to self._changectx no matter if file is modified or not
return self.rev()
def parents(self):
'''return parent filectxs, following copies if necessary'''
def filenode(ctx, path):
return ctx._manifest.get(path, nullid)
path = self._path
fl = self._filelog
pcl = self._changectx._parents
renamed = self.renamed()
if renamed:
pl = [renamed + (None,)]
else:
pl = [(path, filenode(pcl[0], path), fl)]
for pc in pcl[1:]:
pl.append((path, filenode(pc, path), fl))
return [self._parentfilectx(p, fileid=n, filelog=l)
2010-01-25 09:05:27 +03:00
for p, n, l in pl if n != nullid]
def children(self):
return []
class workingfilectx(committablefilectx):
"""A workingfilectx object makes access to data related to a particular
file in the working directory convenient."""
def __init__(self, repo, path, filelog=None, workingctx=None):
super(workingfilectx, self).__init__(repo, path, filelog, workingctx)
@propertycache
def _changectx(self):
return workingctx(self._repo)
def data(self):
return self._repo.wread(self._path)
def renamed(self):
rp = self._repo.dirstate.copied(self._path)
if not rp:
return None
return rp, self._changectx._parents[0]._manifest.get(rp, nullid)
2010-01-25 09:05:27 +03:00
def size(self):
return self._repo.wvfs.lstat(self._path).st_size
2006-12-25 00:55:34 +03:00
def date(self):
t, tz = self._changectx.date()
try:
return (self._repo.wvfs.lstat(self._path).st_mtime, tz)
except OSError as err:
2010-01-25 09:05:27 +03:00
if err.errno != errno.ENOENT:
raise
2006-12-25 00:55:34 +03:00
return (t, tz)
2006-10-10 10:13:03 +04:00
def exists(self):
return self._repo.wvfs.exists(self._path)
def lexists(self):
return self._repo.wvfs.lexists(self._path)
def audit(self):
return self._repo.wvfs.audit(self._path)
def cmp(self, fctx):
"""compare with other file context
returns True if different than fctx.
"""
# fctx should be a filectx (not a workingfilectx)
# invert comparison to reuse the same code path
return fctx.cmp(self)
2008-06-19 02:14:23 +04:00
def remove(self, ignoremissing=False):
"""wraps unlink for a repo's working directory"""
2015-01-14 03:15:26 +03:00
self._repo.wvfs.unlinkpath(self._path, ignoremissing=ignoremissing)
def write(self, data, flags, backgroundclose=False):
"""wraps repo.wwrite"""
self._repo.wwrite(self._path, data, flags,
backgroundclose=backgroundclose)
def setflags(self, l, x):
self._repo.wvfs.setflags(self._path, l, x)
class workingcommitctx(workingctx):
"""A workingcommitctx object makes access to data related to
the revision being committed convenient.
This hides changes in the working directory, if they aren't
committed in this context.
"""
def __init__(self, repo, changes,
text="", user=None, date=None, extra=None):
super(workingctx, self).__init__(repo, text, user, date, extra,
changes)
def _dirstatestatus(self, match, ignored=False, clean=False, unknown=False):
"""Return matched files only in ``self._status``
Uncommitted files appear "clean" via this context, even if
they aren't actually so in the working directory.
"""
if clean:
clean = [f for f in self._manifest if f not in self._changedset]
else:
clean = []
return scmutil.status([f for f in self._status.modified if match(f)],
[f for f in self._status.added if match(f)],
[f for f in self._status.removed if match(f)],
[], [], [], clean)
@propertycache
def _changedset(self):
"""Return the set of files changed in this context
"""
changed = set(self._status.modified)
changed.update(self._status.added)
changed.update(self._status.removed)
return changed
def makecachingfilectxfn(func):
"""Create a filectxfn that caches based on the path.
We can't use util.cachefunc because it uses all arguments as the cache
key and this creates a cycle since the arguments include the repo and
memctx.
"""
cache = {}
def getfilectx(repo, memctx, path):
if path not in cache:
cache[path] = func(repo, memctx, path)
return cache[path]
return getfilectx
def memfilefromctx(ctx):
"""Given a context return a memfilectx for ctx[path]
This is a convenience method for building a memctx based on another
context.
"""
def getfilectx(repo, memctx, path):
fctx = ctx[path]
# this is weird but apparently we only keep track of one parent
# (why not only store that instead of a tuple?)
copied = fctx.renamed()
if copied:
copied = copied[0]
return memfilectx(repo, path, fctx.data(),
islink=fctx.islink(), isexec=fctx.isexec(),
copied=copied, memctx=memctx)
return getfilectx
def memfilefrompatch(patchstore):
"""Given a patch (e.g. patchstore object) return a memfilectx
This is a convenience method for building a memctx based on a patchstore.
"""
def getfilectx(repo, memctx, path):
data, mode, copied = patchstore.getfile(path)
if data is None:
return None
islink, isexec = mode
return memfilectx(repo, path, data, islink=islink,
isexec=isexec, copied=copied,
memctx=memctx)
return getfilectx
class memctx(committablectx):
2008-10-11 15:07:29 +04:00
"""Use memctx to perform in-memory commits via localrepo.commitctx().
Revision information is supplied at initialization time while
related files data and is made available through a callback
mechanism. 'repo' is the current localrepo, 'parents' is a
sequence of two parent revisions identifiers (pass None for every
missing parent), 'text' is the commit message and 'files' lists
names of files touched by the revision (normalized and relative to
repository root).
filectxfn(repo, memctx, path) is a callable receiving the
repository, the current memctx object and the normalized path of
requested file, relative to repository root. It is fired by the
commit function for every file in 'files', but calls order is
undefined. If the file is available in the revision being
committed (updated or added), filectxfn returns a memfilectx
object. If the file was removed, filectxfn return None for recent
Mercurial. Moved files are represented by marking the source file
2008-10-11 15:07:29 +04:00
removed and the new file added with copy information (see
memfilectx).
user receives the committer name and defaults to current
repository username, date is the commit date in any format
supported by util.parsedate() and defaults to current date, extra
is a dictionary of metadata or is left empty.
2008-06-19 02:14:23 +04:00
"""
# Mercurial <= 3.1 expects the filectxfn to raise IOError for missing files.
# Extensions that need to retain compatibility across Mercurial 3.1 can use
# this field to determine what to do in filectxfn.
_returnnoneformissingfiles = True
2008-06-23 15:12:32 +04:00
def __init__(self, repo, parents, text, files, filectxfn, user=None,
date=None, extra=None, branch=None, editor=False):
super(memctx, self).__init__(repo, text, user, date, extra)
2008-06-19 02:14:23 +04:00
self._rev = None
self._node = None
parents = [(p or nullid) for p in parents]
p1, p2 = parents
2008-06-26 23:35:46 +04:00
self._parents = [changectx(self._repo, p) for p in (p1, p2)]
files = sorted(set(files))
self._files = files
if branch is not None:
self._extra['branch'] = encoding.fromlocal(branch)
self.substate = {}
2008-06-19 02:14:23 +04:00
if isinstance(filectxfn, patch.filestore):
filectxfn = memfilefrompatch(filectxfn)
elif not callable(filectxfn):
# if store is not callable, wrap it in a function
filectxfn = memfilefromctx(filectxfn)
# memoizing increases performance for e.g. vcs convert scenarios.
self._filectxfn = makecachingfilectxfn(filectxfn)
if editor:
self._text = editor(self._repo, self, [])
self._repo.savecommitmessage(self._text)
2008-06-19 02:14:23 +04:00
def filectx(self, path, filelog=None):
"""get a file context from the working directory
Returns None if file doesn't exist and should be removed."""
2008-06-19 02:14:23 +04:00
return self._filectxfn(self._repo, self, path)
2010-05-01 16:00:21 +04:00
def commit(self):
"""commit context to the repo"""
return self._repo.commitctx(self)
@propertycache
def _manifest(self):
"""generate a manifest based on the return values of filectxfn"""
# keep this simple for now; just worry about p1
pctx = self._parents[0]
man = pctx.manifest().copy()
for f in self._status.modified:
p1node = nullid
p2node = nullid
2014-06-18 07:55:06 +04:00
p = pctx[f].parents() # if file isn't in pctx, check p2?
if len(p) > 0:
p1node = p[0].filenode()
if len(p) > 1:
p2node = p[1].filenode()
man[f] = revlog.hash(self[f].data(), p1node, p2node)
for f in self._status.added:
man[f] = revlog.hash(self[f].data(), nullid, nullid)
for f in self._status.removed:
if f in man:
del man[f]
return man
@propertycache
def _status(self):
"""Calculate exact status from ``files`` specified at construction
"""
man1 = self.p1().manifest()
p2 = self._parents[1]
# "1 < len(self._parents)" can't be used for checking
# existence of the 2nd parent, because "memctx._parents" is
# explicitly initialized by the list, of which length is 2.
if p2.node() != nullid:
man2 = p2.manifest()
managing = lambda f: f in man1 or f in man2
else:
managing = lambda f: f in man1
modified, added, removed = [], [], []
for f in self._files:
if not managing(f):
added.append(f)
elif self[f]:
modified.append(f)
else:
removed.append(f)
return scmutil.status(modified, added, removed, [], [], [], [])
class memfilectx(committablefilectx):
2008-10-11 15:07:29 +04:00
"""memfilectx represents an in-memory file to commit.
See memctx and committablefilectx for more details.
2008-06-19 02:14:23 +04:00
"""
def __init__(self, repo, path, data, islink=False,
isexec=False, copied=None, memctx=None):
2008-10-11 15:07:29 +04:00
"""
path is the normalized file path relative to repository root.
data is the file content as a string.
islink is True if the file is a symbolic link.
isexec is True if the file is executable.
copied is the source file path if current file was copied in the
revision being committed, or None."""
super(memfilectx, self).__init__(repo, path, None, memctx)
2008-06-19 02:14:23 +04:00
self._data = data
self._flags = (islink and 'l' or '') + (isexec and 'x' or '')
self._copied = None
if copied:
self._copied = (copied, nullid)
2010-01-25 09:05:27 +03:00
def data(self):
return self._data
def remove(self, ignoremissing=False):
"""wraps unlink for a repo's working directory"""
# need to figure out what to do here
del self._changectx[self._path]
def write(self, data, flags):
"""wraps repo.wwrite"""
self._data = data
class overlayfilectx(committablefilectx):
"""Like memfilectx but take an original filectx and optional parameters to
override parts of it. This is useful when fctx.data() is expensive (i.e.
flag processor is expensive) and raw data, flags, and filenode could be
reused (ex. rebase or mode-only amend a REVIDX_EXTSTORED file).
"""
def __init__(self, originalfctx, datafunc=None, path=None, flags=None,
copied=None, ctx=None):
"""originalfctx: filecontext to duplicate
datafunc: None or a function to override data (file content). It is a
function to be lazy. path, flags, copied, ctx: None or overridden value
copied could be (path, rev), or False. copied could also be just path,
and will be converted to (path, nullid). This simplifies some callers.
"""
if path is None:
path = originalfctx.path()
if ctx is None:
ctx = originalfctx.changectx()
ctxmatch = lambda: True
else:
ctxmatch = lambda: ctx == originalfctx.changectx()
repo = originalfctx.repo()
flog = originalfctx.filelog()
super(overlayfilectx, self).__init__(repo, path, flog, ctx)
if copied is None:
copied = originalfctx.renamed()
copiedmatch = lambda: True
else:
if copied and not isinstance(copied, tuple):
# repo._filecommit will recalculate copyrev so nullid is okay
copied = (copied, nullid)
copiedmatch = lambda: copied == originalfctx.renamed()
# When data, copied (could affect data), ctx (could affect filelog
# parents) are not overridden, rawdata, rawflags, and filenode may be
# reused (repo._filecommit should double check filelog parents).
#
# path, flags are not hashed in filelog (but in manifestlog) so they do
# not affect reusable here.
#
# If ctx or copied is overridden to a same value with originalfctx,
# still consider it's reusable. originalfctx.renamed() may be a bit
# expensive so it's not called unless necessary. Assuming datafunc is
# always expensive, do not call it for this "reusable" test.
reusable = datafunc is None and ctxmatch() and copiedmatch()
if datafunc is None:
datafunc = originalfctx.data
if flags is None:
flags = originalfctx.flags()
self._datafunc = datafunc
self._flags = flags
self._copied = copied
if reusable:
# copy extra fields from originalfctx
attrs = ['rawdata', 'rawflags', '_filenode', '_filerev']
for attr in attrs:
if util.safehasattr(originalfctx, attr):
setattr(self, attr, getattr(originalfctx, attr))
def data(self):
return self._datafunc()
class metadataonlyctx(committablectx):
"""Like memctx but it's reusing the manifest of different commit.
Intended to be used by lightweight operations that are creating
metadata-only changes.
Revision information is supplied at initialization time. 'repo' is the
current localrepo, 'ctx' is original revision which manifest we're reuisng
'parents' is a sequence of two parent revisions identifiers (pass None for
every missing parent), 'text' is the commit.
user receives the committer name and defaults to current repository
username, date is the commit date in any format supported by
util.parsedate() and defaults to current date, extra is a dictionary of
metadata or is left empty.
"""
def __new__(cls, repo, originalctx, *args, **kwargs):
return super(metadataonlyctx, cls).__new__(cls, repo)
def __init__(self, repo, originalctx, parents, text, user=None, date=None,
extra=None, editor=False):
super(metadataonlyctx, self).__init__(repo, text, user, date, extra)
self._rev = None
self._node = None
self._originalctx = originalctx
self._manifestnode = originalctx.manifestnode()
parents = [(p or nullid) for p in parents]
p1, p2 = self._parents = [changectx(self._repo, p) for p in parents]
# sanity check to ensure that the reused manifest parents are
# manifests of our commit parents
mp1, mp2 = self.manifestctx().parents
if p1 != nullid and p1.manifestnode() != mp1:
raise RuntimeError('can\'t reuse the manifest: '
'its p1 doesn\'t match the new ctx p1')
if p2 != nullid and p2.manifestnode() != mp2:
raise RuntimeError('can\'t reuse the manifest: '
'its p2 doesn\'t match the new ctx p2')
self._files = originalctx.files()
self.substate = {}
if editor:
self._text = editor(self._repo, self, [])
self._repo.savecommitmessage(self._text)
def manifestnode(self):
return self._manifestnode
@property
def _manifestctx(self):
return self._repo.manifestlog[self._manifestnode]
def filectx(self, path, filelog=None):
return self._originalctx.filectx(path, filelog=filelog)
def commit(self):
"""commit context to the repo"""
return self._repo.commitctx(self)
@property
def _manifest(self):
return self._originalctx.manifest()
@propertycache
def _status(self):
"""Calculate exact status from ``files`` specified in the ``origctx``
and parents manifests.
"""
man1 = self.p1().manifest()
p2 = self._parents[1]
# "1 < len(self._parents)" can't be used for checking
# existence of the 2nd parent, because "metadataonlyctx._parents" is
# explicitly initialized by the list, of which length is 2.
if p2.node() != nullid:
man2 = p2.manifest()
managing = lambda f: f in man1 or f in man2
else:
managing = lambda f: f in man1
modified, added, removed = [], [], []
for f in self._files:
if not managing(f):
added.append(f)
elif self[f]:
modified.append(f)
else:
removed.append(f)
return scmutil.status(modified, added, removed, [], [], [], [])