sapling/eden/scm/edenscm/mercurial/hg.py
Jun Wu 2d6e4db55e clone: stop using hardlinks for svfs
Summary:
Hard link adds complexity for revlog writes. It's not that useful in production
setup. The Rust revlog `flush` API does not break hardlinked files. So let's
just avoid using hard links during local repo clone.

Reviewed By: DurhamG

Differential Revision: D22638460

fbshipit-source-id: 038f4d5c48e9972b14c9e59a9d7ef72b6bc5308d
2020-07-30 20:32:31 -07:00

1119 lines
36 KiB
Python

# Portions Copyright (c) Facebook, Inc. and its affiliates.
#
# This software may be used and distributed according to the terms of the
# GNU General Public License version 2.
# hg.py - repository classes for mercurial
#
# Copyright 2005-2007 Matt Mackall <mpm@selenic.com>
# Copyright 2006 Vadim Gelfer <vadim.gelfer@gmail.com>
#
# This software may be used and distributed according to the terms of the
# GNU General Public License version 2 or any later version.
from __future__ import absolute_import
import errno
import hashlib
import os
import shutil
import weakref
from . import (
bookmarks,
bundlerepo,
clone as clonemod,
cmdutil,
destutil,
discovery,
edenfs,
error,
exchange,
extensions,
httppeer,
localrepo,
lock,
merge as mergemod,
node,
perftrace,
phases,
progress,
pycompat,
scmutil,
sshpeer,
ui as uimod,
url,
util,
verify as verifymod,
vfs as vfsmod,
visibility,
)
from .i18n import _
from .node import nullid
from .pycompat import encodeutf8
release = lock.release
# shared features
sharedbookmarks = "bookmarks"
def _local(path):
path = util.expandpath(util.urllocalpath(path))
return os.path.isfile(path) and bundlerepo or localrepo
def addbranchrevs(lrepo, other, branches, revs):
peer = other.peer() # a courtesy to callers using a localrepo for other
hashbranch, branches = branches
if not hashbranch and not branches:
x = revs or None
if util.safehasattr(revs, "first"):
y = revs.first()
elif revs:
y = revs[0]
else:
y = None
return x, y
if revs:
revs = list(revs)
else:
revs = []
if not peer.capable("branchmap"):
if branches:
raise error.Abort(_("remote branch lookup not supported"))
revs.append(hashbranch)
return revs, revs[0]
branchmap = peer.branchmap()
def primary(branch):
if branch == ".":
if not lrepo:
raise error.Abort(_("dirstate branch not accessible"))
branch = lrepo.dirstate.branch()
if branch in branchmap:
revs.extend(node.hex(r) for r in reversed(branchmap[branch]))
return True
else:
return False
for branch in branches:
if not primary(branch):
raise error.RepoLookupError(_("unknown branch '%s'") % branch)
if hashbranch:
if not primary(hashbranch):
revs.append(hashbranch)
return revs, revs[0]
def parseurl(path, branches=None):
"""parse url#branch, returning (url, (branch, branches))"""
u = util.url(path)
branch = None
if u.fragment:
branch = u.fragment
u.fragment = None
return str(u), (branch, branches or [])
schemes = {
"bundle": bundlerepo,
"file": _local,
"http": httppeer,
"https": httppeer,
"ssh": sshpeer,
}
def _peerlookup(path):
u = util.url(path)
scheme = u.scheme or "file"
thing = schemes.get(scheme) or schemes["file"]
try:
return thing(path)
except TypeError:
# we can't test callable(thing) because 'thing' can be an unloaded
# module that implements __call__
if not util.safehasattr(thing, "instance"):
raise
return thing
def islocal(repo):
"""return true if repo (or path pointing to repo) is local"""
if isinstance(repo, str):
try:
return _peerlookup(repo).islocal(repo)
except AttributeError:
return False
return repo.local()
def openpath(ui, path):
"""open path with open if local, url.open if remote"""
pathurl = util.url(path, parsequery=False, parsefragment=False)
if pathurl.islocal():
return util.posixfile(pathurl.localpath(), "rb")
else:
return url.open(ui, path)
# a list of (ui, repo) functions called for wire peer initialization
wirepeersetupfuncs = []
@perftrace.tracefunc("Repo Setup")
def _peerorrepo(ui, path, create=False, presetupfuncs=None):
"""return a repository object for the specified path"""
obj = _peerlookup(path).instance(ui, path, create)
ui = getattr(obj, "ui", ui)
for f in presetupfuncs or []:
f(ui, obj)
for name, module in extensions.extensions(ui):
hook = getattr(module, "reposetup", None)
if hook:
hook(ui, obj)
if not obj.local():
perftrace.traceflag("remote")
for f in wirepeersetupfuncs:
f(ui, obj)
else:
perftrace.traceflag("local")
return obj
def repository(ui, path="", create=False, presetupfuncs=None):
"""return a repository object for the specified path"""
peer = _peerorrepo(ui, path, create, presetupfuncs=presetupfuncs)
repo = peer.local()
if not repo:
raise error.Abort(_("repository '%s' is not local") % (path or peer.url()))
return repo
def peer(uiorrepo, opts, path, create=False):
"""return a repository peer for the specified path"""
rui = remoteui(uiorrepo, opts)
return _peerorrepo(rui, path, create).peer()
def defaultdest(source):
"""return default destination of clone if none is given
>>> defaultdest(b'foo')
'foo'
>>> defaultdest(b'/foo/bar')
'bar'
>>> defaultdest(b'/')
''
>>> defaultdest(b'')
''
>>> defaultdest(b'http://example.org/')
''
>>> defaultdest(b'http://example.org/foo/')
'foo'
"""
path = util.url(source).path
if not path:
return ""
return os.path.basename(os.path.normpath(path))
def share(
ui, source, dest=None, update=True, bookmarks=True, defaultpath=None, relative=False
):
"""create a shared repository"""
if not islocal(source):
raise error.Abort(_("can only share local repositories"))
if not dest:
dest = defaultdest(source)
else:
dest = ui.expandpath(dest)
if isinstance(source, str):
origsource = ui.expandpath(source)
source, branches = parseurl(origsource)
srcrepo = repository(ui, source)
rev, checkout = addbranchrevs(srcrepo, srcrepo, branches, None)
else:
srcrepo = source.local()
origsource = source = srcrepo.url()
checkout = None
sharedpath = srcrepo.sharedpath # if our source is already sharing
destwvfs = vfsmod.vfs(dest, realpath=True)
destvfs = vfsmod.vfs(os.path.join(destwvfs.base, ".hg"), realpath=True)
if destvfs.lexists():
raise error.Abort(_("destination already exists"))
if not destwvfs.isdir():
destwvfs.mkdir()
destvfs.makedir()
requirements = srcrepo.requirements.copy()
if relative:
try:
sharedpath = os.path.relpath(sharedpath, destvfs.base)
requirements.add("relshared")
except (IOError, ValueError) as e:
# ValueError is raised on Windows if the drive letters differ on
# each path
raise error.Abort(_("cannot calculate relative path"), hint=str(e))
else:
requirements.add("shared")
scmutil.writerequires(destvfs, requirements)
destvfs.writeutf8("sharedpath", sharedpath)
r = repository(ui, destwvfs.base)
postshare(srcrepo, r, bookmarks=bookmarks, defaultpath=defaultpath)
_postshareupdate(r, update, checkout=checkout)
return r
def unshare(ui, repo):
"""convert a shared repository to a normal one
Copy the store data to the repo and remove the sharedpath data.
"""
destlock = lock = None
lock = repo.lock()
try:
# we use locks here because if we race with commit, we
# can end up with extra data in the cloned revlogs that's
# not pointed to by changesets, thus causing verify to
# fail
destlock = copystore(ui, repo, repo.path)
sharefile = repo.localvfs.join("sharedpath")
util.rename(sharefile, sharefile + ".old")
repo.requirements.discard("shared")
repo.requirements.discard("relshared")
repo._writerequirements()
finally:
destlock and destlock.release()
lock and lock.release()
# update store, spath, svfs and sjoin of repo
# invalidate before rerunning __init__
repo.invalidate(clearfilecache=True)
repo.invalidatedirstate()
repo.__init__(repo.baseui, repo.root)
# reinitialize zstore
if repo.ui.configbool("format", "use-zstore-commit-data"):
repo._syncrevlogtozstore()
def postshare(sourcerepo, destrepo, bookmarks=True, defaultpath=None):
"""Called after a new shared repo is created.
The new repo only has a requirements file and pointer to the source.
This function configures additional shared data.
Extensions can wrap this function and write additional entries to
destrepo/.hg/shared to indicate additional pieces of data to be shared.
"""
default = defaultpath or sourcerepo.ui.config("paths", "default")
if default:
fp = destrepo.localvfs("hgrc", "w", text=True)
fp.write("[paths]\n")
fp.write("default = %s\n" % default)
fp.close()
with destrepo.wlock():
if bookmarks:
fp = destrepo.localvfs("shared", "wb")
fp.write(pycompat.encodeutf8(sharedbookmarks + "\n"))
fp.close()
def _postshareupdate(repo, update, checkout=None):
"""Maybe perform a working directory update after a shared repo is created.
``update`` can be a boolean or a revision to update to.
"""
if not update:
return
repo.ui.status(_("updating working directory\n"))
if update is not True:
checkout = update
for test in (checkout, "default", "tip"):
if test is None:
continue
try:
uprev = repo.lookup(test)
break
except error.RepoLookupError:
continue
_update(repo, uprev)
def copystore(ui, srcrepo, destpath):
"""copy files from store of srcrepo in destpath
returns destlock
"""
destlock = None
try:
with progress.bar(ui, _("linking")) as prog:
hardlink = False
num = 0
srcpublishing = srcrepo.publishing()
srcvfs = vfsmod.vfs(srcrepo.sharedpath)
dstvfs = vfsmod.vfs(destpath)
for f in srcrepo.store.copylist():
if srcpublishing and f.endswith("phaseroots"):
continue
dstbase = os.path.dirname(f)
if dstbase and not dstvfs.exists(dstbase):
dstvfs.mkdir(dstbase)
if srcvfs.exists(f):
if f.endswith("data"):
# 'dstbase' may be empty (e.g. revlog format 0)
lockfile = os.path.join(dstbase, "lock")
# lock to avoid premature writing to the target
destlock = lock.lock(dstvfs, lockfile)
hardlink, num = util.copyfiles(
srcvfs.join(f), dstvfs.join(f), hardlink, num, prog
)
if hardlink:
ui.debug("linked %d files\n" % num)
else:
ui.debug("copied %d files\n" % num)
return destlock
except: # re-raises
release(destlock)
raise
def clone(
ui,
peeropts,
source,
dest=None,
pull=False,
rev=None,
update=True,
stream=False,
shallow=False,
):
"""Make a copy of an existing repository.
Create a copy of an existing repository in a new directory. The
source and destination are URLs, as passed to the repository
function. Returns a pair of repository peers, the source and
newly created destination.
The location of the source is added to the new repository's
.hg/hgrc file, as the default to be used for future pulls and
pushes.
If an exception is raised, the partly cloned/updated destination
repository will be deleted.
Arguments:
source: repository object or URL
dest: URL of destination repository to create (defaults to base
name of source repository)
pull: always pull from source repository, even in local case or if the
server prefers streaming
stream: stream raw data uncompressed from repository (fast over
LAN, slow over WAN)
rev: revision to clone up to (implies pull=True)
update: update working directory after clone completes, if
destination is local repository (True means update to default rev,
anything else is treated as a revision)
"""
if isinstance(source, str):
origsource = ui.expandpath(source)
source, mayberevs = parseurl(origsource)
if len(mayberevs) == 1:
rev = rev or mayberevs[0]
srcpeer = peer(ui, peeropts, source)
else:
srcpeer = source.peer() # in case we were called with a localrepo
origsource = source = srcpeer.url()
branch = (None, [])
rev, checkout = addbranchrevs(srcpeer, srcpeer, branch, rev)
if dest is None:
dest = defaultdest(source)
if dest:
ui.status(_("destination directory: %s\n") % dest)
else:
dest = ui.expandpath(dest)
destpeer = None
dest = util.urllocalpath(dest)
source = util.urllocalpath(source)
if not dest:
raise error.Abort(_("empty destination path is not valid"))
destvfs = vfsmod.vfs(dest, expandpath=True)
if destvfs.lexists():
if not destvfs.isdir():
raise error.Abort(_("destination '%s' already exists") % dest)
elif destvfs.listdir():
raise error.Abort(_("destination '%s' is not empty") % dest)
srclock = destlock = destlockw = cleandir = None
srcrepo = srcpeer.local()
try:
abspath = origsource
if islocal(origsource):
abspath = os.path.abspath(util.urllocalpath(origsource))
if islocal(dest):
cleandir = dest
copy = False
if (
srcrepo
and srcrepo.cancopy()
and islocal(dest)
and not phases.hassecret(srcrepo)
):
copy = not pull and not rev
if copy:
try:
# we use a lock here because if we race with commit, we
# can end up with extra data in the cloned revlogs that's
# not pointed to by changesets, thus causing verify to
# fail
srclock = srcrepo.lock(wait=False)
except error.LockError:
copy = False
if copy:
clonecodepath = "copy"
srcrepo.hook("preoutgoing", throw=True, source="clone")
hgdir = os.path.realpath(os.path.join(dest, ".hg"))
if not os.path.exists(dest):
os.mkdir(dest)
else:
# only clean up directories we create ourselves
cleandir = hgdir
try:
destpath = hgdir
util.makedir(destpath, notindexed=True)
except OSError as inst:
if inst.errno == errno.EEXIST:
cleandir = None
raise error.Abort(_("destination '%s' already exists") % dest)
raise
destlock = copystore(ui, srcrepo, destpath)
# repo initialization might also take a lock. Keeping destlock
# outside the repo object can cause deadlock. To avoid deadlock,
# we just release destlock here. The lock will be re-acquired
# soon by `destpeer`, or `local.lock()` below.
if destlock is not None:
destlock.release()
# copy bookmarks over
srcbookmarks = srcrepo.svfs.join("bookmarks")
dstbookmarks = os.path.join(destpath, "store", "bookmarks")
if os.path.exists(srcbookmarks):
util.copyfile(srcbookmarks, dstbookmarks)
# we need to re-init the repo after manually copying the data
# into it
destpeer = peer(srcrepo, peeropts, dest)
local = destpeer.local()
if local:
_writehgrc(local, abspath)
srcrepo.hook("outgoing", source="clone", node=node.hex(node.nullid))
else:
clonecodepath = "legacy-pull"
try:
destpeer = peer(srcrepo or ui, peeropts, dest, create=True)
# only pass ui when no srcrepo
except OSError as inst:
if inst.errno == errno.EEXIST:
cleandir = None
raise error.Abort(_("destination '%s' already exists") % dest)
raise
local = destpeer.local()
# Write [paths]. Code path below might use it.
if local:
_writehgrc(local, abspath)
revs = None
if rev:
if not srcpeer.capable("lookup"):
raise error.Abort(
_(
"src repository does not support "
"revision lookup and so doesn't "
"support clone by revision"
)
)
revs = [srcpeer.lookup(r) for r in rev]
checkout = revs[0]
# Can we use the new code path (stream clone + shallow + no
# update + selective pull)?
if (
local
and not pull
and not update
and not rev
and shallow
and stream is not False
and ui.configbool("experimental", "new-clone-path")
and ui.configbool("remotenames", "selectivepull")
):
clonecodepath = "modern"
clonemod.shallowclone(srcpeer.url(), local)
elif local:
if ui.configbool("experimental", "new-clone-path"):
ui.log(
"features",
fullargs=repr(pycompat.sysargv),
feature="legacy-clone",
traceback=util.smarttraceback(),
)
with local.wlock(), local.lock(), local.transaction("clone"):
if not stream:
if pull:
stream = False
else:
stream = None
overrides = {
# internal config: ui.quietbookmarkmove
("ui", "quietbookmarkmove"): True,
# the normal pull process each commit and so is more expensive
# than streaming bytes from disk to the wire.
# disabling selectivepull allows to run a streamclone
("remotenames", "selectivepull"): False,
}
opargs = {}
if shallow:
opargs["extras"] = {"shallow": True}
with local.ui.configoverride(overrides, "clone"):
exchange.pull(
local,
srcpeer,
revs,
streamclonerequested=stream,
opargs=opargs,
)
elif srcrepo:
exchange.push(
srcrepo, destpeer, revs=revs, bookmarks=srcrepo._bookmarks.keys()
)
else:
raise error.Abort(_("clone from remote to remote not supported"))
cleandir = None
destrepo = destpeer.local()
if destrepo:
with destrepo.wlock(), destrepo.lock(), destrepo.transaction("clone"):
if destrepo.ui.configbool("format", "use-zstore-commit-data"):
destrepo._syncrevlogtozstore()
if update:
if update is not True:
checkout = srcpeer.lookup(update)
uprev = None
status = None
if checkout is not None:
try:
uprev = destrepo.lookup(checkout)
except error.RepoLookupError:
if update is not True:
try:
uprev = destrepo.lookup(update)
except error.RepoLookupError:
pass
if uprev is None:
try:
uprev = destrepo._bookmarks["@"]
update = "@"
bn = destrepo[uprev].branch()
if bn == "default":
status = _("updating to bookmark @\n")
else:
status = _("updating to bookmark @ on branch %s\n") % bn
except KeyError:
try:
uprev = destrepo.branchtip("default")
except error.RepoLookupError:
uprev = destrepo.lookup("tip")
if not status:
bn = destrepo[uprev].branch()
status = _("updating to branch %s\n") % bn
destrepo.ui.status(status)
_update(destrepo, uprev)
if update in destrepo._bookmarks:
bookmarks.activate(destrepo, update)
clonepreclose(
ui,
peeropts,
source,
dest,
pull,
rev,
update,
stream,
srcpeer,
destpeer,
clonecodepath=clonecodepath,
)
finally:
release(srclock, destlockw, destlock)
if srcpeer is not None:
srcpeer.close()
if destpeer is not None:
destpeer.close()
if cleandir is not None:
shutil.rmtree(cleandir, True)
return srcpeer, destpeer
def _writehgrc(repo, abspath):
with repo.wlock(), repo.lock():
template = uimod.samplehgrcs["cloned"]
with repo.localvfs("hgrc", "wb") as fp:
u = util.url(abspath)
u.passwd = None
defaulturl = str(u)
fp.write(pycompat.encodeutf8(util.tonativeeol(template % defaulturl)))
repo.ui.setconfig("paths", "default", defaulturl, "clone")
def clonepreclose(
ui,
peeropts,
source,
dest=None,
pull=False,
rev=None,
update=True,
stream=False,
srcpeer=None,
destpeer=None,
clonecodepath=None,
):
"""Wrapped by extensions like remotenames before closing the peers
clonecodepath is one of:
- "copy": The clone was done by copying local files.
- "legacy-pull": The clone was done by the (legacy) pull code path.
- "modern": The clone was done by the modern clone.streamclone code path,
which is less racy and writes remote bookmarks.
"""
return srcpeer, destpeer
def _showstats(repo, stats, quietempty=False):
if edenfs.requirement in repo.requirements:
return _eden_showstats(repo, stats, quietempty)
if quietempty and not any(stats):
return
repo.ui.status(
_(
"%d files updated, %d files merged, "
"%d files removed, %d files unresolved\n"
)
% stats
)
def _eden_showstats(repo, stats, quietempty=False):
# We hide the updated and removed counts, because they are not accurate
# with eden. One of the primary goals of eden is that the entire working
# directory does not need to be accessed or traversed on update operations.
(updated, merged, removed, unresolved) = stats
if merged or unresolved:
repo.ui.status(
_("%d files merged, %d files unresolved\n") % (merged, unresolved)
)
elif not quietempty:
repo.ui.status(_("update complete\n"))
def updaterepo(repo, node, overwrite, updatecheck=None):
"""Update the working directory to node.
When overwrite is set, changes are clobbered, merged else
returns stats (see pydoc mercurial.merge.applyupdates)"""
return mergemod.update(
repo,
node,
False,
overwrite,
labels=["working copy", "destination"],
updatecheck=updatecheck,
)
def update(repo, node, quietempty=False, updatecheck=None):
"""update the working directory to node
Returns if any files were unresolved.
"""
stats = updaterepo(repo, node, False, updatecheck=updatecheck)
_showstats(repo, stats, quietempty)
if stats[3]:
repo.ui.status(_("use 'hg resolve' to retry unresolved file merges\n"))
return stats[3] > 0
# naming conflict in clone()
_update = update
def clean(repo, node, show_stats=True, quietempty=False):
"""forcibly switch the working directory to node, clobbering changes
Returns if any files were unresolved.
"""
stats = updaterepo(repo, node, True)
repo.localvfs.unlinkpath("graftstate", ignoremissing=True)
if show_stats:
_showstats(repo, stats, quietempty)
return stats[3] > 0
# naming conflict in updatetotally()
_clean = clean
def updatetotally(ui, repo, checkout, brev, clean=False, updatecheck=None):
"""Update the working directory with extra care for non-file components
This takes care of non-file components below:
:bookmark: might be advanced or (in)activated
This takes arguments below:
:checkout: to which revision the working directory is updated
:brev: a name, which might be a bookmark to be activated after updating
:clean: whether changes in the working directory can be discarded
:updatecheck: how to deal with a dirty working directory
Valid values for updatecheck are (None => linear):
* abort: abort if the working directory is dirty
* none: don't check (merge working directory changes into destination)
* linear: check that update is linear before merging working directory
changes into destination
* noconflict: check that the update does not result in file merges
This returns whether conflict is detected at updating or not.
"""
if updatecheck is None:
updatecheck = ui.config("commands", "update.check")
if updatecheck not in ("abort", "none", "linear", "noconflict"):
# If not configured, or invalid value configured
updatecheck = "linear"
with repo.wlock():
movemarkfrom = None
warndest = False
if checkout is None:
updata = destutil.destupdate(repo, clean=clean)
checkout, movemarkfrom, brev = updata
warndest = True
if clean:
hasunresolved = _clean(repo, checkout)
else:
if updatecheck == "abort":
cmdutil.bailifchanged(repo, merge=False)
updatecheck = "none"
hasunresolved = _update(repo, checkout, updatecheck=updatecheck)
if not hasunresolved and movemarkfrom:
if movemarkfrom == repo["."].node():
pass # no-op update
elif bookmarks.update(repo, [movemarkfrom], repo["."].node()):
b = ui.label(repo._activebookmark, "bookmarks.active")
ui.status(_("updating bookmark %s\n") % b)
else:
# this can happen with a non-linear update
b = ui.label(repo._activebookmark, "bookmarks")
ui.status(_("(leaving bookmark %s)\n") % b)
bookmarks.deactivate(repo)
elif brev in repo._bookmarks:
if brev != repo._activebookmark:
b = ui.label(brev, "bookmarks.active")
ui.status(_("(activating bookmark %s)\n") % b)
bookmarks.activate(repo, brev)
else:
if repo._activebookmark:
b = ui.label(repo._activebookmark, "bookmarks")
ui.status(_("(leaving bookmark %s)\n") % b)
bookmarks.deactivate(repo)
if warndest:
destutil.statusotherdests(ui, repo)
return hasunresolved
def merge(repo, node, force=None, remind=True, mergeforce=False, labels=None):
"""Branch merge with node, resolving changes. Return true if any
unresolved conflicts."""
stats = mergemod.update(
repo, node, True, force, mergeforce=mergeforce, labels=labels
)
_showstats(repo, stats)
if stats[3]:
repo.ui.status(
_(
"use 'hg resolve' to retry unresolved file merges "
"or 'hg update -C .' to abandon\n"
)
)
elif remind:
repo.ui.status(_("(branch merge, don't forget to commit)\n"))
return stats[3] > 0
def _incoming(displaychlist, ui, repo, source, opts, buffered=False):
"""
Helper for incoming / gincoming.
displaychlist gets called with
(remoterepo, incomingchangesetlist, displayer) parameters,
and is supposed to contain only code that can't be unified.
"""
source, branches = parseurl(ui.expandpath(source), opts.get("branch"))
other = peer(repo, opts, source)
ui.status(_("comparing with %s\n") % util.hidepassword(source))
revs, checkout = addbranchrevs(repo, other, branches, opts.get("rev"))
if revs:
revs = [other.lookup(rev) for rev in revs]
other, chlist, cleanupfn = bundlerepo.getremotechanges(
ui, repo, other, revs, opts["bundle"], opts["force"]
)
try:
if not chlist:
ui.status(_("no changes found\n"))
return
ui.pager("incoming")
displayer = cmdutil.show_changeset(ui, other, opts, buffered)
displaychlist(other, chlist, displayer)
displayer.close()
finally:
cleanupfn()
return 0 # exit code is zero since we found incoming changes
def incoming(ui, repo, source, opts):
def display(other, chlist, displayer):
limit = cmdutil.loglimit(opts)
if opts.get("newest_first"):
chlist.reverse()
count = 0
for n in chlist:
if limit is not None and count >= limit:
break
parents = [p for p in other.changelog.parents(n) if p != nullid]
if opts.get("no_merges") and len(parents) == 2:
continue
count += 1
displayer.show(other[n])
return _incoming(display, ui, repo, source, opts)
def _outgoing(ui, repo, dest, opts):
path = ui.paths.getpath(dest, default=("default-push", "default"))
if not path:
raise error.Abort(
_("default repository not configured!"),
hint=_("see 'hg help config.paths'"),
)
dest = path.pushloc or path.loc
branches = path.branch, opts.get("branch") or []
ui.status(_("comparing with %s\n") % util.hidepassword(dest))
revs, checkout = addbranchrevs(repo, repo, branches, opts.get("rev"))
if revs:
revs = [repo.lookup(rev) for rev in scmutil.revrange(repo, revs)]
other = peer(repo, opts, dest)
outgoing = discovery.findcommonoutgoing(repo, other, revs, force=opts.get("force"))
o = outgoing.missing
if not o:
scmutil.nochangesfound(repo.ui, repo, outgoing.excluded)
return o, other
def outgoing(ui, repo, dest, opts):
def recurse():
ret = 1
return ret
limit = cmdutil.loglimit(opts)
o, other = _outgoing(ui, repo, dest, opts)
if not o:
cmdutil.outgoinghooks(ui, repo, other, opts, o)
return recurse()
if opts.get("newest_first"):
o.reverse()
ui.pager("outgoing")
displayer = cmdutil.show_changeset(ui, repo, opts)
count = 0
for n in o:
if limit is not None and count >= limit:
break
parents = [p for p in repo.changelog.parents(n) if p != nullid]
if opts.get("no_merges") and len(parents) == 2:
continue
count += 1
displayer.show(repo[n])
displayer.close()
cmdutil.outgoinghooks(ui, repo, other, opts, o)
recurse()
return 0 # exit code is zero since we found outgoing changes
def verify(repo, revs=None):
"""verify the consistency of a repository
If revs is None, verify everything in the repository.
Otherwise, revs is a smartset that specifies revisions to verify
and some checks requiring knowledge about the entire repo will be skipped.
"""
ret = verifymod.verify(repo, revs)
return ret
def remoteui(src, opts):
"build a remote ui from ui or repo and opts"
if util.safehasattr(src, "baseui"): # looks like a repository
dst = src.baseui.copy() # drop repo-specific config
src = src.ui # copy target options from repo
else: # assume it's a global ui object
dst = src.copy() # keep all global options
# copy ssh-specific options
for o in "ssh", "remotecmd":
v = opts.get(o) or src.config("ui", o)
if v:
dst.setconfig("ui", o, v, "copied")
# copy bundle-specific options
r = src.config("bundle", "mainreporoot")
if r:
dst.setconfig("bundle", "mainreporoot", r, "copied")
# copy selected local settings to the remote ui
for sect in ("auth", "hostfingerprints", "hostsecurity", "http_proxy"):
for key, val in src.configitems(sect):
dst.setconfig(sect, key, val, "copied")
v = src.config("web", "cacerts")
if v:
dst.setconfig("web", "cacerts", util.expandpath(v), "copied")
return dst
# Files of interest
# Used to check if the repository has changed looking at mtime and size of
# these files.
foi = [
("spath", "00changelog.i"),
("spath", "phaseroots"), # ! phase can change content at the same size
("spath", "obsstore"),
("path", "bookmarks"), # ! bookmark can change content at the same size
]
class cachedlocalrepo(object):
"""Holds a localrepository that can be cached and reused."""
def __init__(self, repo):
"""Create a new cached repo from an existing repo.
We assume the passed in repo was recently created. If the
repo has changed between when it was created and when it was
turned into a cache, it may not refresh properly.
"""
assert isinstance(repo, localrepo.localrepository)
self._repo = repo
self._state, self.mtime = self._repostate()
def fetch(self):
"""Refresh (if necessary) and return a repository.
If the cached instance is out of date, it will be recreated
automatically and returned.
Returns a tuple of the repo and a boolean indicating whether a new
repo instance was created.
"""
# We compare the mtimes and sizes of some well-known files to
# determine if the repo changed. This is not precise, as mtimes
# are susceptible to clock skew and imprecise filesystems and
# file content can change while maintaining the same size.
state, mtime = self._repostate()
if state == self._state:
return self._repo, False
repo = repository(self._repo.baseui, self._repo.url())
self._repo = repo
self._state = state
self.mtime = mtime
return self._repo, True
def _repostate(self):
state = []
maxmtime = -1
for attr, fname in foi:
prefix = getattr(self._repo, attr)
p = os.path.join(prefix, fname)
try:
st = util.stat(p)
except OSError:
st = util.stat(prefix)
state.append((st.st_mtime, st.st_size))
maxmtime = max(maxmtime, st.st_mtime)
return tuple(state), maxmtime
def copy(self):
"""Obtain a copy of this class instance.
A new localrepository instance is obtained. The new instance should be
completely independent of the original.
"""
repo = repository(self._repo.baseui, self._repo.origroot)
c = cachedlocalrepo(repo)
c._state = self._state
c.mtime = self.mtime
return c