12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100110111021103110411051106110711081109111011111112111311141115111611171118111911201121112211231124112511261127112811291130113111321133113411351136113711381139114011411142114311441145114611471148114911501151115211531154115511561157115811591160116111621163116411651166116711681169117011711172117311741175117611771178117911801181118211831184118511861187118811891190119111921193119411951196119711981199120012011202120312041205120612071208120912101211121212131214121512161217121812191220122112221223122412251226122712281229123012311232123312341235123612371238123912401241124212431244124512461247124812491250125112521253125412551256125712581259126012611262126312641265126612671268126912701271127212731274127512761277127812791280128112821283128412851286128712881289129012911292129312941295129612971298129913001301130213031304130513061307130813091310131113121313131413151316131713181319132013211322132313241325132613271328132913301331133213331334133513361337133813391340134113421343134413451346134713481349135013511352135313541355135613571358135913601361136213631364136513661367136813691370137113721373137413751376137713781379138013811382138313841385138613871388138913901391139213931394139513961397139813991400140114021403140414051406140714081409141014111412141314141415141614171418141914201421142214231424142514261427142814291430143114321433143414351436143714381439144014411442144314441445144614471448144914501451145214531454145514561457145814591460146114621463146414651466146714681469147014711472147314741475147614771478147914801481148214831484148514861487148814891490149114921493149414951496149714981499150015011502150315041505150615071508150915101511151215131514151515161517151815191520152115221523152415251526152715281529153015311532153315341535153615371538153915401541154215431544154515461547154815491550155115521553155415551556155715581559156015611562156315641565156615671568156915701571157215731574157515761577157815791580158115821583158415851586158715881589159015911592159315941595159615971598159916001601160216031604160516061607160816091610161116121613161416151616161716181619162016211622162316241625162616271628162916301631163216331634163516361637163816391640164116421643164416451646164716481649165016511652165316541655165616571658165916601661166216631664166516661667166816691670167116721673167416751676167716781679168016811682168316841685168616871688168916901691169216931694169516961697169816991700170117021703170417051706170717081709171017111712171317141715171617171718171917201721172217231724172517261727172817291730173117321733173417351736173717381739174017411742174317441745174617471748174917501751175217531754175517561757175817591760176117621763176417651766176717681769177017711772177317741775177617771778177917801781178217831784178517861787178817891790179117921793179417951796179717981799180018011802180318041805180618071808180918101811181218131814181518161817181818191820182118221823182418251826182718281829183018311832183318341835183618371838183918401841184218431844184518461847184818491850185118521853185418551856185718581859186018611862186318641865186618671868186918701871187218731874187518761877187818791880188118821883188418851886188718881889189018911892189318941895189618971898189919001901190219031904190519061907190819091910191119121913191419151916191719181919192019211922192319241925192619271928192919301931193219331934193519361937193819391940194119421943194419451946194719481949195019511952195319541955195619571958195919601961196219631964196519661967196819691970197119721973197419751976197719781979198019811982198319841985198619871988198919901991199219931994199519961997199819992000200120022003200420052006200720082009201020112012201320142015201620172018201920202021202220232024202520262027202820292030203120322033203420352036203720382039204020412042204320442045204620472048204920502051205220532054205520562057205820592060206120622063206420652066206720682069207020712072207320742075207620772078207920802081208220832084208520862087208820892090209120922093209420952096209720982099210021012102210321042105210621072108210921102111211221132114211521162117211821192120212121222123212421252126212721282129213021312132213321342135213621372138213921402141214221432144214521462147214821492150215121522153215421552156215721582159216021612162216321642165216621672168216921702171217221732174217521762177217821792180218121822183218421852186218721882189219021912192219321942195219621972198219922002201220222032204220522062207220822092210221122122213221422152216221722182219222022212222222322242225222622272228222922302231223222332234223522362237223822392240224122422243224422452246224722482249225022512252225322542255225622572258225922602261226222632264226522662267226822692270227122722273227422752276227722782279228022812282228322842285228622872288228922902291 |
- # Copyright (c) 2012 The Chromium Authors. All rights reserved.
- # Use of this source code is governed by a BSD-style license that can be
- # found in the LICENSE file.
- """Gclient-specific SCM-specific operations."""
- import collections
- import contextlib
- import errno
- import glob
- import json
- import logging
- import os
- import platform
- import posixpath
- import re
- import shutil
- import sys
- import tempfile
- import threading
- import traceback
- import gclient_utils
- import gerrit_util
- import git_auth
- import git_cache
- import git_common
- import scm
- import subprocess2
- # TODO: Should fix these warnings.
- # pylint: disable=line-too-long
- class NoUsableRevError(gclient_utils.Error):
- """Raised if requested revision isn't found in checkout."""
- class DiffFiltererWrapper(object):
- """Simple base class which tracks which file is being diffed and
- replaces instances of its file name in the original and
- working copy lines of the git diff output."""
- index_string = None
- original_prefix = "--- "
- working_prefix = "+++ "
- def __init__(self, relpath, print_func):
- # Note that we always use '/' as the path separator to be
- # consistent with cygwin-style output on Windows
- self._relpath = relpath.replace("\\", "/")
- self._current_file = None
- self._print_func = print_func
- def SetCurrentFile(self, current_file):
- self._current_file = current_file
- @property
- def _replacement_file(self):
- return posixpath.join(self._relpath, self._current_file)
- def _Replace(self, line):
- return line.replace(self._current_file, self._replacement_file)
- def Filter(self, line):
- if (line.startswith(self.index_string)):
- self.SetCurrentFile(line[len(self.index_string):])
- line = self._Replace(line)
- else:
- if (line.startswith(self.original_prefix)
- or line.startswith(self.working_prefix)):
- line = self._Replace(line)
- self._print_func(line)
- class GitDiffFilterer(DiffFiltererWrapper):
- index_string = "diff --git "
- def SetCurrentFile(self, current_file):
- # Get filename by parsing "a/<filename> b/<filename>"
- self._current_file = current_file[:(len(current_file) / 2)][2:]
- def _Replace(self, line):
- return re.sub("[a|b]/" + self._current_file, self._replacement_file,
- line)
- # SCMWrapper base class
- class SCMWrapper(object):
- """Add necessary glue between all the supported SCM.
- This is the abstraction layer to bind to different SCM.
- """
- def __init__(self,
- url=None,
- root_dir=None,
- relpath=None,
- out_fh=None,
- out_cb=None,
- print_outbuf=False):
- self.url = url
- self._root_dir = root_dir
- if self._root_dir:
- self._root_dir = self._root_dir.replace('/', os.sep)
- self.relpath = relpath
- if self.relpath:
- self.relpath = self.relpath.replace('/', os.sep)
- if self.relpath and self._root_dir:
- self.checkout_path = os.path.join(self._root_dir, self.relpath)
- if out_fh is None:
- out_fh = sys.stdout
- self.out_fh = out_fh
- self.out_cb = out_cb
- self.print_outbuf = print_outbuf
- def Print(self, *args, **kwargs):
- kwargs.setdefault('file', self.out_fh)
- if kwargs.pop('timestamp', True):
- self.out_fh.write('[%s] ' % gclient_utils.Elapsed())
- print(*args, **kwargs)
- def RunCommand(self, command, options, args, file_list=None):
- commands = [
- 'update', 'updatesingle', 'revert', 'revinfo', 'status', 'diff',
- 'pack', 'runhooks'
- ]
- if not command in commands:
- raise gclient_utils.Error('Unknown command %s' % command)
- if not command in dir(self):
- raise gclient_utils.Error(
- 'Command %s not implemented in %s wrapper' %
- (command, self.__class__.__name__))
- return getattr(self, command)(options, args, file_list)
- @staticmethod
- def _get_first_remote_url(checkout_path):
- log = scm.GIT.YieldConfigRegexp(checkout_path, r'remote.*.url')
- return next(log)[1]
- def GetCacheMirror(self):
- if getattr(self, 'cache_dir', None):
- url, _ = gclient_utils.SplitUrlRevision(self.url)
- return git_cache.Mirror(url)
- return None
- def GetActualRemoteURL(self, options):
- """Attempt to determine the remote URL for this SCMWrapper."""
- # Git
- if os.path.exists(os.path.join(self.checkout_path, '.git')):
- actual_remote_url = self._get_first_remote_url(self.checkout_path)
- mirror = self.GetCacheMirror()
- # If the cache is used, obtain the actual remote URL from there.
- if (mirror and mirror.exists() and mirror.mirror_path.replace(
- '\\', '/') == actual_remote_url.replace('\\', '/')):
- actual_remote_url = self._get_first_remote_url(
- mirror.mirror_path)
- return actual_remote_url
- return None
- def DoesRemoteURLMatch(self, options):
- """Determine whether the remote URL of this checkout is the expected URL."""
- if not os.path.exists(self.checkout_path):
- # A checkout which doesn't exist can't be broken.
- return True
- actual_remote_url = self.GetActualRemoteURL(options)
- if actual_remote_url:
- return (gclient_utils.SplitUrlRevision(actual_remote_url)[0].rstrip(
- '/') == gclient_utils.SplitUrlRevision(self.url)[0].rstrip('/'))
- # This may occur if the self.checkout_path exists but does not contain a
- # valid git checkout.
- return False
- def _DeleteOrMove(self, force):
- """Delete the checkout directory or move it out of the way.
- Args:
- force: bool; if True, delete the directory. Otherwise, just move it.
- """
- if force and os.environ.get('CHROME_HEADLESS') == '1':
- self.Print('_____ Conflicting directory found in %s. Removing.' %
- self.checkout_path)
- gclient_utils.AddWarning('Conflicting directory %s deleted.' %
- self.checkout_path)
- gclient_utils.rmtree(self.checkout_path)
- else:
- bad_scm_dir = os.path.join(self._root_dir, '_bad_scm',
- os.path.dirname(self.relpath))
- try:
- os.makedirs(bad_scm_dir)
- except OSError as e:
- if e.errno != errno.EEXIST:
- raise
- dest_path = tempfile.mkdtemp(prefix=os.path.basename(self.relpath),
- dir=bad_scm_dir)
- self.Print(
- '_____ Conflicting directory found in %s. Moving to %s.' %
- (self.checkout_path, dest_path))
- gclient_utils.AddWarning('Conflicting directory %s moved to %s.' %
- (self.checkout_path, dest_path))
- shutil.move(self.checkout_path, dest_path)
- class GitWrapper(SCMWrapper):
- """Wrapper for Git"""
- name = 'git'
- remote = 'origin'
- @property
- def cache_dir(self):
- try:
- return git_cache.Mirror.GetCachePath()
- except RuntimeError:
- return None
- def __init__(self, url=None, *args, **kwargs):
- """Removes 'git+' fake prefix from git URL."""
- if url and (url.startswith('git+http://')
- or url.startswith('git+https://')):
- url = url[4:]
- SCMWrapper.__init__(self, url, *args, **kwargs)
- filter_kwargs = {'time_throttle': 1, 'out_fh': self.out_fh}
- if self.out_cb:
- filter_kwargs['predicate'] = self.out_cb
- self.filter = gclient_utils.GitFilter(**filter_kwargs)
- self._running_under_rosetta = None
- self.current_revision = None
- def GetCheckoutRoot(self):
- return scm.GIT.GetCheckoutRoot(self.checkout_path)
- def GetRevisionDate(self, _revision):
- """Returns the given revision's date in ISO-8601 format (which contains the
- time zone)."""
- # TODO(floitsch): get the time-stamp of the given revision and not just
- # the time-stamp of the currently checked out revision.
- return self._Capture(['log', '-n', '1', '--format=%ai'])
- def _GetDiffFilenames(self, base):
- """Returns the names of files modified since base."""
- return self._Capture(
- # Filter to remove base if it is None.
- list(
- filter(
- bool,
- ['-c', 'core.quotePath=false', 'diff', '--name-only', base])
- )).split()
- def GetSubmoduleStateFromIndex(self):
- """Returns a map where keys are submodule names and values are commit
- hashes. It reads data from the Git index, so only committed values are
- present."""
- out = self._Capture(['ls-files', '-s'])
- result = {}
- for l in out.split('\n'):
- if not l.startswith('160000'):
- # Not a submodule
- continue
- (_, commit, _, filepath) = l.split(maxsplit=3)
- result[filepath] = commit
- return result
- def GetSubmoduleDiff(self):
- """Returns a map where keys are submodule names and values are tuples of
- (old_commit_hash, new_commit_hash). old_commit_hash matches the Git
- index, whereas new_commit_hash matches currently checked out commit
- hash."""
- out = self._Capture([
- 'diff',
- '--no-prefix',
- '--no-ext-diff',
- '--no-color',
- '--ignore-submodules=dirty',
- '--submodule=short',
- '-G',
- 'Subproject commit',
- ])
- NO_COMMIT = 40 * '0'
- committed_submodule = None
- checked_submodule = None
- filepath = None
- state = 0
- diff = {}
- # Parsing git diff uses simple state machine. States:
- # 0 - start state
- # 1 - diff file/line detected, ready to process content
- # 2 - gitlink detected, ready to process gitlink past and current
- # content.
- # 3 - past gitlink content detected. It contains a commit hash that's in
- # git index.
- # 4 - new gitlink content detected. It contains currently checked
- # commit. At this point, we have all information needed, and we can
- # reset state to 0.
- for l in out.split('\n'):
- if l.startswith('diff --git'):
- # New file detected, reset state.
- state = 1
- elif state == 1 and l.startswith('index') and l.endswith('160000'):
- # We detected gitlink
- state = 2
- elif state == 2 and l.startswith('+++ '):
- # This line contains filename
- filepath = l[4:]
- state = 3
- elif state == 3 and l.startswith('-Subproject commit '):
- # This line contains what commit hash Git index expects
- # (ls-files).
- committed_submodule = l.split(' ')[-1]
- state = 4
- elif state == 4 and l.startswith('+Subproject commit '):
- # This line contains currently checked out commit for this submodule.
- checked_submodule = l.split(' ')[-1]
- if NO_COMMIT not in (committed_submodule, checked_submodule):
- diff[filepath] = (committed_submodule, checked_submodule)
- state = 0
- return diff
- def diff(self, options, _args, _file_list):
- _, revision = gclient_utils.SplitUrlRevision(self.url)
- if not revision:
- revision = 'refs/remotes/%s/main' % self.remote
- self._Run(['-c', 'core.quotePath=false', 'diff', revision], options)
- def pack(self, _options, _args, _file_list):
- """Generates a patch file which can be applied to the root of the
- repository.
- The patch file is generated from a diff of the merge base of HEAD and
- its upstream branch.
- """
- try:
- merge_base = [self._Capture(['merge-base', 'HEAD', self.remote])]
- except subprocess2.CalledProcessError:
- merge_base = []
- gclient_utils.CheckCallAndFilter(['git', 'diff'] + merge_base,
- cwd=self.checkout_path,
- filter_fn=GitDiffFilterer(
- self.relpath,
- print_func=self.Print).Filter)
- def _Scrub(self, target, options):
- """Scrubs out all changes in the local repo, back to the state of target."""
- quiet = []
- if not options.verbose:
- quiet = ['--quiet']
- self._Run(['reset', '--hard', target] + quiet, options)
- if options.force and options.delete_unversioned_trees:
- # where `target` is a commit that contains both upper and lower case
- # versions of the same file on a case insensitive filesystem, we are
- # actually in a broken state here. The index will have both 'a' and
- # 'A', but only one of them will exist on the disk. To progress, we
- # delete everything that status thinks is modified.
- output = self._Capture(
- ['-c', 'core.quotePath=false', 'status', '--porcelain'],
- strip=False)
- for line in output.splitlines():
- # --porcelain (v1) looks like:
- # XY filename
- try:
- filename = line[3:]
- self.Print('_____ Deleting residual after reset: %r.' %
- filename)
- gclient_utils.rm_file_or_tree(
- os.path.join(self.checkout_path, filename))
- except OSError:
- pass
- def _FetchAndReset(self, revision, file_list, options):
- """Equivalent to git fetch; git reset."""
- self._SetFetchConfig(options)
- self._Fetch(options, prune=True, quiet=options.verbose)
- revision = self._AutoFetchRef(options, revision)
- self._Scrub(revision, options)
- if file_list is not None:
- files = self._Capture(['-c', 'core.quotePath=false',
- 'ls-files']).splitlines()
- file_list.extend(
- [os.path.join(self.checkout_path, f) for f in files])
- def _DisableHooks(self):
- hook_dir = os.path.join(self.checkout_path, '.git', 'hooks')
- if not os.path.isdir(hook_dir):
- return
- for f in os.listdir(hook_dir):
- if not f.endswith('.sample') and not f.endswith('.disabled'):
- disabled_hook_path = os.path.join(hook_dir, f + '.disabled')
- if os.path.exists(disabled_hook_path):
- os.remove(disabled_hook_path)
- os.rename(os.path.join(hook_dir, f), disabled_hook_path)
- def _maybe_break_locks(self, options):
- """This removes all .lock files from this repo's .git directory, if the
- user passed the --break_repo_locks command line flag.
- In particular, this will cleanup index.lock files, as well as ref lock
- files.
- """
- if options.break_repo_locks:
- git_dir = os.path.join(self.checkout_path, '.git')
- for path, _, filenames in os.walk(git_dir):
- for filename in filenames:
- if filename.endswith('.lock'):
- to_break = os.path.join(path, filename)
- self.Print('breaking lock: %s' % (to_break, ))
- try:
- os.remove(to_break)
- except OSError as ex:
- self.Print('FAILED to break lock: %s: %s' %
- (to_break, ex))
- raise
- def _download_topics(self, patch_rev, googlesource_url):
- """This method returns new patch_revs to process that have the same topic.
- It does the following:
- 1. Finds the topic of the Gerrit change specified in the patch_rev.
- 2. Find all changes with that topic.
- 3. Append patch_rev of the changes with the same topic to the patch_revs
- to process.
- 4. Returns the new patch_revs to process.
- """
- patch_revs_to_process = []
- # Parse the patch_rev to extract the CL and patchset.
- patch_rev_tokens = patch_rev.split('/')
- change = patch_rev_tokens[-2]
- # Parse the googlesource_url.
- tokens = re.search(r'//(.+).googlesource.com/(.+?)(?:\.git)?$',
- googlesource_url)
- if not tokens or len(tokens.groups()) != 2:
- # googlesource_url is not in the expected format.
- return patch_revs_to_process
- # parse the gerrit host and repo out of googlesource_url.
- host, repo = tokens.groups()[:2]
- gerrit_host_url = '%s-review.googlesource.com' % host
- # 1. Find the topic of the Gerrit change specified in the patch_rev.
- change_object = gerrit_util.GetChange(gerrit_host_url, change)
- topic = change_object.get('topic')
- if not topic:
- # This change has no topic set.
- return patch_revs_to_process
- # 2. Find all changes with that topic.
- changes_with_same_topic = gerrit_util.QueryChanges(
- gerrit_host_url, [('topic', topic), ('status', 'open'),
- ('repo', repo)],
- o_params=['ALL_REVISIONS'])
- for c in changes_with_same_topic:
- if str(c['_number']) == change:
- # This change is already in the patch_rev.
- continue
- self.Print('Found CL %d with the topic name %s' %
- (c['_number'], topic))
- # 3. Append patch_rev of the changes with the same topic to the
- # patch_revs to process.
- curr_rev = c['current_revision']
- new_patch_rev = c['revisions'][curr_rev]['ref']
- patch_revs_to_process.append(new_patch_rev)
- # 4. Return the new patch_revs to process.
- return patch_revs_to_process
- def _ref_to_remote_ref(self, target_rev):
- """Helper function for scm.GIT.RefToRemoteRef with error checking.
- Joins the results of scm.GIT.RefToRemoteRef into a string, but raises a
- comprehensible error if RefToRemoteRef fails.
- Args:
- target_rev: a ref somewhere under refs/.
- """
- tmp_ref = scm.GIT.RefToRemoteRef(target_rev, self.remote)
- if not tmp_ref:
- raise gclient_utils.Error(
- 'Failed to turn target revision %r in repo %r into remote ref' %
- (target_rev, self.checkout_path))
- return ''.join(tmp_ref)
- def apply_patch_ref(self, patch_repo, patch_rev, target_rev, options,
- file_list):
- # type: (str, str, str, optparse.Values, Collection[str]) -> str
- """Apply a patch on top of the revision we're synced at.
- The patch ref is given by |patch_repo|@|patch_rev|.
- |target_rev| is usually the branch that the |patch_rev| was uploaded against
- (e.g. 'refs/heads/main'), but this is not required.
- We cherry-pick all commits reachable from |patch_rev| on top of the curret
- HEAD, excluding those reachable from |target_rev|
- (i.e. git cherry-pick target_rev..patch_rev).
- Graphically, it looks like this:
- ... -> o -> [possibly already landed commits] -> target_rev
- \
- -> [possibly not yet landed dependent CLs] -> patch_rev
- The final checkout state is then:
- ... -> HEAD -> [possibly not yet landed dependent CLs] -> patch_rev
- After application, if |options.reset_patch_ref| is specified, we soft reset
- the cherry-picked changes, keeping them in git index only.
- Args:
- patch_repo: The patch origin.
- e.g. 'https://foo.googlesource.com/bar'
- patch_rev: The revision to patch.
- e.g. 'refs/changes/1234/34/1'.
- target_rev: The revision to use when finding the merge base.
- Typically, the branch that the patch was uploaded against.
- e.g. 'refs/heads/main' or 'refs/heads/infra/config'.
- options: The options passed to gclient.
- file_list: A list where modified files will be appended.
- """
- # Abort any cherry-picks in progress.
- try:
- self._Capture(['cherry-pick', '--abort'])
- except subprocess2.CalledProcessError:
- pass
- base_rev = self.revinfo(None, None, None)
- if not target_rev:
- raise gclient_utils.Error(
- 'A target revision for the patch must be given')
- if target_rev.startswith(('refs/heads/', 'refs/branch-heads')):
- # If |target_rev| is in refs/heads/** or refs/branch-heads/**, try
- # first to find the corresponding remote ref for it, since
- # |target_rev| might point to a local ref which is not up to date
- # with the corresponding remote ref.
- remote_ref = self._ref_to_remote_ref(target_rev)
- self.Print('Trying the corresponding remote ref for %r: %r\n' %
- (target_rev, remote_ref))
- if scm.GIT.IsValidRevision(self.checkout_path, remote_ref):
- # refs/remotes may need to be updated to cleanly cherry-pick
- # changes. See https://crbug.com/1255178.
- url, _ = gclient_utils.SplitUrlRevision(self.url)
- mirror = self._GetMirror(url, options, target_rev, remote_ref)
- if mirror:
- self._UpdateMirrorIfNotContains(mirror, options, 'branch',
- target_rev)
- self._Capture(['fetch', '--no-tags', self.remote, target_rev])
- target_rev = remote_ref
- elif not scm.GIT.IsValidRevision(self.checkout_path, target_rev):
- # Fetch |target_rev| if it's not already available.
- url, _ = gclient_utils.SplitUrlRevision(self.url)
- mirror = self._GetMirror(url, options, target_rev, target_rev)
- if mirror:
- rev_type = 'branch' if target_rev.startswith(
- 'refs/') else 'hash'
- self._UpdateMirrorIfNotContains(mirror, options, rev_type,
- target_rev)
- self._Fetch(options, refspec=target_rev)
- patch_revs_to_process = [patch_rev]
- if hasattr(options, 'download_topics') and options.download_topics:
- patch_revs_to_process_from_topics = self._download_topics(
- patch_rev, self.url)
- patch_revs_to_process.extend(patch_revs_to_process_from_topics)
- self._Capture(['reset', '--hard'])
- for pr in patch_revs_to_process:
- self.Print('===Applying patch===')
- self.Print('Revision to patch is %r @ %r.' % (patch_repo, pr))
- self.Print('Current dir is %r' % self.checkout_path)
- self._Capture(['fetch', '--no-tags', patch_repo, pr])
- pr = self._Capture(['rev-parse', 'FETCH_HEAD'])
- if not options.rebase_patch_ref:
- self._Capture(['checkout', pr])
- # Adjust base_rev to be the first parent of our checked out
- # patch ref; This will allow us to correctly extend `file_list`,
- # and will show the correct file-list to programs which do `git
- # diff --cached` expecting to see the patch diff.
- base_rev = self._Capture(['rev-parse', pr + '~'])
- else:
- target_rev_hash = self._Capture(['rev-parse', target_rev])
- commit_list = self._Capture(
- ['log', '--oneline', target_rev + '..' + pr])
- self.Print('Will cherrypick %r (%r) .. %r on top of %r:' %
- (target_rev_hash, target_rev, pr, base_rev))
- self.Print(commit_list)
- try:
- if scm.GIT.IsAncestor(pr,
- target_rev,
- cwd=self.checkout_path):
- if len(patch_revs_to_process) > 1:
- # If there are multiple patch_revs_to_process then
- # we do not want want to invalidate a previous patch
- # so throw an error.
- raise gclient_utils.Error(
- 'patch_rev %s is an ancestor of target_rev %s. This '
- 'situation is unsupported when we need to apply multiple '
- 'patch_revs: %s' %
- (pr, target_rev, patch_revs_to_process))
- # If |patch_rev| is an ancestor of |target_rev|, check
- # it out.
- self._Capture(['checkout', pr])
- else:
- # If a change was uploaded on top of another change,
- # which has already landed, one of the commits in the
- # cherry-pick range will be redundant, since it has
- # already landed and its changes incorporated in the
- # tree. We pass '--keep-redundant-commits' to ignore
- # those changes.
- self._Capture([
- 'cherry-pick', target_rev + '..' + pr,
- '--keep-redundant-commits'
- ])
- except subprocess2.CalledProcessError as e:
- self.Print('Failed to apply patch.')
- self.Print('Revision to patch was %r @ %r.' %
- (patch_repo, pr))
- self.Print('Tried to cherrypick %r .. %r on top of %r.' %
- (target_rev, pr, base_rev))
- self.Print('Current dir is %r' % self.checkout_path)
- self.Print('git returned non-zero exit status %s:\n%s' %
- (e.returncode, e.stderr.decode('utf-8')))
- # Print the current status so that developers know what
- # changes caused the patch failure, since git cherry-pick
- # doesn't show that information.
- self.Print(self._Capture(['status']))
- try:
- self._Capture(['cherry-pick', '--abort'])
- except subprocess2.CalledProcessError:
- pass
- raise
- if file_list is not None:
- file_list.extend(self._GetDiffFilenames(base_rev))
- latest_commit = self.revinfo(None, None, None)
- if options.reset_patch_ref:
- self._Capture(['reset', '--soft', base_rev])
- return latest_commit
- def check_diff(self, previous_commit, files=None):
- # type: (str, Optional[List[str]]) -> bool
- """Check if a diff exists between the current commit and `previous_commit`.
- Returns True if there were diffs or if an error was encountered.
- """
- cmd = ['diff', previous_commit, '--quiet']
- if files:
- cmd += ['--'] + files
- try:
- self._Capture(cmd)
- return False
- except subprocess2.CalledProcessError as e:
- # git diff --quiet exits with 1 if there were diffs.
- if e.returncode != 1:
- self.Print('git returned non-zero exit status %s:\n%s' %
- (e.returncode, e.stderr.decode('utf-8')))
- return True
- def set_config(f):
- def wrapper(*args):
- return_val = f(*args)
- checkout_path = args[0].checkout_path
- if os.path.exists(os.path.join(checkout_path, '.git')):
- # The config updates to the project are stored in this list
- # and updated consecutively after the reads. The updates
- # are done this way because `scm.GIT.GetConfig` caches
- # the config file and `scm.GIT.SetConfig` evicts the cache.
- # This ensures we don't interleave reads and writes causing
- # the cache to set and unset consecutively.
- config_updates = []
- blame_ignore_revs_cfg = scm.GIT.GetConfig(
- checkout_path, 'blame.ignorerevsfile')
- blame_ignore_revs_cfg_set = \
- blame_ignore_revs_cfg == \
- git_common.GIT_BLAME_IGNORE_REV_FILE
- blame_ignore_revs_exists = os.path.isfile(
- os.path.join(checkout_path,
- git_common.GIT_BLAME_IGNORE_REV_FILE))
- if not blame_ignore_revs_cfg_set and blame_ignore_revs_exists:
- config_updates.append(
- ('blame.ignoreRevsFile',
- git_common.GIT_BLAME_IGNORE_REV_FILE))
- elif blame_ignore_revs_cfg_set and not blame_ignore_revs_exists:
- # Some repos may have incorrect config set, unset this
- # value. Moreover, some repositories may decide to remove
- # git_common.GIT_BLAME_IGNORE_REV_FILE, which would break
- # blame without this check.
- # See https://crbug.com/368562244 for more details.
- config_updates.append(('blame.ignoreRevsFile', None))
- ignore_submodules = scm.GIT.GetConfig(checkout_path,
- 'diff.ignoresubmodules',
- None, 'local')
- if not ignore_submodules:
- config_updates.append(('diff.ignoreSubmodules', 'dirty'))
- elif ignore_submodules != 'dirty':
- warning_message = (
- "diff.ignoreSubmodules is not set to 'dirty' "
- "for this repository.\n"
- "This may cause unexpected behavior with submodules; "
- "see //docs/git_submodules.md\n"
- "Consider setting the config:\n"
- "\tgit config diff.ignoreSubmodules dirty\n"
- "or disable this warning by setting the "
- "GCLIENT_SUPPRESS_SUBMODULE_WARNING\n"
- "environment variable to 1.")
- if os.environ.get(
- 'GCLIENT_SUPPRESS_SUBMODULE_WARNING') != '1':
- gclient_utils.AddWarning(warning_message)
- if scm.GIT.GetConfig(checkout_path,
- 'fetch.recursesubmodules') != 'off':
- config_updates.append(('fetch.recurseSubmodules', 'off'))
- if scm.GIT.GetConfig(checkout_path,
- 'push.recursesubmodules') != 'off':
- # The default is off, but if user sets submodules.recurse to
- # on, this becomes on too. We never want to push submodules
- # for gclient managed repositories. Push, even if a no-op,
- # will increase `git cl upload` latency.
- config_updates.append(('push.recurseSubmodules', 'off'))
- for update in config_updates:
- scm.GIT.SetConfig(checkout_path,
- update[0],
- update[1],
- modify_all=True)
- return return_val
- return wrapper
- @set_config
- def update(self, options, args, file_list):
- """Runs git to update or transparently checkout the working copy.
- All updated files will be appended to file_list.
- Raises:
- Error: if can't get URL for relative path.
- """
- if args:
- raise gclient_utils.Error("Unsupported argument(s): %s" %
- ",".join(args))
- url, deps_revision = gclient_utils.SplitUrlRevision(self.url)
- revision = deps_revision
- managed = True
- if options.revision:
- # Override the revision number.
- revision = str(options.revision)
- if revision == 'unmanaged':
- # Check again for a revision in case an initial ref was specified
- # in the url, for example bla.git@refs/heads/custombranch
- revision = deps_revision
- managed = False
- if not revision:
- # If a dependency is not pinned, track the default remote branch.
- revision = scm.GIT.GetRemoteHeadRef(self.checkout_path, self.url,
- self.remote)
- if revision.startswith('origin/'):
- revision = 'refs/remotes/' + revision
- if managed and platform.system() == 'Windows':
- self._DisableHooks()
- printed_path = False
- verbose = []
- if options.verbose:
- self.Print('_____ %s at %s' % (self.relpath, revision),
- timestamp=False)
- verbose = ['--verbose']
- printed_path = True
- revision_ref = revision
- if ':' in revision:
- revision_ref, _, revision = revision.partition(':')
- if revision_ref.startswith('refs/branch-heads'):
- options.with_branch_heads = True
- mirror = self._GetMirror(url, options, revision, revision_ref)
- if mirror:
- url = mirror.mirror_path
- remote_ref = scm.GIT.RefToRemoteRef(revision, self.remote)
- if remote_ref:
- # Rewrite remote refs to their local equivalents.
- revision = ''.join(remote_ref)
- rev_type = "branch"
- elif revision.startswith('refs/heads/'):
- # Local branch? We probably don't want to support, since DEPS should
- # always specify branches as they are in the upstream repo.
- rev_type = "branch"
- else:
- # hash is also a tag, only make a distinction at checkout.
- # Any ref (e.g. /refs/changes/*) not a branch has no difference from
- # a hash.
- rev_type = "hash"
- # If we are going to introduce a new project, there is a possibility
- # that we are syncing back to a state where the project was originally a
- # sub-project rolled by DEPS (realistic case: crossing the Blink merge
- # point syncing backwards, when Blink was a DEPS entry and not part of
- # src.git). In such case, we might have a backup of the former .git
- # folder, which can be used to avoid re-fetching the entire repo again
- # (useful for bisects).
- backup_dir = self.GetGitBackupDirPath()
- target_dir = os.path.join(self.checkout_path, '.git')
- if os.path.exists(backup_dir) and not os.path.exists(target_dir):
- gclient_utils.safe_makedirs(self.checkout_path)
- os.rename(backup_dir, target_dir)
- # Reset to a clean state
- self._Scrub('HEAD', options)
- if (not os.path.exists(self.checkout_path) or
- (os.path.isdir(self.checkout_path)
- and not os.path.exists(os.path.join(self.checkout_path, '.git')))):
- if mirror:
- self._UpdateMirrorIfNotContains(mirror, options, rev_type,
- revision)
- try:
- self.current_revision = self._Clone(revision, url, options)
- except subprocess2.CalledProcessError as e:
- logging.warning('Clone failed due to: %s', e)
- self._DeleteOrMove(options.force)
- self.current_revision = self._Clone(revision, url, options)
- if file_list is not None:
- files = self._Capture(
- ['-c', 'core.quotePath=false', 'ls-files']).splitlines()
- file_list.extend(
- [os.path.join(self.checkout_path, f) for f in files])
- if mirror:
- self._Capture(
- ['remote', 'set-url', '--push', 'origin', mirror.url])
- if not verbose:
- # Make the output a little prettier. It's nice to have some
- # whitespace between projects when cloning.
- self.Print('')
- return self._Capture(['rev-parse', '--verify', 'HEAD'])
- if mirror:
- self._Capture(['remote', 'set-url', '--push', 'origin', mirror.url])
- if not managed:
- self._SetFetchConfig(options)
- self.Print('________ unmanaged solution; skipping %s' %
- self.relpath)
- return self._Capture(['rev-parse', '--verify', 'HEAD'])
- # Special case for rev_type = hash. If we use submodules, we can check
- # information already.
- if rev_type == 'hash':
- if self.current_revision == revision:
- if verbose:
- self.Print('Using submodule information to skip check')
- if options.reset or options.force:
- self._Scrub('HEAD', options)
- return revision
- self._maybe_break_locks(options)
- if mirror:
- self._UpdateMirrorIfNotContains(mirror, options, rev_type, revision)
- # See if the url has changed (the unittests use git://foo for the url,
- # let that through).
- current_url = scm.GIT.GetConfig(self.checkout_path,
- f'remote.{self.remote}.url',
- default='')
- return_early = False
- # TODO(maruel): Delete url != 'git://foo' since it's just to make the
- # unit test pass. (and update the comment above)
- strp_url = url[:-4] if url.endswith('.git') else url
- strp_current_url = current_url[:-4] if current_url.endswith(
- '.git') else current_url
- if (strp_current_url.rstrip('/') != strp_url.rstrip('/')
- and url != 'git://foo'):
- self.Print('_____ switching %s from %s to new upstream %s' %
- (self.relpath, current_url, url))
- if not (options.force or options.reset):
- # Make sure it's clean
- self._CheckClean(revision)
- # Switch over to the new upstream
- self._Run(['remote', 'set-url', self.remote, url], options)
- if mirror:
- if git_cache.Mirror.CacheDirToUrl(current_url.rstrip(
- '/')) == git_cache.Mirror.CacheDirToUrl(
- url.rstrip('/')):
- # Reset alternates when the cache dir is updated.
- with open(
- os.path.join(self.checkout_path, '.git', 'objects',
- 'info', 'alternates'), 'w') as fh:
- fh.write(os.path.join(url, 'objects'))
- else:
- # Because we use Git alternatives, our existing repository
- # is not self-contained. It's possible that new git
- # alternative doesn't have all necessary objects that the
- # current repository needs. Instead of blindly hoping that
- # new alternative contains all necessary objects, keep the
- # old alternative and just append a new one on top of it.
- with open(
- os.path.join(self.checkout_path, '.git', 'objects',
- 'info', 'alternates'), 'a') as fh:
- fh.write("\n" + os.path.join(url, 'objects'))
- current_revision = self._EnsureValidHeadObjectOrCheckout(
- revision, options, url)
- self._FetchAndReset(revision, file_list, options)
- return_early = True
- else:
- current_revision = self._EnsureValidHeadObjectOrCheckout(
- revision, options, url)
- if return_early:
- return current_revision or self._Capture(
- ['rev-parse', '--verify', 'HEAD'])
- cur_branch = self._GetCurrentBranch()
- # Cases:
- # 0) HEAD is detached. Probably from our initial clone.
- # - make sure HEAD is contained by a named ref, then update.
- # Cases 1-4. HEAD is a branch.
- # 1) current branch is not tracking a remote branch
- # - try to rebase onto the new hash or branch
- # 2) current branch is tracking a remote branch with local committed
- # changes, but the DEPS file switched to point to a hash
- # - rebase those changes on top of the hash
- # 3) current branch is tracking a remote branch w/or w/out changes, and
- # no DEPS switch
- # - see if we can FF, if not, prompt the user for rebase, merge, or stop
- # 4) current branch is tracking a remote branch, but DEPS switches to a
- # different remote branch, and a) current branch has no local changes,
- # and --force: - checkout new branch b) current branch has local
- # changes, and --force and --reset: - checkout new branch c) otherwise
- # exit
- # GetUpstreamBranch returns something like 'refs/remotes/origin/main'
- # for a tracking branch or 'main' if not a tracking branch (it's based
- # on a specific rev/hash) or it returns None if it couldn't find an
- # upstream
- if cur_branch is None:
- upstream_branch = None
- current_type = "detached"
- logging.debug("Detached HEAD")
- else:
- upstream_branch = scm.GIT.GetUpstreamBranch(self.checkout_path)
- if not upstream_branch or not upstream_branch.startswith(
- 'refs/remotes'):
- current_type = "hash"
- logging.debug(
- "Current branch is not tracking an upstream (remote)"
- " branch.")
- elif upstream_branch.startswith('refs/remotes'):
- current_type = "branch"
- else:
- raise gclient_utils.Error('Invalid Upstream: %s' %
- upstream_branch)
- self._SetFetchConfig(options)
- # Fetch upstream if we don't already have |revision|.
- if not scm.GIT.IsValidRevision(
- self.checkout_path, revision, sha_only=True):
- self._Fetch(options, prune=options.force)
- if not scm.GIT.IsValidRevision(
- self.checkout_path, revision, sha_only=True):
- # Update the remotes first so we have all the refs.
- remote_output = scm.GIT.Capture(['remote'] + verbose +
- ['update'],
- cwd=self.checkout_path)
- if verbose:
- self.Print(remote_output)
- revision = self._AutoFetchRef(options, revision)
- # This is a big hammer, debatable if it should even be here...
- if options.force or options.reset:
- target = 'HEAD'
- if options.upstream and upstream_branch:
- target = upstream_branch
- self._Scrub(target, options)
- if current_type == 'detached':
- # case 0
- # We just did a Scrub, this is as clean as it's going to get. In
- # particular if HEAD is a commit that contains two versions of the
- # same file on a case-insensitive filesystem (e.g. 'a' and 'A'),
- # there's no way to actually "Clean" the checkout; that commit is
- # uncheckoutable on this system. The best we can do is carry forward
- # to the checkout step.
- if not (options.force or options.reset):
- self._CheckClean(revision)
- self._CheckDetachedHead(revision, options)
- if not current_revision:
- current_revision = self._Capture(
- ['rev-list', '-n', '1', 'HEAD'])
- if current_revision == revision:
- self.Print('Up-to-date; skipping checkout.')
- else:
- # 'git checkout' may need to overwrite existing untracked files.
- # Allow it only when nuclear options are enabled.
- self._Checkout(
- options,
- revision,
- force=(options.force and options.delete_unversioned_trees),
- quiet=True,
- )
- if not printed_path:
- self.Print('_____ %s at %s' % (self.relpath, revision),
- timestamp=False)
- elif current_type == 'hash':
- # case 1
- # Can't find a merge-base since we don't know our upstream. That
- # makes this command VERY likely to produce a rebase failure. For
- # now we assume origin is our upstream since that's what the old
- # behavior was.
- upstream_branch = self.remote
- if options.revision or deps_revision:
- upstream_branch = revision
- self._AttemptRebase(upstream_branch,
- file_list,
- options,
- printed_path=printed_path,
- merge=options.merge)
- printed_path = True
- elif rev_type == 'hash':
- # case 2
- self._AttemptRebase(upstream_branch,
- file_list,
- options,
- newbase=revision,
- printed_path=printed_path,
- merge=options.merge)
- printed_path = True
- elif remote_ref and ''.join(remote_ref) != upstream_branch:
- # case 4
- new_base = ''.join(remote_ref)
- if not printed_path:
- self.Print('_____ %s at %s' % (self.relpath, revision),
- timestamp=False)
- switch_error = (
- "Could not switch upstream branch from %s to %s\n" %
- (upstream_branch, new_base) +
- "Please use --force or merge or rebase manually:\n" +
- "cd %s; git rebase %s\n" % (self.checkout_path, new_base) +
- "OR git checkout -b <some new branch> %s" % new_base)
- force_switch = False
- if options.force:
- try:
- self._CheckClean(revision)
- # case 4a
- force_switch = True
- except gclient_utils.Error as e:
- if options.reset:
- # case 4b
- force_switch = True
- else:
- switch_error = '%s\n%s' % (str(e), switch_error)
- if force_switch:
- self.Print("Switching upstream branch from %s to %s" %
- (upstream_branch, new_base))
- switch_branch = 'gclient_' + remote_ref[1]
- self._Capture(['branch', '-f', switch_branch, new_base])
- self._Checkout(options, switch_branch, force=True, quiet=True)
- else:
- # case 4c
- raise gclient_utils.Error(switch_error)
- else:
- # case 3 - the default case
- rebase_files = self._GetDiffFilenames(upstream_branch)
- if verbose:
- self.Print('Trying fast-forward merge to branch : %s' %
- upstream_branch)
- try:
- merge_args = ['merge']
- if options.merge:
- merge_args.append('--ff')
- else:
- merge_args.append('--ff-only')
- merge_args.append(upstream_branch)
- merge_output = self._Capture(merge_args)
- except subprocess2.CalledProcessError as e:
- rebase_files = []
- if re.search(b'fatal: Not possible to fast-forward, aborting.',
- e.stderr):
- if not printed_path:
- self.Print('_____ %s at %s' % (self.relpath, revision),
- timestamp=False)
- printed_path = True
- while True:
- if not options.auto_rebase:
- try:
- action = self._AskForData(
- 'Cannot %s, attempt to rebase? '
- '(y)es / (q)uit / (s)kip : ' %
- ('merge' if options.merge else
- 'fast-forward merge'), options)
- except ValueError:
- raise gclient_utils.Error('Invalid Character')
- if options.auto_rebase or re.match(
- r'yes|y', action, re.I):
- self._AttemptRebase(upstream_branch,
- rebase_files,
- options,
- printed_path=printed_path,
- merge=False)
- printed_path = True
- break
- if re.match(r'quit|q', action, re.I):
- raise gclient_utils.Error(
- "Can't fast-forward, please merge or "
- "rebase manually.\n"
- "cd %s && git " % self.checkout_path +
- "rebase %s" % upstream_branch)
- if re.match(r'skip|s', action, re.I):
- self.Print('Skipping %s' % self.relpath)
- return
- self.Print('Input not recognized')
- elif re.match(
- b"error: Your local changes to '.*' would be "
- b"overwritten by merge. Aborting.\nPlease, commit your "
- b"changes or stash them before you can merge.\n",
- e.stderr):
- if not printed_path:
- self.Print('_____ %s at %s' % (self.relpath, revision),
- timestamp=False)
- printed_path = True
- raise gclient_utils.Error(e.stderr.decode('utf-8'))
- else:
- # Some other problem happened with the merge
- logging.error("Error during fast-forward merge in %s!" %
- self.relpath)
- self.Print(e.stderr.decode('utf-8'))
- raise
- else:
- # Fast-forward merge was successful
- if not re.match('Already up-to-date.', merge_output) or verbose:
- if not printed_path:
- self.Print('_____ %s at %s' % (self.relpath, revision),
- timestamp=False)
- printed_path = True
- self.Print(merge_output.strip())
- if not verbose:
- # Make the output a little prettier. It's nice to have
- # some whitespace between projects when syncing.
- self.Print('')
- if file_list is not None:
- file_list.extend(
- [os.path.join(self.checkout_path, f) for f in rebase_files])
- # If the rebase generated a conflict, abort and ask user to fix
- if self._IsRebasing():
- raise gclient_utils.Error(
- '\n____ %s at %s\n'
- '\nConflict while rebasing this branch.\n'
- 'Fix the conflict and run gclient again.\n'
- 'See man git-rebase for details.\n' % (self.relpath, revision))
- # If --reset and --delete_unversioned_trees are specified, remove any
- # untracked directories.
- if options.reset and options.delete_unversioned_trees:
- # GIT.CaptureStatus() uses 'dit diff' to compare to a specific SHA1
- # (the merge-base by default), so doesn't include untracked files.
- # So we use 'git ls-files --directory --others --exclude-standard'
- # here directly.
- paths = scm.GIT.Capture([
- '-c', 'core.quotePath=false', 'ls-files', '--directory',
- '--others', '--exclude-standard'
- ], self.checkout_path)
- for path in (p for p in paths.splitlines() if p.endswith('/')):
- full_path = os.path.join(self.checkout_path, path)
- if not os.path.islink(full_path):
- self.Print('_____ removing unversioned directory %s' % path)
- gclient_utils.rmtree(full_path)
- if not current_revision:
- current_revision = self._Capture(['rev-parse', '--verify', 'HEAD'])
- if verbose:
- self.Print(f'Checked out revision {current_revision}',
- timestamp=False)
- return current_revision
- def revert(self, options, _args, file_list):
- """Reverts local modifications.
- All reverted files will be appended to file_list.
- """
- if not os.path.isdir(self.checkout_path):
- # revert won't work if the directory doesn't exist. It needs to
- # checkout instead.
- self.Print('_____ %s is missing, syncing instead' % self.relpath)
- # Don't reuse the args.
- return self.update(options, [], file_list)
- default_rev = "refs/heads/main"
- if options.upstream:
- if self._GetCurrentBranch():
- upstream_branch = scm.GIT.GetUpstreamBranch(self.checkout_path)
- default_rev = upstream_branch or default_rev
- _, deps_revision = gclient_utils.SplitUrlRevision(self.url)
- if not deps_revision:
- deps_revision = default_rev
- if deps_revision.startswith('refs/heads/'):
- deps_revision = deps_revision.replace('refs/heads/',
- self.remote + '/')
- try:
- deps_revision = self.GetUsableRev(deps_revision, options)
- except NoUsableRevError as e:
- # If the DEPS entry's url and hash changed, try to update the
- # origin. See also http://crbug.com/520067.
- logging.warning(
- "Couldn't find usable revision, will retrying to update instead: %s",
- str(e))
- return self.update(options, [], file_list)
- if file_list is not None:
- files = self._GetDiffFilenames(deps_revision)
- self._Scrub(deps_revision, options)
- self._Run(['clean', '-f', '-d'], options)
- if file_list is not None:
- file_list.extend(
- [os.path.join(self.checkout_path, f) for f in files])
- def revinfo(self, _options, _args, _file_list):
- """Returns revision"""
- return self._Capture(['rev-parse', 'HEAD'])
- def runhooks(self, options, args, file_list):
- self.status(options, args, file_list)
- def status(self, options, _args, file_list):
- """Display status information."""
- if not os.path.isdir(self.checkout_path):
- self.Print('________ couldn\'t run status in %s:\n'
- 'The directory does not exist.' % self.checkout_path)
- else:
- merge_base = []
- if self.url:
- _, base_rev = gclient_utils.SplitUrlRevision(self.url)
- if base_rev:
- if base_rev.startswith('refs/'):
- base_rev = self._ref_to_remote_ref(base_rev)
- merge_base = [base_rev]
- self._Run(['-c', 'core.quotePath=false', 'diff', '--name-status'] +
- merge_base,
- options,
- always_show_header=options.verbose)
- if file_list is not None:
- files = self._GetDiffFilenames(
- merge_base[0] if merge_base else None)
- file_list.extend(
- [os.path.join(self.checkout_path, f) for f in files])
- def GetUsableRev(self, rev, options):
- """Finds a useful revision for this repository."""
- sha1 = None
- if not os.path.isdir(self.checkout_path):
- raise NoUsableRevError(
- 'This is not a git repo, so we cannot get a usable rev.')
- if scm.GIT.IsValidRevision(cwd=self.checkout_path, rev=rev):
- sha1 = rev
- else:
- # May exist in origin, but we don't have it yet, so fetch and look
- # again.
- self._Fetch(options)
- if scm.GIT.IsValidRevision(cwd=self.checkout_path, rev=rev):
- sha1 = rev
- if not sha1:
- raise NoUsableRevError(
- 'Hash %s does not appear to be a valid hash in this repo.' %
- rev)
- return sha1
- def GetGitBackupDirPath(self):
- """Returns the path where the .git folder for the current project can be
- staged/restored. Use case: subproject moved from DEPS <-> outer project."""
- return os.path.join(self._root_dir,
- 'old_' + self.relpath.replace(os.sep, '_')) + '.git'
- def _GetMirror(self, url, options, revision=None, revision_ref=None):
- """Get a git_cache.Mirror object for the argument url."""
- if not self.cache_dir:
- return None
- mirror_kwargs = {
- 'print_func': self.filter,
- 'refs': [],
- 'commits': [],
- }
- if hasattr(options, 'with_branch_heads') and options.with_branch_heads:
- mirror_kwargs['refs'].append('refs/branch-heads/*')
- elif revision_ref and revision_ref.startswith('refs/branch-heads/'):
- mirror_kwargs['refs'].append(revision_ref)
- if hasattr(options, 'with_tags') and options.with_tags:
- mirror_kwargs['refs'].append('refs/tags/*')
- elif revision_ref and revision_ref.startswith('refs/tags/'):
- mirror_kwargs['refs'].append(revision_ref)
- if revision and not revision.startswith('refs/'):
- mirror_kwargs['commits'].append(revision)
- return git_cache.Mirror(url, **mirror_kwargs)
- def _UpdateMirrorIfNotContains(self, mirror, options, rev_type, revision):
- """Update a git mirror unless it already contains a hash revision.
- This raises an error if a SHA-1 revision isn't present even after
- fetching from the remote.
- """
- # 'hash' is overloaded and can refer to a SHA-1 hash or refs/changes/*.
- is_sha = gclient_utils.IsFullGitSha(revision)
- if rev_type == 'hash' and is_sha and mirror.contains_revision(revision):
- if options.verbose:
- self.Print('skipping mirror update, it has rev=%s already' %
- revision,
- timestamp=False)
- return
- if getattr(options, 'shallow', False):
- depth = 10000
- else:
- depth = None
- mirror.populate(verbose=False,
- bootstrap=not getattr(options, 'no_bootstrap', False),
- depth=depth,
- lock_timeout=getattr(options, 'lock_timeout', 0))
- # Make sure we've actually fetched the revision we want, but only if it
- # was specified as an explicit commit hash.
- if rev_type == 'hash' and is_sha and not mirror.contains_revision(
- revision):
- raise gclient_utils.Error(f'Failed to fetch {revision}.')
- def _Clone(self, revision, url, options):
- """Clone a git repository from the given URL.
- Once we've cloned the repo, we checkout a working branch if the
- specified revision is a branch head. If it is a tag or a specific
- commit, then we leave HEAD detached as it makes future updates simpler
- -- in this case the user should first create a new branch or switch to
- an existing branch before making changes in the repo."""
- if self.print_outbuf:
- print_stdout = True
- filter_fn = None
- else:
- print_stdout = False
- filter_fn = self.filter
- if not options.verbose:
- # git clone doesn't seem to insert a newline properly before
- # printing to stdout
- self.Print('')
- # If the parent directory does not exist, Git clone on Windows will not
- # create it, so we need to do it manually.
- parent_dir = os.path.dirname(self.checkout_path)
- gclient_utils.safe_makedirs(parent_dir)
- if hasattr(options, 'no_history') and options.no_history:
- self._Run(['init', self.checkout_path], options, cwd=self._root_dir)
- self._Run(['remote', 'add', 'origin', url], options)
- revision = self._AutoFetchRef(options, revision, depth=1)
- remote_ref = scm.GIT.RefToRemoteRef(revision, self.remote)
- self._Checkout(options, ''.join(remote_ref or revision), quiet=True)
- else:
- cfg = gclient_utils.DefaultIndexPackConfig(url)
- clone_cmd = cfg + ['clone', '--no-checkout', '--progress']
- if self.cache_dir:
- clone_cmd.append('--shared')
- if options.verbose:
- clone_cmd.append('--verbose')
- clone_cmd.append(url)
- tmp_dir = tempfile.mkdtemp(prefix='_gclient_%s_' %
- os.path.basename(self.checkout_path),
- dir=parent_dir)
- clone_cmd.append(tmp_dir)
- try:
- self._Run(clone_cmd,
- options,
- cwd=self._root_dir,
- retry=True,
- print_stdout=print_stdout,
- filter_fn=filter_fn)
- logging.debug(
- 'Cloned into temporary dir, moving to checkout_path')
- gclient_utils.safe_makedirs(self.checkout_path)
- gclient_utils.safe_rename(
- os.path.join(tmp_dir, '.git'),
- os.path.join(self.checkout_path, '.git'))
- except:
- traceback.print_exc(file=self.out_fh)
- raise
- finally:
- if os.listdir(tmp_dir):
- self.Print('_____ removing non-empty tmp dir %s' % tmp_dir)
- gclient_utils.rmtree(tmp_dir)
- self._SetFetchConfig(options)
- self._Fetch(options, prune=options.force)
- revision = self._AutoFetchRef(options, revision)
- remote_ref = scm.GIT.RefToRemoteRef(revision, self.remote)
- self._Checkout(options, ''.join(remote_ref or revision), quiet=True)
- if self._GetCurrentBranch() is None:
- # Squelch git's very verbose detached HEAD warning and use our own
- self.Print((
- 'Checked out %s to a detached HEAD. Before making any commits\n'
- 'in this repo, you should use \'git checkout <branch>\' to switch \n'
- 'to an existing branch or use \'git checkout %s -b <branch>\' to\n'
- 'create a new branch for your work.') % (revision, self.remote))
- return revision
- def _AskForData(self, prompt, options):
- if options.jobs > 1:
- self.Print(prompt)
- raise gclient_utils.Error("Background task requires input. Rerun "
- "gclient with --jobs=1 so that\n"
- "interaction is possible.")
- return gclient_utils.AskForData(prompt)
- def _AttemptRebase(self,
- upstream,
- files,
- options,
- newbase=None,
- branch=None,
- printed_path=False,
- merge=False):
- """Attempt to rebase onto either upstream or, if specified, newbase."""
- if files is not None:
- files.extend(self._GetDiffFilenames(upstream))
- revision = upstream
- if newbase:
- revision = newbase
- action = 'merge' if merge else 'rebase'
- if not printed_path:
- self.Print('_____ %s : Attempting %s onto %s...' %
- (self.relpath, action, revision))
- printed_path = True
- else:
- self.Print('Attempting %s onto %s...' % (action, revision))
- if merge:
- merge_output = self._Capture(['merge', revision])
- if options.verbose:
- self.Print(merge_output)
- return
- # Build the rebase command here using the args
- # git rebase [options] [--onto <newbase>] <upstream> [<branch>]
- rebase_cmd = ['rebase']
- if options.verbose:
- rebase_cmd.append('--verbose')
- if newbase:
- rebase_cmd.extend(['--onto', newbase])
- rebase_cmd.append(upstream)
- if branch:
- rebase_cmd.append(branch)
- try:
- rebase_output = scm.GIT.Capture(rebase_cmd, cwd=self.checkout_path)
- except subprocess2.CalledProcessError as e:
- if (re.match(
- br'cannot rebase: you have unstaged changes', e.stderr
- ) or re.match(
- br'cannot rebase: your index contains uncommitted changes',
- e.stderr)):
- while True:
- rebase_action = self._AskForData(
- 'Cannot rebase because of unstaged changes.\n'
- '\'git reset --hard HEAD\' ?\n'
- 'WARNING: destroys any uncommitted work in your current branch!'
- ' (y)es / (q)uit / (s)how : ', options)
- if re.match(r'yes|y', rebase_action, re.I):
- self._Scrub('HEAD', options)
- # Should this be recursive?
- rebase_output = scm.GIT.Capture(rebase_cmd,
- cwd=self.checkout_path)
- break
- if re.match(r'quit|q', rebase_action, re.I):
- raise gclient_utils.Error(
- "Please merge or rebase manually\n"
- "cd %s && git " % self.checkout_path +
- "%s" % ' '.join(rebase_cmd))
- if re.match(r'show|s', rebase_action, re.I):
- self.Print('%s' % e.stderr.decode('utf-8').strip())
- continue
- gclient_utils.Error("Input not recognized")
- continue
- elif re.search(br'^CONFLICT', e.stdout, re.M):
- raise gclient_utils.Error(
- "Conflict while rebasing this branch.\n"
- "Fix the conflict and run gclient again.\n"
- "See 'man git-rebase' for details.\n")
- else:
- self.Print(e.stdout.decode('utf-8').strip())
- self.Print('Rebase produced error output:\n%s' %
- e.stderr.decode('utf-8').strip())
- raise gclient_utils.Error(
- "Unrecognized error, please merge or rebase "
- "manually.\ncd %s && git " % self.checkout_path +
- "%s" % ' '.join(rebase_cmd))
- self.Print(rebase_output.strip())
- if not options.verbose:
- # Make the output a little prettier. It's nice to have some
- # whitespace between projects when syncing.
- self.Print('')
- def _EnsureValidHeadObjectOrCheckout(self, revision, options, url):
- # Special case handling if all 3 conditions are met:
- # * the mirros have recently changed, but deps destination remains same,
- # * the git histories of mirrors are conflicting. * git cache is used
- # This manifests itself in current checkout having invalid HEAD commit
- # on most git operations. Since git cache is used, just deleted the .git
- # folder, and re-create it by cloning.
- try:
- return self._Capture(['rev-list', '-n', '1', 'HEAD'])
- except subprocess2.CalledProcessError as e:
- if (b'fatal: bad object HEAD' in e.stderr and self.cache_dir
- and self.cache_dir in url):
- self.Print(
- ('Likely due to DEPS change with git cache_dir, '
- 'the current commit points to no longer existing object.\n'
- '%s' % e))
- self._DeleteOrMove(options.force)
- return self._Clone(revision, url, options)
- raise
- def _IsRebasing(self):
- # Check for any of REBASE-i/REBASE-m/REBASE/AM. Unfortunately git
- # doesn't have a plumbing command to determine whether a rebase is in
- # progress, so for now emualate (more-or-less) git-rebase.sh /
- # git-completion.bash
- g = os.path.join(self.checkout_path, '.git')
- return (os.path.isdir(os.path.join(g, "rebase-merge"))
- or os.path.isdir(os.path.join(g, "rebase-apply")))
- def _CheckClean(self, revision):
- lockfile = os.path.join(self.checkout_path, ".git", "index.lock")
- if os.path.exists(lockfile):
- raise gclient_utils.Error(
- '\n____ %s at %s\n'
- '\tYour repo is locked, possibly due to a concurrent git process.\n'
- '\tIf no git executable is running, then clean up %r and try again.\n'
- % (self.relpath, revision, lockfile))
- # Ensure that the tree is clean.
- if scm.GIT.Capture([
- 'status', '--porcelain', '--untracked-files=no',
- '--ignore-submodules'
- ],
- cwd=self.checkout_path):
- raise gclient_utils.Error(
- '\n____ %s at %s\n'
- '\tYou have uncommitted changes.\n'
- '\tcd into %s, run git status to see changes,\n'
- '\tand commit, stash, or reset.\n' %
- (self.relpath, revision, self.relpath))
- def _CheckDetachedHead(self, revision, _options):
- # HEAD is detached. Make sure it is safe to move away from (i.e., it is
- # reference by a commit). If not, error out -- most likely a rebase is
- # in progress, try to detect so we can give a better error.
- try:
- scm.GIT.Capture(['name-rev', '--no-undefined', 'HEAD'],
- cwd=self.checkout_path)
- except subprocess2.CalledProcessError:
- # Commit is not contained by any rev. See if the user is rebasing:
- if self._IsRebasing():
- # Punt to the user
- raise gclient_utils.Error(
- '\n____ %s at %s\n'
- '\tAlready in a conflict, i.e. (no branch).\n'
- '\tFix the conflict and run gclient again.\n'
- '\tOr to abort run:\n\t\tgit-rebase --abort\n'
- '\tSee man git-rebase for details.\n' %
- (self.relpath, revision))
- # Let's just save off the commit so we can proceed.
- name = ('saved-by-gclient-' +
- self._Capture(['rev-parse', '--short', 'HEAD']))
- self._Capture(['branch', '-f', name])
- self.Print(
- '_____ found an unreferenced commit and saved it as \'%s\'' %
- name)
- def _GetCurrentBranch(self):
- # Returns name of current branch or None for detached HEAD
- branch = self._Capture(['rev-parse', '--abbrev-ref=strict', 'HEAD'])
- if branch == 'HEAD':
- return None
- return branch
- def _Capture(self, args, **kwargs):
- set_git_dir = 'cwd' not in kwargs
- kwargs.setdefault('cwd', self.checkout_path)
- kwargs.setdefault('stderr', subprocess2.PIPE)
- strip = kwargs.pop('strip', True)
- env = scm.GIT.ApplyEnvVars(kwargs)
- # If an explicit cwd isn't set, then default to the .git/ subdir so we
- # get stricter behavior. This can be useful in cases of slight
- # corruption -- we don't accidentally go corrupting parent git checks
- # too. See https://crbug.com/1000825 for an example.
- if set_git_dir:
- env.setdefault(
- 'GIT_DIR',
- os.path.abspath(os.path.join(self.checkout_path, '.git')))
- kwargs.setdefault('env', env)
- ret = git_common.run(*args, **kwargs)
- if strip:
- ret = ret.strip()
- self.Print('Finished running: %s %s' % ('git', ' '.join(args)))
- return ret
- def _Checkout(self, options, ref, force=False, quiet=None):
- """Performs a 'git-checkout' operation.
- Args:
- options: The configured option set
- ref: (str) The branch/commit to checkout
- quiet: (bool/None) Whether or not the checkout should pass '--quiet'; if
- 'None', the behavior is inferred from 'options.verbose'.
- Returns: (str) The output of the checkout operation
- """
- if quiet is None:
- quiet = (not options.verbose)
- checkout_args = ['checkout']
- if force:
- checkout_args.append('--force')
- if quiet:
- checkout_args.append('--quiet')
- checkout_args.append(ref)
- return self._Capture(checkout_args)
- def _Fetch(self,
- options,
- remote=None,
- prune=False,
- quiet=False,
- refspec=None,
- depth=None):
- cfg = gclient_utils.DefaultIndexPackConfig(self.url)
- # When updating, the ref is modified to be a remote ref .
- # (e.g. refs/heads/NAME becomes refs/remotes/REMOTE/NAME).
- # Try to reverse that mapping.
- original_ref = scm.GIT.RemoteRefToRef(refspec, self.remote)
- if original_ref:
- refspec = original_ref + ':' + refspec
- # When a mirror is configured, it only fetches
- # refs/{heads,branch-heads,tags}/*.
- # If asked to fetch other refs, we must fetch those directly from
- # the repository, and not from the mirror.
- if not original_ref.startswith(
- ('refs/heads/', 'refs/branch-heads/', 'refs/tags/')):
- remote, _ = gclient_utils.SplitUrlRevision(self.url)
- fetch_cmd = cfg + [
- 'fetch',
- remote or self.remote,
- ]
- if refspec:
- fetch_cmd.append(refspec)
- if prune:
- fetch_cmd.append('--prune')
- if options.verbose:
- fetch_cmd.append('--verbose')
- if not hasattr(options, 'with_tags') or not options.with_tags:
- fetch_cmd.append('--no-tags')
- elif quiet:
- fetch_cmd.append('--quiet')
- if depth:
- fetch_cmd.append('--depth=' + str(depth))
- self._Run(fetch_cmd, options, show_header=options.verbose, retry=True)
- def _SetFetchConfig(self, options):
- """Adds, and optionally fetches, "branch-heads" and "tags" refspecs
- if requested."""
- if options.force or options.reset:
- try:
- scm.GIT.SetConfig(self.checkout_path,
- f'remote.{self.remote}.fetch',
- modify_all=True)
- scm.GIT.SetConfig(
- self.checkout_path, f'remote.{self.remote}.fetch',
- f'+refs/heads/*:refs/remotes/{self.remote}/*')
- except subprocess2.CalledProcessError as e:
- # If exit code was 5, it means we attempted to unset a config
- # that didn't exist. Ignore it.
- if e.returncode != 5:
- raise
- if hasattr(options, 'with_branch_heads') and options.with_branch_heads:
- scm.GIT.SetConfig(
- self.checkout_path,
- f'remote.{self.remote}.fetch',
- '+refs/branch-heads/*:refs/remotes/branch-heads/*',
- value_pattern='^\\+refs/branch-heads/\\*:.*$',
- modify_all=True)
- if hasattr(options, 'with_tags') and options.with_tags:
- scm.GIT.SetConfig(self.checkout_path,
- f'remote.{self.remote}.fetch',
- '+refs/tags/*:refs/tags/*',
- value_pattern='^\\+refs/tags/\\*:.*$',
- modify_all=True)
- def _AutoFetchRef(self, options, revision, depth=None):
- """Attempts to fetch |revision| if not available in local repo.
- Returns possibly updated revision."""
- if not scm.GIT.IsValidRevision(self.checkout_path, revision):
- self._Fetch(options, refspec=revision, depth=depth)
- revision = self._Capture(['rev-parse', 'FETCH_HEAD'])
- return revision
- def _Run(self, args, options, **kwargs):
- # Disable 'unused options' warning | pylint: disable=unused-argument
- kwargs.setdefault('cwd', self.checkout_path)
- kwargs.setdefault('filter_fn', self.filter)
- kwargs.setdefault('show_header', True)
- env = scm.GIT.ApplyEnvVars(kwargs)
- cmd = ['git'] + args
- gclient_utils.CheckCallAndFilter(cmd, env=env, **kwargs)
- class CipdPackage(object):
- """A representation of a single CIPD package."""
- def __init__(self, name, version, authority_for_subdir):
- self._authority_for_subdir = authority_for_subdir
- self._name = name
- self._version = version
- @property
- def authority_for_subdir(self):
- """Whether this package has authority to act on behalf of its subdir.
- Some operations should only be performed once per subdirectory. A package
- that has authority for its subdirectory is the only package that should
- perform such operations.
- Returns:
- bool; whether this package has subdir authority.
- """
- return self._authority_for_subdir
- @property
- def name(self):
- return self._name
- @property
- def version(self):
- return self._version
- class CipdRoot(object):
- """A representation of a single CIPD root."""
- def __init__(self, root_dir, service_url, log_level=None):
- self._all_packages = set()
- self._mutator_lock = threading.Lock()
- self._packages_by_subdir = collections.defaultdict(list)
- self._root_dir = root_dir
- self._service_url = service_url
- self._resolved_packages = None
- self._log_level = log_level or 'error'
- def add_package(self, subdir, package, version):
- """Adds a package to this CIPD root.
- As far as clients are concerned, this grants both root and subdir authority
- to packages arbitrarily. (The implementation grants root authority to the
- first package added and subdir authority to the first package added for that
- subdir, but clients should not depend on or expect that behavior.)
- Args:
- subdir: str; relative path to where the package should be installed from
- the cipd root directory.
- package: str; the cipd package name.
- version: str; the cipd package version.
- Returns:
- CipdPackage; the package that was created and added to this root.
- """
- with self._mutator_lock:
- cipd_package = CipdPackage(package, version,
- not self._packages_by_subdir[subdir])
- self._all_packages.add(cipd_package)
- self._packages_by_subdir[subdir].append(cipd_package)
- return cipd_package
- def packages(self, subdir):
- """Get the list of configured packages for the given subdir."""
- return list(self._packages_by_subdir[subdir])
- def resolved_packages(self):
- if not self._resolved_packages:
- self._resolved_packages = self.ensure_file_resolve()
- return self._resolved_packages
- def clobber(self):
- """Remove the .cipd directory.
- This is useful for forcing ensure to redownload and reinitialize all
- packages.
- """
- with self._mutator_lock:
- cipd_cache_dir = os.path.join(self.root_dir, '.cipd')
- try:
- gclient_utils.rmtree(os.path.join(cipd_cache_dir))
- except OSError:
- if os.path.exists(cipd_cache_dir):
- raise
- def expand_package_name(self, package_name_string, **kwargs):
- """Run `cipd expand-package-name`.
- CIPD package names can be declared with placeholder variables
- such as '${platform}', this cmd will return the package name
- with the variables resolved. The resolution is based on the host
- the command is executing on.
- """
- kwargs.setdefault('stderr', subprocess2.PIPE)
- cmd = ['cipd', 'expand-package-name', package_name_string]
- ret = subprocess2.check_output(cmd, **kwargs).decode('utf-8')
- return ret.strip()
- @contextlib.contextmanager
- def _create_ensure_file(self):
- try:
- contents = '$ParanoidMode CheckPresence\n'
- # TODO(crbug/1329641): Remove once cipd packages have been updated
- # to always be created in copy mode.
- contents += '$OverrideInstallMode copy\n\n'
- for subdir, packages in sorted(self._packages_by_subdir.items()):
- contents += '@Subdir %s\n' % subdir
- for package in sorted(packages, key=lambda p: p.name):
- contents += '%s %s\n' % (package.name, package.version)
- contents += '\n'
- ensure_file = None
- with tempfile.NamedTemporaryFile(suffix='.ensure',
- delete=False,
- mode='wb') as ensure_file:
- ensure_file.write(contents.encode('utf-8', 'replace'))
- yield ensure_file.name
- finally:
- if ensure_file is not None and os.path.exists(ensure_file.name):
- os.remove(ensure_file.name)
- def ensure(self):
- """Run `cipd ensure`."""
- with self._mutator_lock:
- with self._create_ensure_file() as ensure_file:
- cmd = [
- 'cipd',
- 'ensure',
- '-log-level',
- self._log_level,
- '-root',
- self.root_dir,
- '-ensure-file',
- ensure_file,
- ]
- gclient_utils.CheckCallAndFilter(cmd,
- print_stdout=True,
- show_header=True)
- @contextlib.contextmanager
- def _create_ensure_file_for_resolve(self):
- try:
- contents = '$ResolvedVersions %s\n' % os.devnull
- for subdir, packages in sorted(self._packages_by_subdir.items()):
- contents += '@Subdir %s\n' % subdir
- for package in sorted(packages, key=lambda p: p.name):
- contents += '%s %s\n' % (package.name, package.version)
- contents += '\n'
- ensure_file = None
- with tempfile.NamedTemporaryFile(suffix='.ensure',
- delete=False,
- mode='wb') as ensure_file:
- ensure_file.write(contents.encode('utf-8', 'replace'))
- yield ensure_file.name
- finally:
- if ensure_file is not None and os.path.exists(ensure_file.name):
- os.remove(ensure_file.name)
- def _create_resolved_file(self):
- return tempfile.NamedTemporaryFile(suffix='.resolved',
- delete=False,
- mode='wb')
- def ensure_file_resolve(self):
- """Run `cipd ensure-file-resolve`."""
- with self._mutator_lock:
- with self._create_resolved_file() as output_file:
- with self._create_ensure_file_for_resolve() as ensure_file:
- cmd = [
- 'cipd',
- 'ensure-file-resolve',
- '-log-level',
- self._log_level,
- '-ensure-file',
- ensure_file,
- '-json-output',
- output_file.name,
- ]
- gclient_utils.CheckCallAndFilter(cmd,
- print_stdout=False,
- show_header=False)
- with open(output_file.name) as f:
- output_json = json.load(f)
- return output_json.get('result', {})
- def run(self, command):
- if command == 'update':
- self.ensure()
- elif command == 'revert':
- self.clobber()
- self.ensure()
- def created_package(self, package):
- """Checks whether this root created the given package.
- Args:
- package: CipdPackage; the package to check.
- Returns:
- bool; whether this root created the given package.
- """
- return package in self._all_packages
- @property
- def root_dir(self):
- return self._root_dir
- @property
- def service_url(self):
- return self._service_url
- class CipdWrapper(SCMWrapper):
- """Wrapper for CIPD.
- Currently only supports chrome-infra-packages.appspot.com.
- """
- name = 'cipd'
- def __init__(self,
- url=None,
- root_dir=None,
- relpath=None,
- out_fh=None,
- out_cb=None,
- root=None,
- package=None):
- super(CipdWrapper, self).__init__(url=url,
- root_dir=root_dir,
- relpath=relpath,
- out_fh=out_fh,
- out_cb=out_cb)
- assert root.created_package(package)
- self._package = package
- self._root = root
- #override
- def GetCacheMirror(self):
- return None
- #override
- def GetActualRemoteURL(self, options):
- return self._root.service_url
- #override
- def DoesRemoteURLMatch(self, options):
- del options
- return True
- def revert(self, options, args, file_list):
- """Does nothing.
- CIPD packages should be reverted at the root by running
- `CipdRoot.run('revert')`.
- """
- def diff(self, options, args, file_list):
- """CIPD has no notion of diffing."""
- def pack(self, options, args, file_list):
- """CIPD has no notion of diffing."""
- def revinfo(self, options, args, file_list):
- """Grab the instance ID."""
- try:
- tmpdir = tempfile.mkdtemp()
- # Attempt to get instance_id from the root resolved cache.
- # Resolved cache will not match on any CIPD packages with
- # variables such as ${platform}, they will fall back to
- # the slower method below.
- resolved = self._root.resolved_packages()
- if resolved:
- # CIPD uses POSIX separators across all platforms, so
- # replace any Windows separators.
- path_split = self.relpath.replace(os.sep, "/").split(":")
- if len(path_split) > 1:
- src_path, package = path_split
- if src_path in resolved:
- for resolved_package in resolved[src_path]:
- if package == resolved_package.get(
- 'pin', {}).get('package'):
- return resolved_package.get(
- 'pin', {}).get('instance_id')
- describe_json_path = os.path.join(tmpdir, 'describe.json')
- cmd = [
- 'cipd', 'describe', self._package.name, '-log-level', 'error',
- '-version', self._package.version, '-json-output',
- describe_json_path
- ]
- gclient_utils.CheckCallAndFilter(cmd)
- with open(describe_json_path) as f:
- describe_json = json.load(f)
- return describe_json.get('result', {}).get('pin',
- {}).get('instance_id')
- finally:
- gclient_utils.rmtree(tmpdir)
- def status(self, options, args, file_list):
- pass
- def update(self, options, args, file_list):
- """Does nothing.
- CIPD packages should be updated at the root by running
- `CipdRoot.run('update')`.
- """
- class GcsRoot(object):
- """Root to keep track of all GCS objects, per checkout"""
- def __init__(self, root_dir):
- self._mutator_lock = threading.Lock()
- self._root_dir = root_dir
- # Populated when the DEPS file is parsed
- # The objects here have not yet been downloaded and written into
- # the .gcs_entries file
- self._parsed_objects = {}
- # .gcs_entries keeps track of which GCS deps have already been installed
- # Maps checkout_name -> {GCS dep path -> [object_name]}
- # This file is in the same directory as .gclient
- self._gcs_entries_file = os.path.join(self._root_dir, '.gcs_entries')
- # Contents of the .gcs_entries file
- self._gcs_entries = self.read_gcs_entries()
- @property
- def root_dir(self):
- return self._root_dir
- def add_object(self, checkout_name, dep_path, object_name):
- """Records the object in the _parsed_objects variable
- This does not actually download the object"""
- with self._mutator_lock:
- if checkout_name not in self._parsed_objects:
- self._parsed_objects[checkout_name] = {}
- if dep_path not in self._parsed_objects[checkout_name]:
- self._parsed_objects[checkout_name][dep_path] = [object_name]
- else:
- self._parsed_objects[checkout_name][dep_path].append(
- object_name)
- def read_gcs_entries(self):
- """Reads .gcs_entries file and loads the content into _gcs_entries"""
- if not os.path.exists(self._gcs_entries_file):
- return {}
- with open(self._gcs_entries_file, 'r') as f:
- content = f.read().rstrip()
- if content:
- return json.loads(content)
- return {}
- def resolve_objects(self, checkout_name):
- """Updates .gcs_entries with objects in _parsed_objects
- This should only be called after the objects have been downloaded
- and extracted."""
- with self._mutator_lock:
- object_dict = self._parsed_objects.get(checkout_name)
- if not object_dict:
- return
- self._gcs_entries[checkout_name] = object_dict
- with open(self._gcs_entries_file, 'w') as f:
- f.write(json.dumps(self._gcs_entries, indent=2))
- self._parsed_objects[checkout_name] = {}
- def clobber_deps_with_updated_objects(self, checkout_name):
- """Clobber the path if an object or GCS dependency is removed/added
- This must be called before the GCS dependencies are
- downloaded and extracted."""
- with self._mutator_lock:
- parsed_object_dict = self._parsed_objects.get(checkout_name, {})
- parsed_paths = set(parsed_object_dict.keys())
- resolved_object_dict = self._gcs_entries.get(checkout_name, {})
- resolved_paths = set(resolved_object_dict.keys())
- # If any GCS deps are added or removed entirely, clobber that path
- intersected_paths = parsed_paths.intersection(resolved_paths)
- # If any objects within a GCS dep are added/removed, clobber its
- # extracted contents and relevant gcs dotfiles
- for path in intersected_paths:
- resolved_objects = resolved_object_dict[path]
- parsed_objects = parsed_object_dict[path]
- full_path = os.path.join(self.root_dir, path)
- if (len(resolved_objects) != len(parsed_objects)
- and os.path.exists(full_path)):
- self.clobber_tar_content_names(full_path)
- self.clobber_hash_files(full_path)
- self.clobber_migration_files(full_path)
- def clobber_tar_content_names(self, entry_directory):
- """Delete paths written in .*_content_names files"""
- content_names_files = glob.glob(
- os.path.join(entry_directory, '.*_content_names'))
- for file in content_names_files:
- with open(file, 'r') as f:
- names = json.loads(f.read().strip())
- for name in names:
- name_path = os.path.join(entry_directory, name)
- if os.path.isdir(
- name_path) or not os.path.exists(name_path):
- continue
- os.remove(os.path.join(entry_directory, name))
- os.remove(file)
- def clobber_hash_files(self, entry_directory):
- files = glob.glob(os.path.join(entry_directory, '.*_hash'))
- for f in files:
- os.remove(f)
- def clobber_migration_files(self, entry_directory):
- files = glob.glob(os.path.join(entry_directory,
- '.*_is_first_class_gcs'))
- for f in files:
- os.remove(f)
- def clobber(self):
- """Remove all dep path gcs items and clear .gcs_entries"""
- for _, objects_dict in self._gcs_entries.items():
- for dep_path, _ in objects_dict.items():
- full_path = os.path.join(self.root_dir, dep_path)
- self.clobber_tar_content_names(full_path)
- self.clobber_hash_files(full_path)
- self.clobber_migration_files(full_path)
- if os.path.exists(self._gcs_entries_file):
- os.remove(self._gcs_entries_file)
- with self._mutator_lock:
- self._gcs_entries = {}
- class GcsWrapper(SCMWrapper):
- """Wrapper for GCS.
- Currently only supports content from Google Cloud Storage.
- """
- name = 'gcs'
- def __init__(self,
- url=None,
- root_dir=None,
- relpath=None,
- out_fh=None,
- out_cb=None):
- super(GcsWrapper, self).__init__(url=url,
- root_dir=root_dir,
- relpath=relpath,
- out_fh=out_fh,
- out_cb=out_cb)
- #override
- def GetCacheMirror(self):
- return None
- #override
- def GetActualRemoteURL(self, options):
- return None
- #override
- def DoesRemoteURLMatch(self, options):
- del options
- return True
- def revert(self, options, args, file_list):
- """Does nothing."""
- def diff(self, options, args, file_list):
- """GCS has no notion of diffing."""
- def pack(self, options, args, file_list):
- """GCS has no notion of diffing."""
- def revinfo(self, options, args, file_list):
- """Does nothing"""
- def status(self, options, args, file_list):
- pass
- def update(self, options, args, file_list):
- """Does nothing."""
- class CogWrapper(SCMWrapper):
- """Wrapper for Cog, all no-op."""
- name = 'cog'
- def __init__(self):
- super(CogWrapper, self).__init__()
- #override
- def GetCacheMirror(self):
- return None
- #override
- def GetActualRemoteURL(self, options):
- return None
- #override
- def GetSubmoduleDiff(self):
- return None
- #override
- def GetSubmoduleStateFromIndex(self):
- return None
- #override
- def DoesRemoteURLMatch(self, options):
- del options
- return True
- def revert(self, options, args, file_list):
- pass
- def diff(self, options, args, file_list):
- pass
- def pack(self, options, args, file_list):
- pass
- def revinfo(self, options, args, file_list):
- pass
- def status(self, options, args, file_list):
- pass
- def update(self, options, args, file_list):
- pass
|