Mandoline: Introduce a tab-level DevTools agent and persist state across navigations.
[chromium-blink-merge.git] / build / get_syzygy_binaries.py
blob1cab3fcf48dfd21f8c54e6492101f9d5df2e6300
1 #!/usr/bin/env python
2 # Copyright 2014 The Chromium Authors. All rights reserved.
3 # Use of this source code is governed by a BSD-style license that can be
4 # found in the LICENSE file.
6 """A utility script for downloading versioned Syzygy binaries."""
8 import hashlib
9 import errno
10 import json
11 import logging
12 import optparse
13 import os
14 import re
15 import shutil
16 import stat
17 import sys
18 import subprocess
19 import tempfile
20 import time
21 import zipfile
24 _LOGGER = logging.getLogger(os.path.basename(__file__))
26 # The relative path where official builds are archived in their GS bucket.
27 _SYZYGY_ARCHIVE_PATH = ('/builds/official/%(revision)s')
29 # A JSON file containing the state of the download directory. If this file and
30 # directory state do not agree, then the binaries will be downloaded and
31 # installed again.
32 _STATE = '.state'
34 # This matches an integer (an SVN revision number) or a SHA1 value (a GIT hash).
35 # The archive exclusively uses lowercase GIT hashes.
36 _REVISION_RE = re.compile('^(?:\d+|[a-f0-9]{40})$')
38 # This matches an MD5 hash.
39 _MD5_RE = re.compile('^[a-f0-9]{32}$')
41 # List of reources to be downloaded and installed. These are tuples with the
42 # following format:
43 # (basename, logging name, relative installation path, extraction filter)
44 _RESOURCES = [
45 ('benchmark.zip', 'benchmark', '', None),
46 ('binaries.zip', 'binaries', 'exe', None),
47 ('symbols.zip', 'symbols', 'exe',
48 lambda x: x.filename.endswith('.dll.pdb'))]
51 def _LoadState(output_dir):
52 """Loads the contents of the state file for a given |output_dir|, returning
53 None if it doesn't exist.
54 """
55 path = os.path.join(output_dir, _STATE)
56 if not os.path.exists(path):
57 _LOGGER.debug('No state file found.')
58 return None
59 with open(path, 'rb') as f:
60 _LOGGER.debug('Reading state file: %s', path)
61 try:
62 return json.load(f)
63 except ValueError:
64 _LOGGER.debug('Invalid state file.')
65 return None
68 def _SaveState(output_dir, state, dry_run=False):
69 """Saves the |state| dictionary to the given |output_dir| as a JSON file."""
70 path = os.path.join(output_dir, _STATE)
71 _LOGGER.debug('Writing state file: %s', path)
72 if dry_run:
73 return
74 with open(path, 'wb') as f:
75 f.write(json.dumps(state, sort_keys=True, indent=2))
78 def _Md5(path):
79 """Returns the MD5 hash of the file at |path|, which must exist."""
80 return hashlib.md5(open(path, 'rb').read()).hexdigest()
83 def _StateIsValid(state):
84 """Returns true if the given state structure is valid."""
85 if not isinstance(state, dict):
86 _LOGGER.debug('State must be a dict.')
87 return False
88 r = state.get('revision', None)
89 if not isinstance(r, basestring) or not _REVISION_RE.match(r):
90 _LOGGER.debug('State contains an invalid revision.')
91 return False
92 c = state.get('contents', None)
93 if not isinstance(c, dict):
94 _LOGGER.debug('State must contain a contents dict.')
95 return False
96 for (relpath, md5) in c.iteritems():
97 if not isinstance(relpath, basestring) or len(relpath) == 0:
98 _LOGGER.debug('State contents dict contains an invalid path.')
99 return False
100 if not isinstance(md5, basestring) or not _MD5_RE.match(md5):
101 _LOGGER.debug('State contents dict contains an invalid MD5 digest.')
102 return False
103 return True
106 def _BuildActualState(stored, revision, output_dir):
107 """Builds the actual state using the provided |stored| state as a template.
108 Only examines files listed in the stored state, causing the script to ignore
109 files that have been added to the directories locally. |stored| must be a
110 valid state dictionary.
112 contents = {}
113 state = { 'revision': revision, 'contents': contents }
114 for relpath, md5 in stored['contents'].iteritems():
115 abspath = os.path.abspath(os.path.join(output_dir, relpath))
116 if os.path.isfile(abspath):
117 m = _Md5(abspath)
118 contents[relpath] = m
120 return state
123 def _StatesAreConsistent(stored, actual):
124 """Validates whether two state dictionaries are consistent. Both must be valid
125 state dictionaries. Additional entries in |actual| are ignored.
127 if stored['revision'] != actual['revision']:
128 _LOGGER.debug('Mismatched revision number.')
129 return False
130 cont_stored = stored['contents']
131 cont_actual = actual['contents']
132 for relpath, md5 in cont_stored.iteritems():
133 if relpath not in cont_actual:
134 _LOGGER.debug('Missing content: %s', relpath)
135 return False
136 if md5 != cont_actual[relpath]:
137 _LOGGER.debug('Modified content: %s', relpath)
138 return False
139 return True
142 def _GetCurrentState(revision, output_dir):
143 """Loads the current state and checks to see if it is consistent. Returns
144 a tuple (state, bool). The returned state will always be valid, even if an
145 invalid state is present on disk.
147 stored = _LoadState(output_dir)
148 if not _StateIsValid(stored):
149 _LOGGER.debug('State is invalid.')
150 # Return a valid but empty state.
151 return ({'revision': '0', 'contents': {}}, False)
152 actual = _BuildActualState(stored, revision, output_dir)
153 # If the script has been modified consider the state invalid.
154 path = os.path.join(output_dir, _STATE)
155 if os.path.getmtime(__file__) > os.path.getmtime(path):
156 return (stored, False)
157 # Otherwise, explicitly validate the state.
158 if not _StatesAreConsistent(stored, actual):
159 return (stored, False)
160 return (stored, True)
163 def _DirIsEmpty(path):
164 """Returns true if the given directory is empty, false otherwise."""
165 for root, dirs, files in os.walk(path):
166 return not dirs and not files
169 def _RmTreeHandleReadOnly(func, path, exc):
170 """An error handling function for use with shutil.rmtree. This will
171 detect failures to remove read-only files, and will change their properties
172 prior to removing them. This is necessary on Windows as os.remove will return
173 an access error for read-only files, and git repos contain read-only
174 pack/index files.
176 excvalue = exc[1]
177 if func in (os.rmdir, os.remove) and excvalue.errno == errno.EACCES:
178 _LOGGER.debug('Removing read-only path: %s', path)
179 os.chmod(path, stat.S_IRWXU | stat.S_IRWXG | stat.S_IRWXO)
180 func(path)
181 else:
182 raise
185 def _RmTree(path):
186 """A wrapper of shutil.rmtree that handles read-only files."""
187 shutil.rmtree(path, ignore_errors=False, onerror=_RmTreeHandleReadOnly)
190 def _CleanState(output_dir, state, dry_run=False):
191 """Cleans up files/directories in |output_dir| that are referenced by
192 the given |state|. Raises an error if there are local changes. Returns a
193 dictionary of files that were deleted.
195 _LOGGER.debug('Deleting files from previous installation.')
196 deleted = {}
198 # Generate a list of files to delete, relative to |output_dir|.
199 contents = state['contents']
200 files = sorted(contents.keys())
202 # Try to delete the files. Keep track of directories to delete as well.
203 dirs = {}
204 for relpath in files:
205 fullpath = os.path.join(output_dir, relpath)
206 fulldir = os.path.dirname(fullpath)
207 dirs[fulldir] = True
208 if os.path.exists(fullpath):
209 # If somehow the file has become a directory complain about it.
210 if os.path.isdir(fullpath):
211 raise Exception('Directory exists where file expected: %s' % fullpath)
213 # Double check that the file doesn't have local changes. If it does
214 # then refuse to delete it.
215 if relpath in contents:
216 stored_md5 = contents[relpath]
217 actual_md5 = _Md5(fullpath)
218 if actual_md5 != stored_md5:
219 raise Exception('File has local changes: %s' % fullpath)
221 # The file is unchanged so it can safely be deleted.
222 _LOGGER.debug('Deleting file "%s".', fullpath)
223 deleted[relpath] = True
224 if not dry_run:
225 os.unlink(fullpath)
227 # Sort directories from longest name to shortest. This lets us remove empty
228 # directories from the most nested paths first.
229 dirs = sorted(dirs.keys(), key=lambda x: len(x), reverse=True)
230 for p in dirs:
231 if os.path.exists(p) and _DirIsEmpty(p):
232 _LOGGER.debug('Deleting empty directory "%s".', p)
233 if not dry_run:
234 _RmTree(p)
236 return deleted
239 def _FindGsUtil():
240 """Looks for depot_tools and returns the absolute path to gsutil.py."""
241 for path in os.environ['PATH'].split(os.pathsep):
242 path = os.path.abspath(path)
243 git_cl = os.path.join(path, 'git_cl.py')
244 gs_util = os.path.join(path, 'gsutil.py')
245 if os.path.exists(git_cl) and os.path.exists(gs_util):
246 return gs_util
247 return None
250 def _GsUtil(*cmd):
251 """Runs the given command in gsutil with exponential backoff and retries."""
252 gs_util = _FindGsUtil()
253 cmd = [sys.executable, gs_util] + list(cmd)
255 retries = 3
256 timeout = 4 # Seconds.
257 while True:
258 _LOGGER.debug('Running %s', cmd)
259 prog = subprocess.Popen(cmd, shell=False)
260 prog.communicate()
262 # Stop retrying on success.
263 if prog.returncode == 0:
264 return
266 # Raise a permanent failure if retries have been exhausted.
267 if retries == 0:
268 raise RuntimeError('Command "%s" returned %d.' % (cmd, prog.returncode))
270 _LOGGER.debug('Sleeping %d seconds and trying again.', timeout)
271 time.sleep(timeout)
272 retries -= 1
273 timeout *= 2
276 def _Download(resource):
277 """Downloads the given GS resource to a temporary file, returning its path."""
278 tmp = tempfile.mkstemp(suffix='syzygy_archive')
279 os.close(tmp[0])
280 url = 'gs://syzygy-archive' + resource
281 _GsUtil('cp', url, tmp[1])
282 return tmp[1]
285 def _InstallBinaries(options, deleted={}):
286 """Installs Syzygy binaries. This assumes that the output directory has
287 already been cleaned, as it will refuse to overwrite existing files."""
288 contents = {}
289 state = { 'revision': options.revision, 'contents': contents }
290 archive_path = _SYZYGY_ARCHIVE_PATH % { 'revision': options.revision }
291 if options.resources:
292 resources = [(resource, resource, '', None)
293 for resource in options.resources]
294 else:
295 resources = _RESOURCES
296 for (base, name, subdir, filt) in resources:
297 # Create the output directory if it doesn't exist.
298 fulldir = os.path.join(options.output_dir, subdir)
299 if os.path.isfile(fulldir):
300 raise Exception('File exists where a directory needs to be created: %s' %
301 fulldir)
302 if not os.path.exists(fulldir):
303 _LOGGER.debug('Creating directory: %s', fulldir)
304 if not options.dry_run:
305 os.makedirs(fulldir)
307 # Download and read the archive.
308 resource = archive_path + '/' + base
309 _LOGGER.debug('Retrieving %s archive at "%s".', name, resource)
310 path = _Download(resource)
312 _LOGGER.debug('Unzipping %s archive.', name)
313 with open(path, 'rb') as data:
314 archive = zipfile.ZipFile(data)
315 for entry in archive.infolist():
316 if not filt or filt(entry):
317 fullpath = os.path.normpath(os.path.join(fulldir, entry.filename))
318 relpath = os.path.relpath(fullpath, options.output_dir)
319 if os.path.exists(fullpath):
320 # If in a dry-run take into account the fact that the file *would*
321 # have been deleted.
322 if options.dry_run and relpath in deleted:
323 pass
324 else:
325 raise Exception('Path already exists: %s' % fullpath)
327 # Extract the file and update the state dictionary.
328 _LOGGER.debug('Extracting "%s".', fullpath)
329 if not options.dry_run:
330 archive.extract(entry.filename, fulldir)
331 md5 = _Md5(fullpath)
332 contents[relpath] = md5
333 if sys.platform == 'cygwin':
334 os.chmod(fullpath, os.stat(fullpath).st_mode | stat.S_IXUSR)
336 _LOGGER.debug('Removing temporary file "%s".', path)
337 os.remove(path)
339 return state
342 def _ParseCommandLine():
343 """Parses the command-line and returns an options structure."""
344 option_parser = optparse.OptionParser()
345 option_parser.add_option('--dry-run', action='store_true', default=False,
346 help='If true then will simply list actions that would be performed.')
347 option_parser.add_option('--force', action='store_true', default=False,
348 help='Force an installation even if the binaries are up to date.')
349 option_parser.add_option('--no-cleanup', action='store_true', default=False,
350 help='Allow installation on non-Windows platforms, and skip the forced '
351 'cleanup step.')
352 option_parser.add_option('--output-dir', type='string',
353 help='The path where the binaries will be replaced. Existing binaries '
354 'will only be overwritten if not up to date.')
355 option_parser.add_option('--overwrite', action='store_true', default=False,
356 help='If specified then the installation will happily delete and rewrite '
357 'the entire output directory, blasting any local changes.')
358 option_parser.add_option('--revision', type='string',
359 help='The SVN revision or GIT hash associated with the required version.')
360 option_parser.add_option('--revision-file', type='string',
361 help='A text file containing an SVN revision or GIT hash.')
362 option_parser.add_option('--resource', type='string', action='append',
363 dest='resources', help='A resource to be downloaded.')
364 option_parser.add_option('--verbose', dest='log_level', action='store_const',
365 default=logging.INFO, const=logging.DEBUG,
366 help='Enables verbose logging.')
367 option_parser.add_option('--quiet', dest='log_level', action='store_const',
368 default=logging.INFO, const=logging.ERROR,
369 help='Disables all output except for errors.')
370 options, args = option_parser.parse_args()
371 if args:
372 option_parser.error('Unexpected arguments: %s' % args)
373 if not options.output_dir:
374 option_parser.error('Must specify --output-dir.')
375 if not options.revision and not options.revision_file:
376 option_parser.error('Must specify one of --revision or --revision-file.')
377 if options.revision and options.revision_file:
378 option_parser.error('Must not specify both --revision and --revision-file.')
380 # Configure logging.
381 logging.basicConfig(level=options.log_level)
383 # If a revision file has been specified then read it.
384 if options.revision_file:
385 options.revision = open(options.revision_file, 'rb').read().strip()
386 _LOGGER.debug('Parsed revision "%s" from file "%s".',
387 options.revision, options.revision_file)
389 # Ensure that the specified SVN revision or GIT hash is valid.
390 if not _REVISION_RE.match(options.revision):
391 option_parser.error('Must specify a valid SVN or GIT revision.')
393 # This just makes output prettier to read.
394 options.output_dir = os.path.normpath(options.output_dir)
396 return options
399 def _RemoveOrphanedFiles(options):
400 """This is run on non-Windows systems to remove orphaned files that may have
401 been downloaded by a previous version of this script.
403 # Reconfigure logging to output info messages. This will allow inspection of
404 # cleanup status on non-Windows buildbots.
405 _LOGGER.setLevel(logging.INFO)
407 output_dir = os.path.abspath(options.output_dir)
409 # We only want to clean up the folder in 'src/third_party/syzygy', and we
410 # expect to be called with that as an output directory. This is an attempt to
411 # not start deleting random things if the script is run from an alternate
412 # location, or not called from the gclient hooks.
413 expected_syzygy_dir = os.path.abspath(os.path.join(
414 os.path.dirname(__file__), '..', 'third_party', 'syzygy'))
415 expected_output_dir = os.path.join(expected_syzygy_dir, 'binaries')
416 if expected_output_dir != output_dir:
417 _LOGGER.info('Unexpected output directory, skipping cleanup.')
418 return
420 if not os.path.isdir(expected_syzygy_dir):
421 _LOGGER.info('Output directory does not exist, skipping cleanup.')
422 return
424 def OnError(function, path, excinfo):
425 """Logs error encountered by shutil.rmtree."""
426 _LOGGER.error('Error when running %s(%s)', function, path, exc_info=excinfo)
428 _LOGGER.info('Removing orphaned files from %s', expected_syzygy_dir)
429 if not options.dry_run:
430 shutil.rmtree(expected_syzygy_dir, True, OnError)
433 def main():
434 options = _ParseCommandLine()
436 if options.dry_run:
437 _LOGGER.debug('Performing a dry-run.')
439 # We only care about Windows platforms, as the Syzygy binaries aren't used
440 # elsewhere. However, there was a short period of time where this script
441 # wasn't gated on OS types, and those OSes downloaded and installed binaries.
442 # This will cleanup orphaned files on those operating systems.
443 if sys.platform not in ('win32', 'cygwin'):
444 if options.no_cleanup:
445 _LOGGER.debug('Skipping usual cleanup for non-Windows platforms.')
446 else:
447 return _RemoveOrphanedFiles(options)
449 # Load the current installation state, and validate it against the
450 # requested installation.
451 state, is_consistent = _GetCurrentState(options.revision, options.output_dir)
453 # Decide whether or not an install is necessary.
454 if options.force:
455 _LOGGER.debug('Forcing reinstall of binaries.')
456 elif is_consistent:
457 # Avoid doing any work if the contents of the directory are consistent.
458 _LOGGER.debug('State unchanged, no reinstall necessary.')
459 return
461 # Under normal logging this is the only only message that will be reported.
462 _LOGGER.info('Installing revision %s Syzygy binaries.',
463 options.revision[0:12])
465 # Clean up the old state to begin with.
466 deleted = []
467 if options.overwrite:
468 if os.path.exists(options.output_dir):
469 # If overwrite was specified then take a heavy-handed approach.
470 _LOGGER.debug('Deleting entire installation directory.')
471 if not options.dry_run:
472 _RmTree(options.output_dir)
473 else:
474 # Otherwise only delete things that the previous installation put in place,
475 # and take care to preserve any local changes.
476 deleted = _CleanState(options.output_dir, state, options.dry_run)
478 # Install the new binaries. In a dry-run this will actually download the
479 # archives, but it won't write anything to disk.
480 state = _InstallBinaries(options, deleted)
482 # Build and save the state for the directory.
483 _SaveState(options.output_dir, state, options.dry_run)
486 if __name__ == '__main__':
487 main()