blob: 1cab3fcf48dfd21f8c54e6492101f9d5df2e6300 [file] [log] [blame]
[email protected]9c47c5fd2014-05-15 00:13:511#!/usr/bin/env python
2# Copyright 2014 The Chromium Authors. All rights reserved.
3# Use of this source code is governed by a BSD-style license that can be
4# found in the LICENSE file.
5
6"""A utility script for downloading versioned Syzygy binaries."""
7
[email protected]9c47c5fd2014-05-15 00:13:518import hashlib
[email protected]f1d116b2014-06-13 18:07:269import errno
[email protected]9c47c5fd2014-05-15 00:13:5110import json
11import logging
12import optparse
13import os
14import re
15import shutil
[email protected]f1d116b2014-06-13 18:07:2616import stat
17import sys
[email protected]9c47c5fd2014-05-15 00:13:5118import subprocess
chrisha3641d5f2015-07-14 18:35:1719import tempfile
20import time
[email protected]9c47c5fd2014-05-15 00:13:5121import zipfile
22
23
24_LOGGER = logging.getLogger(os.path.basename(__file__))
25
chrisha3641d5f2015-07-14 18:35:1726# The relative path where official builds are archived in their GS bucket.
27_SYZYGY_ARCHIVE_PATH = ('/builds/official/%(revision)s')
[email protected]9c47c5fd2014-05-15 00:13:5128
29# A JSON file containing the state of the download directory. If this file and
30# directory state do not agree, then the binaries will be downloaded and
31# installed again.
32_STATE = '.state'
33
34# This matches an integer (an SVN revision number) or a SHA1 value (a GIT hash).
35# The archive exclusively uses lowercase GIT hashes.
36_REVISION_RE = re.compile('^(?:\d+|[a-f0-9]{40})$')
37
38# This matches an MD5 hash.
39_MD5_RE = re.compile('^[a-f0-9]{32}$')
40
41# List of reources to be downloaded and installed. These are tuples with the
42# following format:
43# (basename, logging name, relative installation path, extraction filter)
44_RESOURCES = [
45 ('benchmark.zip', 'benchmark', '', None),
46 ('binaries.zip', 'binaries', 'exe', None),
[email protected]fcee7862014-06-15 18:28:4847 ('symbols.zip', 'symbols', 'exe',
sebmarchandf07597f2014-10-23 23:04:4448 lambda x: x.filename.endswith('.dll.pdb'))]
[email protected]9c47c5fd2014-05-15 00:13:5149
50
[email protected]9c47c5fd2014-05-15 00:13:5151def _LoadState(output_dir):
52 """Loads the contents of the state file for a given |output_dir|, returning
53 None if it doesn't exist.
54 """
55 path = os.path.join(output_dir, _STATE)
56 if not os.path.exists(path):
57 _LOGGER.debug('No state file found.')
58 return None
59 with open(path, 'rb') as f:
60 _LOGGER.debug('Reading state file: %s', path)
61 try:
62 return json.load(f)
63 except ValueError:
64 _LOGGER.debug('Invalid state file.')
65 return None
66
67
68def _SaveState(output_dir, state, dry_run=False):
69 """Saves the |state| dictionary to the given |output_dir| as a JSON file."""
70 path = os.path.join(output_dir, _STATE)
71 _LOGGER.debug('Writing state file: %s', path)
72 if dry_run:
73 return
74 with open(path, 'wb') as f:
75 f.write(json.dumps(state, sort_keys=True, indent=2))
76
77
78def _Md5(path):
79 """Returns the MD5 hash of the file at |path|, which must exist."""
80 return hashlib.md5(open(path, 'rb').read()).hexdigest()
81
82
83def _StateIsValid(state):
84 """Returns true if the given state structure is valid."""
85 if not isinstance(state, dict):
86 _LOGGER.debug('State must be a dict.')
87 return False
88 r = state.get('revision', None)
89 if not isinstance(r, basestring) or not _REVISION_RE.match(r):
90 _LOGGER.debug('State contains an invalid revision.')
91 return False
92 c = state.get('contents', None)
93 if not isinstance(c, dict):
94 _LOGGER.debug('State must contain a contents dict.')
95 return False
96 for (relpath, md5) in c.iteritems():
97 if not isinstance(relpath, basestring) or len(relpath) == 0:
98 _LOGGER.debug('State contents dict contains an invalid path.')
99 return False
100 if not isinstance(md5, basestring) or not _MD5_RE.match(md5):
101 _LOGGER.debug('State contents dict contains an invalid MD5 digest.')
102 return False
103 return True
104
105
106def _BuildActualState(stored, revision, output_dir):
107 """Builds the actual state using the provided |stored| state as a template.
108 Only examines files listed in the stored state, causing the script to ignore
109 files that have been added to the directories locally. |stored| must be a
110 valid state dictionary.
111 """
112 contents = {}
113 state = { 'revision': revision, 'contents': contents }
114 for relpath, md5 in stored['contents'].iteritems():
115 abspath = os.path.abspath(os.path.join(output_dir, relpath))
116 if os.path.isfile(abspath):
117 m = _Md5(abspath)
118 contents[relpath] = m
119
120 return state
121
122
123def _StatesAreConsistent(stored, actual):
124 """Validates whether two state dictionaries are consistent. Both must be valid
125 state dictionaries. Additional entries in |actual| are ignored.
126 """
127 if stored['revision'] != actual['revision']:
128 _LOGGER.debug('Mismatched revision number.')
129 return False
130 cont_stored = stored['contents']
131 cont_actual = actual['contents']
132 for relpath, md5 in cont_stored.iteritems():
133 if relpath not in cont_actual:
134 _LOGGER.debug('Missing content: %s', relpath)
135 return False
136 if md5 != cont_actual[relpath]:
137 _LOGGER.debug('Modified content: %s', relpath)
138 return False
139 return True
140
141
142def _GetCurrentState(revision, output_dir):
143 """Loads the current state and checks to see if it is consistent. Returns
144 a tuple (state, bool). The returned state will always be valid, even if an
145 invalid state is present on disk.
146 """
147 stored = _LoadState(output_dir)
148 if not _StateIsValid(stored):
149 _LOGGER.debug('State is invalid.')
150 # Return a valid but empty state.
151 return ({'revision': '0', 'contents': {}}, False)
152 actual = _BuildActualState(stored, revision, output_dir)
[email protected]fcee7862014-06-15 18:28:48153 # If the script has been modified consider the state invalid.
154 path = os.path.join(output_dir, _STATE)
155 if os.path.getmtime(__file__) > os.path.getmtime(path):
156 return (stored, False)
157 # Otherwise, explicitly validate the state.
[email protected]9c47c5fd2014-05-15 00:13:51158 if not _StatesAreConsistent(stored, actual):
159 return (stored, False)
160 return (stored, True)
161
162
163def _DirIsEmpty(path):
164 """Returns true if the given directory is empty, false otherwise."""
165 for root, dirs, files in os.walk(path):
166 return not dirs and not files
167
168
[email protected]f1d116b2014-06-13 18:07:26169def _RmTreeHandleReadOnly(func, path, exc):
170 """An error handling function for use with shutil.rmtree. This will
171 detect failures to remove read-only files, and will change their properties
172 prior to removing them. This is necessary on Windows as os.remove will return
173 an access error for read-only files, and git repos contain read-only
174 pack/index files.
175 """
176 excvalue = exc[1]
177 if func in (os.rmdir, os.remove) and excvalue.errno == errno.EACCES:
178 _LOGGER.debug('Removing read-only path: %s', path)
179 os.chmod(path, stat.S_IRWXU | stat.S_IRWXG | stat.S_IRWXO)
180 func(path)
181 else:
182 raise
183
184
185def _RmTree(path):
186 """A wrapper of shutil.rmtree that handles read-only files."""
187 shutil.rmtree(path, ignore_errors=False, onerror=_RmTreeHandleReadOnly)
188
189
[email protected]9c47c5fd2014-05-15 00:13:51190def _CleanState(output_dir, state, dry_run=False):
191 """Cleans up files/directories in |output_dir| that are referenced by
192 the given |state|. Raises an error if there are local changes. Returns a
193 dictionary of files that were deleted.
194 """
195 _LOGGER.debug('Deleting files from previous installation.')
196 deleted = {}
197
198 # Generate a list of files to delete, relative to |output_dir|.
199 contents = state['contents']
200 files = sorted(contents.keys())
201
202 # Try to delete the files. Keep track of directories to delete as well.
203 dirs = {}
204 for relpath in files:
205 fullpath = os.path.join(output_dir, relpath)
206 fulldir = os.path.dirname(fullpath)
207 dirs[fulldir] = True
208 if os.path.exists(fullpath):
209 # If somehow the file has become a directory complain about it.
210 if os.path.isdir(fullpath):
211 raise Exception('Directory exists where file expected: %s' % fullpath)
212
213 # Double check that the file doesn't have local changes. If it does
214 # then refuse to delete it.
215 if relpath in contents:
216 stored_md5 = contents[relpath]
217 actual_md5 = _Md5(fullpath)
218 if actual_md5 != stored_md5:
219 raise Exception('File has local changes: %s' % fullpath)
220
221 # The file is unchanged so it can safely be deleted.
222 _LOGGER.debug('Deleting file "%s".', fullpath)
223 deleted[relpath] = True
224 if not dry_run:
225 os.unlink(fullpath)
226
227 # Sort directories from longest name to shortest. This lets us remove empty
228 # directories from the most nested paths first.
229 dirs = sorted(dirs.keys(), key=lambda x: len(x), reverse=True)
230 for p in dirs:
231 if os.path.exists(p) and _DirIsEmpty(p):
232 _LOGGER.debug('Deleting empty directory "%s".', p)
233 if not dry_run:
[email protected]f1d116b2014-06-13 18:07:26234 _RmTree(p)
[email protected]9c47c5fd2014-05-15 00:13:51235
236 return deleted
237
238
chrisha3641d5f2015-07-14 18:35:17239def _FindGsUtil():
240 """Looks for depot_tools and returns the absolute path to gsutil.py."""
241 for path in os.environ['PATH'].split(os.pathsep):
242 path = os.path.abspath(path)
243 git_cl = os.path.join(path, 'git_cl.py')
244 gs_util = os.path.join(path, 'gsutil.py')
245 if os.path.exists(git_cl) and os.path.exists(gs_util):
246 return gs_util
247 return None
248
249
250def _GsUtil(*cmd):
251 """Runs the given command in gsutil with exponential backoff and retries."""
252 gs_util = _FindGsUtil()
253 cmd = [sys.executable, gs_util] + list(cmd)
254
255 retries = 3
256 timeout = 4 # Seconds.
257 while True:
258 _LOGGER.debug('Running %s', cmd)
bcwhite2fc996192015-07-16 18:44:59259 prog = subprocess.Popen(cmd, shell=False)
chrisha3641d5f2015-07-14 18:35:17260 prog.communicate()
261
262 # Stop retrying on success.
263 if prog.returncode == 0:
264 return
265
266 # Raise a permanent failure if retries have been exhausted.
267 if retries == 0:
268 raise RuntimeError('Command "%s" returned %d.' % (cmd, prog.returncode))
269
270 _LOGGER.debug('Sleeping %d seconds and trying again.', timeout)
271 time.sleep(timeout)
272 retries -= 1
273 timeout *= 2
274
275
276def _Download(resource):
277 """Downloads the given GS resource to a temporary file, returning its path."""
278 tmp = tempfile.mkstemp(suffix='syzygy_archive')
279 os.close(tmp[0])
280 url = 'gs://syzygy-archive' + resource
281 _GsUtil('cp', url, tmp[1])
282 return tmp[1]
[email protected]9c47c5fd2014-05-15 00:13:51283
284
285def _InstallBinaries(options, deleted={}):
286 """Installs Syzygy binaries. This assumes that the output directory has
287 already been cleaned, as it will refuse to overwrite existing files."""
288 contents = {}
289 state = { 'revision': options.revision, 'contents': contents }
chrisha3641d5f2015-07-14 18:35:17290 archive_path = _SYZYGY_ARCHIVE_PATH % { 'revision': options.revision }
erikwright9bdc2bf2015-01-22 20:20:04291 if options.resources:
292 resources = [(resource, resource, '', None)
293 for resource in options.resources]
294 else:
295 resources = _RESOURCES
296 for (base, name, subdir, filt) in resources:
[email protected]9c47c5fd2014-05-15 00:13:51297 # Create the output directory if it doesn't exist.
298 fulldir = os.path.join(options.output_dir, subdir)
299 if os.path.isfile(fulldir):
300 raise Exception('File exists where a directory needs to be created: %s' %
301 fulldir)
302 if not os.path.exists(fulldir):
303 _LOGGER.debug('Creating directory: %s', fulldir)
304 if not options.dry_run:
305 os.makedirs(fulldir)
306
chrisha3641d5f2015-07-14 18:35:17307 # Download and read the archive.
308 resource = archive_path + '/' + base
309 _LOGGER.debug('Retrieving %s archive at "%s".', name, resource)
310 path = _Download(resource)
[email protected]9c47c5fd2014-05-15 00:13:51311
312 _LOGGER.debug('Unzipping %s archive.', name)
chrisha3641d5f2015-07-14 18:35:17313 with open(path, 'rb') as data:
314 archive = zipfile.ZipFile(data)
315 for entry in archive.infolist():
316 if not filt or filt(entry):
317 fullpath = os.path.normpath(os.path.join(fulldir, entry.filename))
318 relpath = os.path.relpath(fullpath, options.output_dir)
319 if os.path.exists(fullpath):
320 # If in a dry-run take into account the fact that the file *would*
321 # have been deleted.
322 if options.dry_run and relpath in deleted:
323 pass
324 else:
325 raise Exception('Path already exists: %s' % fullpath)
[email protected]9c47c5fd2014-05-15 00:13:51326
chrisha3641d5f2015-07-14 18:35:17327 # Extract the file and update the state dictionary.
328 _LOGGER.debug('Extracting "%s".', fullpath)
329 if not options.dry_run:
330 archive.extract(entry.filename, fulldir)
331 md5 = _Md5(fullpath)
332 contents[relpath] = md5
333 if sys.platform == 'cygwin':
334 os.chmod(fullpath, os.stat(fullpath).st_mode | stat.S_IXUSR)
335
336 _LOGGER.debug('Removing temporary file "%s".', path)
337 os.remove(path)
[email protected]9c47c5fd2014-05-15 00:13:51338
339 return state
340
341
342def _ParseCommandLine():
343 """Parses the command-line and returns an options structure."""
344 option_parser = optparse.OptionParser()
345 option_parser.add_option('--dry-run', action='store_true', default=False,
346 help='If true then will simply list actions that would be performed.')
347 option_parser.add_option('--force', action='store_true', default=False,
348 help='Force an installation even if the binaries are up to date.')
chrisha3641d5f2015-07-14 18:35:17349 option_parser.add_option('--no-cleanup', action='store_true', default=False,
350 help='Allow installation on non-Windows platforms, and skip the forced '
351 'cleanup step.')
[email protected]9c47c5fd2014-05-15 00:13:51352 option_parser.add_option('--output-dir', type='string',
353 help='The path where the binaries will be replaced. Existing binaries '
354 'will only be overwritten if not up to date.')
355 option_parser.add_option('--overwrite', action='store_true', default=False,
356 help='If specified then the installation will happily delete and rewrite '
357 'the entire output directory, blasting any local changes.')
358 option_parser.add_option('--revision', type='string',
359 help='The SVN revision or GIT hash associated with the required version.')
360 option_parser.add_option('--revision-file', type='string',
361 help='A text file containing an SVN revision or GIT hash.')
erikwright9bdc2bf2015-01-22 20:20:04362 option_parser.add_option('--resource', type='string', action='append',
363 dest='resources', help='A resource to be downloaded.')
[email protected]9c47c5fd2014-05-15 00:13:51364 option_parser.add_option('--verbose', dest='log_level', action='store_const',
365 default=logging.INFO, const=logging.DEBUG,
366 help='Enables verbose logging.')
367 option_parser.add_option('--quiet', dest='log_level', action='store_const',
368 default=logging.INFO, const=logging.ERROR,
369 help='Disables all output except for errors.')
370 options, args = option_parser.parse_args()
371 if args:
372 option_parser.error('Unexpected arguments: %s' % args)
373 if not options.output_dir:
374 option_parser.error('Must specify --output-dir.')
375 if not options.revision and not options.revision_file:
376 option_parser.error('Must specify one of --revision or --revision-file.')
377 if options.revision and options.revision_file:
378 option_parser.error('Must not specify both --revision and --revision-file.')
379
380 # Configure logging.
381 logging.basicConfig(level=options.log_level)
382
383 # If a revision file has been specified then read it.
384 if options.revision_file:
385 options.revision = open(options.revision_file, 'rb').read().strip()
386 _LOGGER.debug('Parsed revision "%s" from file "%s".',
387 options.revision, options.revision_file)
388
389 # Ensure that the specified SVN revision or GIT hash is valid.
390 if not _REVISION_RE.match(options.revision):
391 option_parser.error('Must specify a valid SVN or GIT revision.')
392
393 # This just makes output prettier to read.
394 options.output_dir = os.path.normpath(options.output_dir)
395
396 return options
397
398
chrishaf3b3b3bb2014-09-19 19:27:58399def _RemoveOrphanedFiles(options):
400 """This is run on non-Windows systems to remove orphaned files that may have
401 been downloaded by a previous version of this script.
402 """
403 # Reconfigure logging to output info messages. This will allow inspection of
404 # cleanup status on non-Windows buildbots.
405 _LOGGER.setLevel(logging.INFO)
406
407 output_dir = os.path.abspath(options.output_dir)
408
409 # We only want to clean up the folder in 'src/third_party/syzygy', and we
410 # expect to be called with that as an output directory. This is an attempt to
411 # not start deleting random things if the script is run from an alternate
412 # location, or not called from the gclient hooks.
413 expected_syzygy_dir = os.path.abspath(os.path.join(
414 os.path.dirname(__file__), '..', 'third_party', 'syzygy'))
415 expected_output_dir = os.path.join(expected_syzygy_dir, 'binaries')
416 if expected_output_dir != output_dir:
417 _LOGGER.info('Unexpected output directory, skipping cleanup.')
[email protected]f1d116b2014-06-13 18:07:26418 return
419
chrishaf3b3b3bb2014-09-19 19:27:58420 if not os.path.isdir(expected_syzygy_dir):
421 _LOGGER.info('Output directory does not exist, skipping cleanup.')
422 return
423
424 def OnError(function, path, excinfo):
425 """Logs error encountered by shutil.rmtree."""
426 _LOGGER.error('Error when running %s(%s)', function, path, exc_info=excinfo)
427
428 _LOGGER.info('Removing orphaned files from %s', expected_syzygy_dir)
429 if not options.dry_run:
430 shutil.rmtree(expected_syzygy_dir, True, OnError)
431
432
433def main():
[email protected]9c47c5fd2014-05-15 00:13:51434 options = _ParseCommandLine()
435
436 if options.dry_run:
437 _LOGGER.debug('Performing a dry-run.')
438
chrishaf3b3b3bb2014-09-19 19:27:58439 # We only care about Windows platforms, as the Syzygy binaries aren't used
440 # elsewhere. However, there was a short period of time where this script
441 # wasn't gated on OS types, and those OSes downloaded and installed binaries.
442 # This will cleanup orphaned files on those operating systems.
443 if sys.platform not in ('win32', 'cygwin'):
chrisha3641d5f2015-07-14 18:35:17444 if options.no_cleanup:
445 _LOGGER.debug('Skipping usual cleanup for non-Windows platforms.')
446 else:
447 return _RemoveOrphanedFiles(options)
chrishaf3b3b3bb2014-09-19 19:27:58448
[email protected]9c47c5fd2014-05-15 00:13:51449 # Load the current installation state, and validate it against the
450 # requested installation.
451 state, is_consistent = _GetCurrentState(options.revision, options.output_dir)
452
453 # Decide whether or not an install is necessary.
454 if options.force:
455 _LOGGER.debug('Forcing reinstall of binaries.')
456 elif is_consistent:
457 # Avoid doing any work if the contents of the directory are consistent.
458 _LOGGER.debug('State unchanged, no reinstall necessary.')
459 return
460
461 # Under normal logging this is the only only message that will be reported.
462 _LOGGER.info('Installing revision %s Syzygy binaries.',
463 options.revision[0:12])
464
465 # Clean up the old state to begin with.
466 deleted = []
467 if options.overwrite:
468 if os.path.exists(options.output_dir):
469 # If overwrite was specified then take a heavy-handed approach.
470 _LOGGER.debug('Deleting entire installation directory.')
471 if not options.dry_run:
[email protected]f1d116b2014-06-13 18:07:26472 _RmTree(options.output_dir)
[email protected]9c47c5fd2014-05-15 00:13:51473 else:
474 # Otherwise only delete things that the previous installation put in place,
475 # and take care to preserve any local changes.
476 deleted = _CleanState(options.output_dir, state, options.dry_run)
477
478 # Install the new binaries. In a dry-run this will actually download the
479 # archives, but it won't write anything to disk.
480 state = _InstallBinaries(options, deleted)
481
482 # Build and save the state for the directory.
483 _SaveState(options.output_dir, state, options.dry_run)
484
485
486if __name__ == '__main__':
487 main()