| # Copyright (c) 2012 The Chromium OS Authors. All rights reserved. |
| # Use of this source code is governed by a BSD-style license that can be |
| # found in the LICENSE file. |
| |
| import glob |
| import logging |
| import os |
| import re |
| import urlparse |
| import urllib2 |
| |
| from autotest_lib.client.bin import utils |
| from autotest_lib.client.common_lib import error, global_config |
| from autotest_lib.client.common_lib.cros import dev_server |
| from autotest_lib.server import utils as server_utils |
| from chromite.lib import retry_util |
| |
| try: |
| from chromite.lib import metrics |
| except ImportError: |
| metrics = utils.metrics_mock |
| |
| try: |
| import devserver |
| STATEFUL_UPDATE_PATH = devserver.__path__[0] |
| except ImportError: |
| STATEFUL_UPDATE_PATH = '/usr/bin' |
| |
| # Local stateful update path is relative to the CrOS source directory. |
| STATEFUL_UPDATE_SCRIPT = 'stateful_update' |
| UPDATER_IDLE = 'UPDATE_STATUS_IDLE' |
| UPDATER_NEED_REBOOT = 'UPDATE_STATUS_UPDATED_NEED_REBOOT' |
| # A list of update engine client states that occur after an update is triggered. |
| UPDATER_PROCESSING_UPDATE = ['UPDATE_STATUS_CHECKING_FORUPDATE', |
| 'UPDATE_STATUS_UPDATE_AVAILABLE', |
| 'UPDATE_STATUS_DOWNLOADING', |
| 'UPDATE_STATUS_FINALIZING'] |
| |
| class ChromiumOSError(error.InstallError): |
| """Generic error for ChromiumOS-specific exceptions.""" |
| |
| |
| class RootFSUpdateError(ChromiumOSError): |
| """Raised when the RootFS fails to update.""" |
| |
| |
| class StatefulUpdateError(ChromiumOSError): |
| """Raised when the stateful partition fails to update.""" |
| |
| |
| def url_to_version(update_url): |
| """Return the version based on update_url. |
| |
| @param update_url: url to the image to update to. |
| |
| """ |
| # The Chrome OS version is generally the last element in the URL. The only |
| # exception is delta update URLs, which are rooted under the version; e.g., |
| # http://.../update/.../0.14.755.0/au/0.14.754.0. In this case we want to |
| # strip off the au section of the path before reading the version. |
| return re.sub('/au/.*', '', |
| urlparse.urlparse(update_url).path).split('/')[-1].strip() |
| |
| |
| def url_to_image_name(update_url): |
| """Return the image name based on update_url. |
| |
| From a URL like: |
| http://172.22.50.205:8082/update/lumpy-release/R27-3837.0.0 |
| return lumpy-release/R27-3837.0.0 |
| |
| @param update_url: url to the image to update to. |
| @returns a string representing the image name in the update_url. |
| |
| """ |
| return '/'.join(urlparse.urlparse(update_url).path.split('/')[-2:]) |
| |
| |
| def _get_devserver_build_from_update_url(update_url): |
| """Get the devserver and build from the update url. |
| |
| @param update_url: The url for update. |
| Eg: http://devserver:port/update/build. |
| |
| @return: A tuple of (devserver url, build) or None if the update_url |
| doesn't match the expected pattern. |
| |
| @raises ValueError: If the update_url doesn't match the expected pattern. |
| @raises ValueError: If no global_config was found, or it doesn't contain an |
| image_url_pattern. |
| """ |
| pattern = global_config.global_config.get_config_value( |
| 'CROS', 'image_url_pattern', type=str, default='') |
| if not pattern: |
| raise ValueError('Cannot parse update_url, the global config needs ' |
| 'an image_url_pattern.') |
| re_pattern = pattern.replace('%s', '(\S+)') |
| parts = re.search(re_pattern, update_url) |
| if not parts or len(parts.groups()) < 2: |
| raise ValueError('%s is not an update url' % update_url) |
| return parts.groups() |
| |
| |
| def list_image_dir_contents(update_url): |
| """Lists the contents of the devserver for a given build/update_url. |
| |
| @param update_url: An update url. Eg: http://devserver:port/update/build. |
| """ |
| if not update_url: |
| logging.warning('Need update_url to list contents of the devserver.') |
| return |
| error_msg = 'Cannot check contents of devserver, update url %s' % update_url |
| try: |
| devserver_url, build = _get_devserver_build_from_update_url(update_url) |
| except ValueError as e: |
| logging.warning('%s: %s', error_msg, e) |
| return |
| devserver = dev_server.ImageServer(devserver_url) |
| try: |
| devserver.list_image_dir(build) |
| # The devserver will retry on URLError to avoid flaky connections, but will |
| # eventually raise the URLError if it persists. All HTTPErrors get |
| # converted to DevServerExceptions. |
| except (dev_server.DevServerException, urllib2.URLError) as e: |
| logging.warning('%s: %s', error_msg, e) |
| |
| |
| # TODO(garnold) This implements shared updater functionality needed for |
| # supporting the autoupdate_EndToEnd server-side test. We should probably |
| # migrate more of the existing ChromiumOSUpdater functionality to it as we |
| # expand non-CrOS support in other tests. |
| class BaseUpdater(object): |
| """Platform-agnostic DUT update functionality.""" |
| |
| def __init__(self, updater_ctrl_bin, update_url, host, interactive=True): |
| """Initializes the object. |
| |
| @param updater_ctrl_bin: Path to update_engine_client. |
| @param update_url: The URL we want the update to use. |
| @param host: A client.common_lib.hosts.Host implementation. |
| @param interactive: Bool whether we are doing an interactive update. |
| """ |
| self.updater_ctrl_bin = updater_ctrl_bin |
| self.update_url = update_url |
| self.host = host |
| self.interactive = interactive |
| |
| |
| def check_update_status(self): |
| """Returns the current update engine state. |
| |
| We use the `update_engine_client -status' command and parse the line |
| indicating the update state, e.g. "CURRENT_OP=UPDATE_STATUS_IDLE". |
| """ |
| update_status = self.host.run(command='%s -status | grep CURRENT_OP' % |
| self.updater_ctrl_bin) |
| return update_status.stdout.strip().split('=')[-1] |
| |
| |
| def get_last_update_error(self): |
| """Get the last autoupdate error code.""" |
| error_msg = self.host.run( |
| '%s --last_attempt_error' % self.updater_ctrl_bin) |
| error_msg = (error_msg.stdout.strip()).replace('\n', ', ') |
| return error_msg |
| |
| |
| def _base_update_handler_no_retry(self, run_args): |
| """Base function to handle a remote update ssh call. |
| |
| @param run_args: Dictionary of args passed to ssh_host.run function. |
| |
| @throws: intercepts and re-throws all exceptions |
| """ |
| try: |
| self.host.run(**run_args) |
| except Exception as e: |
| logging.debug('exception in update handler: %s', e) |
| raise e |
| |
| |
| def _base_update_handler(self, run_args, err_msg_prefix=None): |
| """Handle a remote update ssh call, possibly with retries. |
| |
| @param run_args: Dictionary of args passed to ssh_host.run function. |
| @param err_msg_prefix: Prefix of the exception error message. |
| """ |
| def exception_handler(e): |
| """Examines exceptions and returns True if the update handler |
| should be retried. |
| |
| @param e: the exception intercepted by the retry util. |
| """ |
| return (isinstance(e, error.AutoservSSHTimeout) or |
| (isinstance(e, error.GenericHostRunError) and |
| hasattr(e, 'description') and |
| (re.search('ERROR_CODE=37', e.description) or |
| re.search('generic error .255.', e.description)))) |
| |
| try: |
| # Try the update twice (arg 2 is max_retry, not including the first |
| # call). Some exceptions may be caught by the retry handler. |
| retry_util.GenericRetry(exception_handler, 1, |
| self._base_update_handler_no_retry, |
| run_args) |
| except Exception as e: |
| message = err_msg_prefix + ': ' + str(e) |
| raise RootFSUpdateError(message) |
| |
| |
| def _wait_for_update_service(self): |
| """Ensure that the update engine daemon is running, possibly |
| by waiting for it a bit in case the DUT just rebooted and the |
| service hasn't started yet. |
| """ |
| def handler(e): |
| """Retry exception handler. |
| |
| Assumes that the error is due to the update service not having |
| started yet. |
| |
| @param e: the exception intercepted by the retry util. |
| """ |
| if isinstance(e, error.AutoservRunError): |
| logging.debug('update service check exception: %s\n' |
| 'retrying...', e) |
| return True |
| else: |
| return False |
| |
| # Retry at most three times, every 5s. |
| status = retry_util.GenericRetry(handler, 3, |
| self.check_update_status, |
| sleep=5) |
| |
| # Expect the update engine to be idle. |
| if status != UPDATER_IDLE: |
| raise ChromiumOSError('%s is not in an installable state' % |
| self.host.hostname) |
| |
| |
| def trigger_update(self): |
| """Triggers a background update. |
| |
| @raise RootFSUpdateError or unknown Exception if anything went wrong. |
| """ |
| # If this function is called immediately after reboot (which it is at |
| # this time), there is no guarantee that the update service is up and |
| # running yet, so wait for it. |
| self._wait_for_update_service() |
| |
| autoupdate_cmd = ('%s --check_for_update --omaha_url=%s' % |
| (self.updater_ctrl_bin, self.update_url)) |
| run_args = {'command': autoupdate_cmd} |
| err_prefix = 'Failed to trigger an update on %s. ' % self.host.hostname |
| logging.info('Triggering update via: %s', autoupdate_cmd) |
| metric_fields = {'success': False} |
| try: |
| self._base_update_handler(run_args, err_prefix) |
| metric_fields['success'] = True |
| finally: |
| c = metrics.Counter('chromeos/autotest/autoupdater/trigger') |
| metric_fields.update(self._get_metric_fields()) |
| c.increment(fields=metric_fields) |
| |
| |
| def _get_metric_fields(self): |
| """Return a dict of metric fields. |
| |
| This is used for sending autoupdate metrics for this instance. |
| """ |
| build_name = url_to_image_name(self.update_url) |
| try: |
| board, build_type, milestone, _ = server_utils.ParseBuildName( |
| build_name) |
| except server_utils.ParseBuildNameException: |
| logging.warning('Unable to parse build name %s for metrics. ' |
| 'Continuing anyway.', build_name) |
| board, build_type, milestone = ('', '', '') |
| return { |
| 'dev_server': dev_server.get_hostname(self.update_url), |
| 'board': board, |
| 'build_type': build_type, |
| 'milestone': milestone, |
| } |
| |
| |
| def _verify_update_completed(self): |
| """Verifies that an update has completed. |
| |
| @raise RootFSUpdateError: if verification fails. |
| """ |
| status = self.check_update_status() |
| if status != UPDATER_NEED_REBOOT: |
| error_msg = '' |
| if status == UPDATER_IDLE: |
| error_msg = 'Update error: %s' % self.get_last_update_error() |
| raise RootFSUpdateError('Update did not complete with correct ' |
| 'status. Expecting %s, actual %s. %s' % |
| (UPDATER_NEED_REBOOT, status, error_msg)) |
| |
| |
| def update_image(self): |
| """Updates the device image and verifies success.""" |
| autoupdate_cmd = ('%s --update --omaha_url=%s' % |
| (self.updater_ctrl_bin, self.update_url)) |
| if not self.interactive: |
| autoupdate_cmd = '%s --interactive=false' % autoupdate_cmd |
| run_args = {'command': autoupdate_cmd, 'timeout': 3600} |
| err_prefix = ('Failed to install device image using payload at %s ' |
| 'on %s. ' % (self.update_url, self.host.hostname)) |
| logging.info('Updating image via: %s', autoupdate_cmd) |
| metric_fields = {'success': False} |
| try: |
| self._base_update_handler(run_args, err_prefix) |
| metric_fields['success'] = True |
| finally: |
| c = metrics.Counter('chromeos/autotest/autoupdater/update') |
| metric_fields.update(self._get_metric_fields()) |
| c.increment(fields=metric_fields) |
| |
| self._verify_update_completed() |
| |
| |
| class ChromiumOSUpdater(BaseUpdater): |
| """Helper class used to update DUT with image of desired version.""" |
| REMOTE_STATEFUL_UPDATE_PATH = os.path.join( |
| '/usr/local/bin', STATEFUL_UPDATE_SCRIPT) |
| REMOTE_TMP_STATEFUL_UPDATE = os.path.join( |
| '/tmp', STATEFUL_UPDATE_SCRIPT) |
| UPDATER_BIN = '/usr/bin/update_engine_client' |
| UPDATED_MARKER = '/run/update_engine_autoupdate_completed' |
| UPDATER_LOGS = ['/var/log/messages', '/var/log/update_engine'] |
| |
| KERNEL_A = {'name': 'KERN-A', 'kernel': 2, 'root': 3} |
| KERNEL_B = {'name': 'KERN-B', 'kernel': 4, 'root': 5} |
| # Time to wait for new kernel to be marked successful after |
| # auto update. |
| KERNEL_UPDATE_TIMEOUT = 120 |
| |
| def __init__(self, update_url, host=None, local_devserver=False, |
| interactive=True): |
| super(ChromiumOSUpdater, self).__init__(self.UPDATER_BIN, update_url, |
| host, interactive=interactive) |
| self.local_devserver = local_devserver |
| if not local_devserver: |
| self.update_version = url_to_version(update_url) |
| else: |
| self.update_version = None |
| |
| |
| def reset_update_engine(self): |
| """Resets the host to prepare for a clean update regardless of state.""" |
| self._run('rm -f %s' % self.UPDATED_MARKER) |
| self._run('stop ui || true') |
| self._run('stop update-engine || true') |
| self._run('start update-engine') |
| |
| # Wait for update engine to be ready. |
| self._wait_for_update_service() |
| |
| |
| def _run(self, cmd, *args, **kwargs): |
| """Abbreviated form of self.host.run(...)""" |
| return self.host.run(cmd, *args, **kwargs) |
| |
| |
| def rootdev(self, options=''): |
| """Returns the stripped output of rootdev <options>. |
| |
| @param options: options to run rootdev. |
| |
| """ |
| return self._run('rootdev %s' % options).stdout.strip() |
| |
| |
| def get_kernel_state(self): |
| """Returns the (<active>, <inactive>) kernel state as a pair.""" |
| active_root = int(re.findall('\d+\Z', self.rootdev('-s'))[0]) |
| if active_root == self.KERNEL_A['root']: |
| return self.KERNEL_A, self.KERNEL_B |
| elif active_root == self.KERNEL_B['root']: |
| return self.KERNEL_B, self.KERNEL_A |
| else: |
| raise ChromiumOSError('Encountered unknown root partition: %s' % |
| active_root) |
| |
| |
| def _cgpt(self, flag, kernel, dev='$(rootdev -s -d)'): |
| """Return numeric cgpt value for the specified flag, kernel, device. """ |
| return int(self._run('cgpt show -n -i %d %s %s' % ( |
| kernel['kernel'], flag, dev)).stdout.strip()) |
| |
| |
| def get_kernel_priority(self, kernel): |
| """Return numeric priority for the specified kernel. |
| |
| @param kernel: information of the given kernel, KERNEL_A or KERNEL_B. |
| |
| """ |
| return self._cgpt('-P', kernel) |
| |
| |
| def get_kernel_success(self, kernel): |
| """Return boolean success flag for the specified kernel. |
| |
| @param kernel: information of the given kernel, KERNEL_A or KERNEL_B. |
| |
| """ |
| return self._cgpt('-S', kernel) != 0 |
| |
| |
| def get_kernel_tries(self, kernel): |
| """Return tries count for the specified kernel. |
| |
| @param kernel: information of the given kernel, KERNEL_A or KERNEL_B. |
| |
| """ |
| return self._cgpt('-T', kernel) |
| |
| |
| def get_stateful_update_script(self): |
| """Returns the path to the stateful update script on the target. |
| |
| When runnning test_that, stateful_update is in chroot /usr/sbin, |
| as installed by chromeos-base/devserver packages. |
| In the lab, it is installed with the python module devserver, by |
| build_externals.py command. |
| |
| If we can find it, we hope it exists already on the DUT, we assert |
| otherwise. |
| """ |
| stateful_update_file = os.path.join(STATEFUL_UPDATE_PATH, |
| STATEFUL_UPDATE_SCRIPT) |
| if os.path.exists(stateful_update_file): |
| self.host.send_file( |
| stateful_update_file, self.REMOTE_TMP_STATEFUL_UPDATE, |
| delete_dest=True) |
| return self.REMOTE_TMP_STATEFUL_UPDATE |
| |
| if self.host.path_exists(self.REMOTE_STATEFUL_UPDATE_PATH): |
| logging.warning('Could not chroot %s script, falling back on %s', |
| STATEFUL_UPDATE_SCRIPT, self.REMOTE_STATEFUL_UPDATE_PATH) |
| return self.REMOTE_STATEFUL_UPDATE_PATH |
| else: |
| raise ChromiumOSError('Could not locate %s', |
| STATEFUL_UPDATE_SCRIPT) |
| |
| |
| def reset_stateful_partition(self): |
| """Clear any pending stateful update request.""" |
| statefuldev_cmd = [self.get_stateful_update_script()] |
| statefuldev_cmd += ['--stateful_change=reset', '2>&1'] |
| self._run(' '.join(statefuldev_cmd)) |
| |
| |
| def revert_boot_partition(self): |
| """Revert the boot partition.""" |
| part = self.rootdev('-s') |
| logging.warning('Reverting update; Boot partition will be %s', part) |
| return self._run('/postinst %s 2>&1' % part) |
| |
| |
| def rollback_rootfs(self, powerwash): |
| """Triggers rollback and waits for it to complete. |
| |
| @param powerwash: If true, powerwash as part of rollback. |
| |
| @raise RootFSUpdateError if anything went wrong. |
| |
| """ |
| version = self.host.get_release_version() |
| # Introduced can_rollback in M36 (build 5772). # etc/lsb-release matches |
| # X.Y.Z. This version split just pulls the first part out. |
| try: |
| build_number = int(version.split('.')[0]) |
| except ValueError: |
| logging.error('Could not parse build number.') |
| build_number = 0 |
| |
| if build_number >= 5772: |
| can_rollback_cmd = '%s --can_rollback' % self.UPDATER_BIN |
| logging.info('Checking for rollback.') |
| try: |
| self._run(can_rollback_cmd) |
| except error.AutoservRunError as e: |
| raise RootFSUpdateError("Rollback isn't possible on %s: %s" % |
| (self.host.hostname, str(e))) |
| |
| rollback_cmd = '%s --rollback --follow' % self.UPDATER_BIN |
| if not powerwash: |
| rollback_cmd += ' --nopowerwash' |
| |
| logging.info('Performing rollback.') |
| try: |
| self._run(rollback_cmd) |
| except error.AutoservRunError as e: |
| raise RootFSUpdateError('Rollback failed on %s: %s' % |
| (self.host.hostname, str(e))) |
| |
| self._verify_update_completed() |
| |
| |
| # TODO(garnold) This is here for backward compatibility and should be |
| # deprecated once we shift to using update_image() everywhere. |
| def update_rootfs(self): |
| """Run the standard command to force an update.""" |
| return self.update_image() |
| |
| |
| def update_stateful(self, clobber=True): |
| """Updates the stateful partition. |
| |
| @param clobber: If True, a clean stateful installation. |
| """ |
| logging.info('Updating stateful partition...') |
| statefuldev_url = self.update_url.replace('update', |
| 'static') |
| |
| # Attempt stateful partition update; this must succeed so that the newly |
| # installed host is testable after update. |
| statefuldev_cmd = [self.get_stateful_update_script(), statefuldev_url] |
| if clobber: |
| statefuldev_cmd.append('--stateful_change=clean') |
| |
| statefuldev_cmd.append('2>&1') |
| try: |
| self._run(' '.join(statefuldev_cmd), timeout=1200) |
| except error.AutoservRunError: |
| update_error = StatefulUpdateError( |
| 'Failed to perform stateful update on %s' % |
| self.host.hostname) |
| raise update_error |
| |
| def run_update(self, update_root=True): |
| """Update the DUT with image of specific version. |
| |
| @param update_root: True to force a rootfs update. |
| """ |
| booted_version = self.host.get_release_version() |
| if self.update_version: |
| logging.info('Updating from version %s to %s.', |
| booted_version, self.update_version) |
| |
| # Check that Dev Server is accepting connections (from autoserv's host). |
| # If we can't talk to it, the machine host probably can't either. |
| auserver_host = 'http://%s' % urlparse.urlparse(self.update_url)[1] |
| try: |
| if not dev_server.ImageServer.devserver_healthy(auserver_host): |
| raise ChromiumOSError( |
| 'Update server at %s not healthy' % auserver_host) |
| except Exception as e: |
| logging.debug('Error happens in connection to devserver: %r', e) |
| raise ChromiumOSError( |
| 'Update server at %s not available' % auserver_host) |
| |
| logging.info('Installing from %s to %s', self.update_url, |
| self.host.hostname) |
| |
| # Reset update state. |
| self.reset_update_engine() |
| self.reset_stateful_partition() |
| |
| try: |
| try: |
| if not update_root: |
| logging.info('Root update is skipped.') |
| else: |
| self.update_rootfs() |
| |
| self.update_stateful() |
| except: |
| self.revert_boot_partition() |
| self.reset_stateful_partition() |
| raise |
| |
| logging.info('Update complete.') |
| except: |
| # Collect update engine logs in the event of failure. |
| if self.host.job: |
| logging.info('Collecting update engine logs due to failure...') |
| self.host.get_file( |
| self.UPDATER_LOGS, self.host.job.sysinfo.sysinfodir, |
| preserve_perm=False) |
| list_image_dir_contents(self.update_url) |
| raise |
| finally: |
| logging.info('Update engine log has downloaded in ' |
| 'sysinfo/update_engine dir. Check the lastest.') |
| |
| |
| def check_version(self): |
| """Check the image running in DUT has the desired version. |
| |
| @returns: True if the DUT's image version matches the version that |
| the autoupdater tries to update to. |
| |
| """ |
| booted_version = self.host.get_release_version() |
| return (self.update_version and |
| self.update_version.endswith(booted_version)) |
| |
| |
| def check_version_to_confirm_install(self): |
| """Check image running in DUT has the desired version to be installed. |
| |
| The method should not be used to check if DUT needs to have a full |
| reimage. Only use it to confirm a image is installed. |
| |
| The method is designed to verify version for following 6 scenarios with |
| samples of version to update to and expected booted version: |
| 1. trybot paladin build. |
| update version: trybot-lumpy-paladin/R27-3837.0.0-b123 |
| booted version: 3837.0.2013_03_21_1340 |
| |
| 2. trybot release build. |
| update version: trybot-lumpy-release/R27-3837.0.0-b456 |
| booted version: 3837.0.0 |
| |
| 3. buildbot official release build. |
| update version: lumpy-release/R27-3837.0.0 |
| booted version: 3837.0.0 |
| |
| 4. non-official paladin rc build. |
| update version: lumpy-paladin/R27-3878.0.0-rc7 |
| booted version: 3837.0.0-rc7 |
| |
| 5. chrome-perf build. |
| update version: lumpy-chrome-perf/R28-3837.0.0-b2996 |
| booted version: 3837.0.0 |
| |
| 6. pgo-generate build. |
| update version: lumpy-release-pgo-generate/R28-3837.0.0-b2996 |
| booted version: 3837.0.0-pgo-generate |
| |
| When we are checking if a DUT needs to do a full install, we should NOT |
| use this method to check if the DUT is running the same version, since |
| it may return false positive for a DUT running trybot paladin build to |
| be updated to another trybot paladin build. |
| |
| TODO: This logic has a bug if a trybot paladin build failed to be |
| installed in a DUT running an older trybot paladin build with same |
| platform number, but different build number (-b###). So to conclusively |
| determine if a tryjob paladin build is imaged successfully, we may need |
| to find out the date string from update url. |
| |
| @returns: True if the DUT's image version (without the date string if |
| the image is a trybot build), matches the version that the |
| autoupdater is trying to update to. |
| |
| """ |
| # In the local_devserver case, we can't know the expected |
| # build, so just pass. |
| if not self.update_version: |
| return True |
| |
| # Always try the default check_version method first, this prevents |
| # any backward compatibility issue. |
| if self.check_version(): |
| return True |
| |
| return utils.version_match(self.update_version, |
| self.host.get_release_version(), |
| self.update_url) |
| |
| |
| def verify_boot_expectations(self, expected_kernel_state, rollback_message): |
| """Verifies that we fully booted given expected kernel state. |
| |
| This method both verifies that we booted using the correct kernel |
| state and that the OS has marked the kernel as good. |
| |
| @param expected_kernel_state: kernel state that we are verifying with |
| i.e. I expect to be booted onto partition 4 etc. See output of |
| get_kernel_state. |
| @param rollback_message: string to raise as a ChromiumOSError |
| if we booted with the wrong partition. |
| |
| @raises ChromiumOSError: If we didn't. |
| """ |
| # Figure out the newly active kernel. |
| active_kernel_state = self.get_kernel_state()[0] |
| |
| # Check for rollback due to a bad build. |
| if (expected_kernel_state and |
| active_kernel_state != expected_kernel_state): |
| |
| # Kernel crash reports should be wiped between test runs, but |
| # may persist from earlier parts of the test, or from problems |
| # with provisioning. |
| # |
| # Kernel crash reports will NOT be present if the crash happened |
| # before encrypted stateful is mounted. |
| # |
| # TODO(dgarrett): Integrate with server/crashcollect.py at some |
| # point. |
| kernel_crashes = glob.glob('/var/spool/crash/kernel.*.kcrash') |
| if kernel_crashes: |
| rollback_message += ': kernel_crash' |
| logging.debug('Found %d kernel crash reports:', |
| len(kernel_crashes)) |
| # The crash names contain timestamps that may be useful: |
| # kernel.20131207.005945.0.kcrash |
| for crash in kernel_crashes: |
| logging.debug(' %s', os.path.basename(crash)) |
| |
| # Print out some information to make it easier to debug |
| # the rollback. |
| logging.debug('Dumping partition table.') |
| self._run('cgpt show $(rootdev -s -d)') |
| logging.debug('Dumping crossystem for firmware debugging.') |
| self._run('crossystem --all') |
| raise ChromiumOSError(rollback_message) |
| |
| # Make sure chromeos-setgoodkernel runs. |
| try: |
| utils.poll_for_condition( |
| lambda: (self.get_kernel_tries(active_kernel_state) == 0 |
| and self.get_kernel_success(active_kernel_state)), |
| exception=ChromiumOSError(), |
| timeout=self.KERNEL_UPDATE_TIMEOUT, sleep_interval=5) |
| except ChromiumOSError: |
| services_status = self._run('status system-services').stdout |
| if services_status != 'system-services start/running\n': |
| event = ('Chrome failed to reach login screen') |
| else: |
| event = ('update-engine failed to call ' |
| 'chromeos-setgoodkernel') |
| raise ChromiumOSError( |
| 'After update and reboot, %s ' |
| 'within %d seconds' % (event, |
| self.KERNEL_UPDATE_TIMEOUT)) |
| |