| import abc |
| import datetime |
| import glob |
| import json |
| import os |
| import re |
| import shutil |
| import time |
| |
| import common |
| from autotest_lib.client.common_lib import time_utils |
| from autotest_lib.client.common_lib import utils |
| from autotest_lib.server.cros.dynamic_suite import constants |
| from autotest_lib.server.cros.dynamic_suite import frontend_wrappers |
| |
| |
| _AFE = frontend_wrappers.RetryingAFE() |
| |
| SPECIAL_TASK_PATTERN = '.*/hosts/[^/]+/(\d+)-[^/]+' |
| JOB_PATTERN = '.*/(\d+)-[^/]+' |
| # Pattern of a job folder, e.g., 123-debug_user, where 123 is job id and |
| # debug_user is the name of user starts the job. |
| JOB_FOLDER_PATTERN = '.*/(\d+-[^/]+)' |
| |
| def is_job_expired(age_limit, timestamp): |
| """Check whether a job timestamp is older than an age limit. |
| |
| @param age_limit: Minimum age, measured in days. If the value is |
| not positive, the job is always expired. |
| @param timestamp: Timestamp of the job whose age we are checking. |
| The format must match time_utils.TIME_FMT. |
| |
| @returns True iff the job is old enough to be expired. |
| """ |
| if age_limit <= 0: |
| return True |
| job_time = time_utils.time_string_to_datetime(timestamp) |
| expiration = job_time + datetime.timedelta(days=age_limit) |
| return datetime.datetime.now() >= expiration |
| |
| |
| def get_job_id_or_task_id(result_dir): |
| """Extract job id or special task id from result_dir |
| |
| @param result_dir: path to the result dir. |
| For test job: |
| /usr/local/autotest/results/2032-chromeos-test/chromeos1-rack5-host6 |
| The hostname at the end is optional. |
| For special task: |
| /usr/local/autotest/results/hosts/chromeos1-rack5-host6/1343-cleanup |
| |
| @returns: integer representing the job id or task id. Returns None if fail |
| to parse job or task id from the result_dir. |
| """ |
| if not result_dir: |
| return |
| result_dir = os.path.abspath(result_dir) |
| # Result folder for job running inside container has only job id. |
| ssp_job_pattern = '.*/(\d+)$' |
| # Try to get the job ID from the last pattern of number-text. This avoids |
| # issue with path like 123-results/456-debug_user, in which 456 is the real |
| # job ID. |
| m_job = re.findall(JOB_PATTERN, result_dir) |
| if m_job: |
| return int(m_job[-1]) |
| m_special_task = re.match(SPECIAL_TASK_PATTERN, result_dir) |
| if m_special_task: |
| return int(m_special_task.group(1)) |
| m_ssp_job_pattern = re.match(ssp_job_pattern, result_dir) |
| if m_ssp_job_pattern and utils.is_in_container(): |
| return int(m_ssp_job_pattern.group(1)) |
| |
| |
| def get_job_folder_name(result_dir): |
| """Extract folder name of a job from result_dir. |
| |
| @param result_dir: path to the result dir. |
| For test job: |
| /usr/local/autotest/results/2032-chromeos-test/chromeos1-rack5-host6 |
| The hostname at the end is optional. |
| For special task: |
| /usr/local/autotest/results/hosts/chromeos1-rack5-host6/1343-cleanup |
| |
| @returns: The name of the folder of a job. Returns None if fail to parse |
| the name matching pattern JOB_FOLDER_PATTERN from the result_dir. |
| """ |
| if not result_dir: |
| return |
| m_job = re.findall(JOB_FOLDER_PATTERN, result_dir) |
| if m_job: |
| return m_job[-1] |
| |
| |
| class _JobDirectory(object): |
| """State associated with a job to be offloaded. |
| |
| The full life-cycle of a job (including failure events that |
| normally don't occur) looks like this: |
| 1. The job's results directory is discovered by |
| `get_job_directories()`, and a job instance is created for it. |
| 2. Calls to `offload()` have no effect so long as the job |
| isn't complete in the database and the job isn't expired |
| according to the `age_limit` parameter. |
| 3. Eventually, the job is both finished and expired. The next |
| call to `offload()` makes the first attempt to offload the |
| directory to GS. Offload is attempted, but fails to complete |
| (e.g. because of a GS problem). |
| 4. Finally, a call to `offload()` succeeds, and the directory no |
| longer exists. Now `is_offloaded()` is true, so the job |
| instance is deleted, and future failures will not mention this |
| directory any more. |
| |
| Only steps 1. and 4. are guaranteed to occur. The others depend |
| on the timing of calls to `offload()`, and on the reliability of |
| the actual offload process. |
| |
| """ |
| |
| __metaclass__ = abc.ABCMeta |
| |
| GLOB_PATTERN = None # must be redefined in subclass |
| |
| def __init__(self, resultsdir): |
| self._dirname = resultsdir |
| self._id = get_job_id_or_task_id(resultsdir) |
| self._offload_count = 0 |
| self._first_offload_start = 0 |
| |
| @classmethod |
| def get_job_directories(cls): |
| """Return a list of directories of jobs that need offloading.""" |
| return [d for d in glob.glob(cls.GLOB_PATTERN) if os.path.isdir(d)] |
| |
| @abc.abstractmethod |
| def get_timestamp_if_finished(self): |
| """Return this job's timestamp from the database. |
| |
| If the database has not marked the job as finished, return |
| `None`. Otherwise, return a timestamp for the job. The |
| timestamp is to be used to determine expiration in |
| `is_job_expired()`. |
| |
| @return Return `None` if the job is still running; otherwise |
| return a string with a timestamp in the appropriate |
| format. |
| """ |
| raise NotImplementedError("_JobDirectory.get_timestamp_if_finished") |
| |
| def enqueue_offload(self, queue, age_limit): |
| """Enqueue the job for offload, if it's eligible. |
| |
| The job is eligible for offloading if the database has marked |
| it finished, and the job is older than the `age_limit` |
| parameter. |
| |
| If the job is eligible, offload processing is requested by |
| passing the `queue` parameter's `put()` method a sequence with |
| the job's `_dirname` attribute and its directory name. |
| |
| @param queue If the job should be offloaded, put the offload |
| parameters into this queue for processing. |
| @param age_limit Minimum age for a job to be offloaded. A value |
| of 0 means that the job will be offloaded as |
| soon as it is finished. |
| |
| """ |
| timestamp = self.get_timestamp_if_finished() |
| if not self._offload_count: |
| if not timestamp: |
| return |
| if not is_job_expired(age_limit, timestamp): |
| return |
| self._first_offload_start = time.time() |
| self._offload_count += 1 |
| if self.process_gs_instructions(): |
| queue.put([self._dirname, os.path.dirname(self._dirname), timestamp]) |
| |
| def is_offloaded(self): |
| """Return whether this job has been successfully offloaded.""" |
| return not os.path.exists(self._dirname) |
| |
| def get_failure_time(self): |
| """Return the time of the first offload failure.""" |
| return self._first_offload_start |
| |
| def get_failure_count(self): |
| """Return the number of times this job has failed to offload.""" |
| return self._offload_count |
| |
| def get_job_directory(self): |
| """Return the name of this job's results directory.""" |
| return self._dirname |
| |
| def process_gs_instructions(self): |
| """Process any gs_offloader instructions for this special task. |
| |
| @returns True/False if there is anything left to offload. |
| """ |
| # Default support is to still offload the directory. |
| return True |
| |
| |
| NO_OFFLOAD_README = """These results have been deleted rather than offloaded. |
| This is the expected behavior for passing jobs from the Commit Queue.""" |
| |
| |
| class RegularJobDirectory(_JobDirectory): |
| """Subclass of _JobDirectory for regular test jobs.""" |
| |
| GLOB_PATTERN = '[0-9]*-*' |
| |
| def process_gs_instructions(self): |
| """Process any gs_offloader instructions for this job. |
| |
| @returns True/False if there is anything left to offload. |
| """ |
| # Go through the gs_offloader instructions file for each test in this job. |
| for path in glob.glob(os.path.join(self._dirname, '*', |
| constants.GS_OFFLOADER_INSTRUCTIONS)): |
| with open(path, 'r') as f: |
| gs_off_instructions = json.load(f) |
| if gs_off_instructions.get(constants.GS_OFFLOADER_NO_OFFLOAD): |
| dirname = os.path.dirname(path) |
| shutil.rmtree(dirname) |
| os.mkdir(dirname) |
| breadcrumb_name = os.path.join(dirname, 'logs-removed-readme.txt') |
| with open(breadcrumb_name, 'w') as f: |
| f.write(NO_OFFLOAD_README) |
| |
| # Finally check if there's anything left to offload. |
| if not os.listdir(self._dirname): |
| shutil.rmtree(self._dirname) |
| return False |
| return True |
| |
| |
| def get_timestamp_if_finished(self): |
| """Get the timestamp to use for finished jobs. |
| |
| @returns the latest hqe finished_on time. If the finished_on times are null |
| returns the job's created_on time. |
| """ |
| entry = _AFE.get_jobs(id=self._id, finished=True) |
| if not entry: |
| return None |
| hqes = _AFE.get_host_queue_entries(finished_on__isnull=False, |
| job_id=self._id) |
| if not hqes: |
| return entry[0].created_on |
| # While most Jobs have 1 HQE, some can have multiple, so check them all. |
| return max([hqe.finished_on for hqe in hqes]) |
| |
| |
| class SpecialJobDirectory(_JobDirectory): |
| """Subclass of _JobDirectory for special (per-host) jobs.""" |
| |
| GLOB_PATTERN = 'hosts/*/[0-9]*-*' |
| |
| def __init__(self, resultsdir): |
| super(SpecialJobDirectory, self).__init__(resultsdir) |
| |
| def get_timestamp_if_finished(self): |
| entry = _AFE.get_special_tasks(id=self._id, is_complete=True) |
| return entry[0].time_finished if entry else None |