| # Copyright 2013 The ChromiumOS Authors |
| # Use of this source code is governed by a BSD-style license that can be |
| # found in the LICENSE file. |
| """A reproducing entity. |
| |
| Part of the Chrome build flags optimization. |
| |
| The Task class is used by different modules. Each module fills in the |
| corresponding information into a Task instance. Class Task contains the bit set |
| representing the flags selection. The builder module is responsible for filling |
| the image and the checksum field of a Task. The executor module will put the |
| execution output to the execution field. |
| """ |
| |
| __author__ = "[email protected] (Yuheng Long)" |
| |
| import os |
| import subprocess |
| import sys |
| from uuid import uuid4 |
| |
| |
| BUILD_STAGE = 1 |
| TEST_STAGE = 2 |
| |
| # Message indicating that the build or test failed. |
| ERROR_STRING = "error" |
| |
| # The maximum number of tries a build can have. Some compilations may fail due |
| # to unexpected environment circumstance. This variable defines how many tries |
| # the build should attempt before giving up. |
| BUILD_TRIES = 3 |
| |
| # The maximum number of tries a test can have. Some tests may fail due to |
| # unexpected environment circumstance. This variable defines how many tries the |
| # test should attempt before giving up. |
| TEST_TRIES = 3 |
| |
| |
| # Create the file/directory if it does not already exist. |
| def _CreateDirectory(file_name): |
| directory = os.path.dirname(file_name) |
| if not os.path.exists(directory): |
| os.makedirs(directory) |
| |
| |
| class Task(object): |
| """A single reproducing entity. |
| |
| A single test of performance with a particular set of flags. It records the |
| flag set, the image, the check sum of the image and the cost. |
| """ |
| |
| # The command that will be used in the build stage to compile the tasks. |
| BUILD_COMMAND = None |
| # The command that will be used in the test stage to test the tasks. |
| TEST_COMMAND = None |
| # The directory to log the compilation and test results. |
| LOG_DIRECTORY = None |
| |
| @staticmethod |
| def InitLogCommand(build_command, test_command, log_directory): |
| """Set up the build and test command for the task and the log directory. |
| |
| This framework is generic. It lets the client specify application specific |
| compile and test methods by passing different build_command and |
| test_command. |
| |
| Args: |
| build_command: The command that will be used in the build stage to compile |
| this task. |
| test_command: The command that will be used in the test stage to test this |
| task. |
| log_directory: The directory to log the compilation and test results. |
| """ |
| |
| Task.BUILD_COMMAND = build_command |
| Task.TEST_COMMAND = test_command |
| Task.LOG_DIRECTORY = log_directory |
| |
| def __init__(self, flag_set): |
| """Set up the optimization flag selection for this task. |
| |
| Args: |
| flag_set: The optimization flag set that is encapsulated by this task. |
| """ |
| |
| self._flag_set = flag_set |
| |
| # A unique identifier that distinguishes this task from other tasks. |
| self._task_identifier = uuid4() |
| |
| self._log_path = (Task.LOG_DIRECTORY, self._task_identifier) |
| |
| # Initiate the hash value. The hash value is used so as not to recompute it |
| # every time the hash method is called. |
| self._hash_value = None |
| |
| # Indicate that the task has not been compiled/tested. |
| self._build_cost = None |
| self._exe_cost = None |
| self._checksum = None |
| self._image = None |
| self._file_length = None |
| self._text_length = None |
| |
| def __eq__(self, other): |
| """Test whether two tasks are equal. |
| |
| Two tasks are equal if their flag_set are equal. |
| |
| Args: |
| other: The other task with which this task is tested equality. |
| Returns: |
| True if the encapsulated flag sets are equal. |
| """ |
| if isinstance(other, Task): |
| return self.GetFlags() == other.GetFlags() |
| return False |
| |
| def __hash__(self): |
| if self._hash_value is None: |
| # Cache the hash value of the flags, so as not to recompute them. |
| self._hash_value = hash(self._flag_set) |
| return self._hash_value |
| |
| def GetIdentifier(self, stage): |
| """Get the identifier of the task in the stage. |
| |
| The flag set uniquely identifies a task in the build stage. The checksum of |
| the image of the task uniquely identifies the task in the test stage. |
| |
| Args: |
| stage: The stage (build/test) in which this method is called. |
| Returns: |
| Return the flag set in build stage and return the checksum in test stage. |
| """ |
| |
| # Define the dictionary for different stage function lookup. |
| get_identifier_functions = { |
| BUILD_STAGE: self.FormattedFlags, |
| TEST_STAGE: self.__GetCheckSum, |
| } |
| |
| assert stage in get_identifier_functions |
| return get_identifier_functions[stage]() |
| |
| def GetResult(self, stage): |
| """Get the performance results of the task in the stage. |
| |
| Args: |
| stage: The stage (build/test) in which this method is called. |
| Returns: |
| Performance results. |
| """ |
| |
| # Define the dictionary for different stage function lookup. |
| get_result_functions = { |
| BUILD_STAGE: self.__GetBuildResult, |
| TEST_STAGE: self.GetTestResult, |
| } |
| |
| assert stage in get_result_functions |
| |
| return get_result_functions[stage]() |
| |
| def SetResult(self, stage, result): |
| """Set the performance results of the task in the stage. |
| |
| This method is called by the pipeling_worker to set the results for |
| duplicated tasks. |
| |
| Args: |
| stage: The stage (build/test) in which this method is called. |
| result: The performance results of the stage. |
| """ |
| |
| # Define the dictionary for different stage function lookup. |
| set_result_functions = { |
| BUILD_STAGE: self.__SetBuildResult, |
| TEST_STAGE: self.__SetTestResult, |
| } |
| |
| assert stage in set_result_functions |
| |
| set_result_functions[stage](result) |
| |
| def Done(self, stage): |
| """Check whether the stage is done. |
| |
| Args: |
| stage: The stage to be checked, build or test. |
| Returns: |
| True if the stage is done. |
| """ |
| |
| # Define the dictionary for different result string lookup. |
| done_string = { |
| BUILD_STAGE: self._build_cost, |
| TEST_STAGE: self._exe_cost, |
| } |
| |
| assert stage in done_string |
| |
| return done_string[stage] is not None |
| |
| def Work(self, stage): |
| """Perform the task. |
| |
| Args: |
| stage: The stage in which the task is performed, compile or test. |
| """ |
| |
| # Define the dictionary for different stage function lookup. |
| work_functions = {BUILD_STAGE: self.__Compile, TEST_STAGE: self.__Test} |
| |
| assert stage in work_functions |
| |
| work_functions[stage]() |
| |
| def FormattedFlags(self): |
| """Format the optimization flag set of this task. |
| |
| Returns: |
| The formatted optimization flag set that is encapsulated by this task. |
| """ |
| return str(self._flag_set.FormattedForUse()) |
| |
| def GetFlags(self): |
| """Get the optimization flag set of this task. |
| |
| Returns: |
| The optimization flag set that is encapsulated by this task. |
| """ |
| |
| return self._flag_set |
| |
| def __GetCheckSum(self): |
| """Get the compilation image checksum of this task. |
| |
| Returns: |
| The compilation image checksum of this task. |
| """ |
| |
| # The checksum should be computed before this method is called. |
| assert self._checksum is not None |
| return self._checksum |
| |
| def __Compile(self): |
| """Run a compile. |
| |
| This method compile an image using the present flags, get the image, |
| test the existent of the image and gathers monitoring information, and sets |
| the internal cost (fitness) for this set of flags. |
| """ |
| |
| # Format the flags as a string as input to compile command. The unique |
| # identifier is passed to the compile command. If concurrent processes are |
| # used to compile different tasks, these processes can use the identifier to |
| # write to different file. |
| flags = self._flag_set.FormattedForUse() |
| command = "%s %s %s" % ( |
| Task.BUILD_COMMAND, |
| " ".join(flags), |
| self._task_identifier, |
| ) |
| |
| # Try BUILD_TRIES number of times before confirming that the build fails. |
| for _ in range(BUILD_TRIES): |
| try: |
| # Execute the command and get the execution status/results. |
| p = subprocess.Popen( |
| command.split(), |
| stdout=subprocess.PIPE, |
| stderr=subprocess.PIPE, |
| ) |
| (out, err) = p.communicate() |
| |
| if out: |
| out = out.strip() |
| if out != ERROR_STRING: |
| # Each build results contains the checksum of the result image, the |
| # performance cost of the build, the compilation image, the length |
| # of the build, and the length of the text section of the build. |
| ( |
| checksum, |
| cost, |
| image, |
| file_length, |
| text_length, |
| ) = out.split() |
| # Build successfully. |
| break |
| |
| # Build failed. |
| cost = ERROR_STRING |
| except _: |
| # If there is exception getting the cost information of the build, the |
| # build failed. |
| cost = ERROR_STRING |
| |
| # Convert the build cost from String to integer. The build cost is used to |
| # compare a task with another task. Set the build cost of the failing task |
| # to the max integer. The for loop will keep trying until either there is a |
| # success or BUILD_TRIES number of tries have been conducted. |
| self._build_cost = sys.maxint if cost == ERROR_STRING else float(cost) |
| |
| self._checksum = checksum |
| self._file_length = file_length |
| self._text_length = text_length |
| self._image = image |
| |
| self.__LogBuildCost(err) |
| |
| def __Test(self): |
| """__Test the task against benchmark(s) using the input test command.""" |
| |
| # Ensure that the task is compiled before being tested. |
| assert self._image is not None |
| |
| # If the task does not compile, no need to test. |
| if self._image == ERROR_STRING: |
| self._exe_cost = ERROR_STRING |
| return |
| |
| # The unique identifier is passed to the test command. If concurrent |
| # processes are used to compile different tasks, these processes can use the |
| # identifier to write to different file. |
| command = "%s %s %s" % ( |
| Task.TEST_COMMAND, |
| self._image, |
| self._task_identifier, |
| ) |
| |
| # Try TEST_TRIES number of times before confirming that the build fails. |
| for _ in range(TEST_TRIES): |
| try: |
| p = subprocess.Popen( |
| command.split(), |
| stdout=subprocess.PIPE, |
| stderr=subprocess.PIPE, |
| ) |
| (out, err) = p.communicate() |
| |
| if out: |
| out = out.strip() |
| if out != ERROR_STRING: |
| # The test results contains the performance cost of the test. |
| cost = out |
| # Test successfully. |
| break |
| |
| # Test failed. |
| cost = ERROR_STRING |
| except _: |
| # If there is exception getting the cost information of the test, the |
| # test failed. The for loop will keep trying until either there is a |
| # success or TEST_TRIES number of tries have been conducted. |
| cost = ERROR_STRING |
| |
| self._exe_cost = sys.maxint if (cost == ERROR_STRING) else float(cost) |
| |
| self.__LogTestCost(err) |
| |
| def __SetBuildResult( |
| self, (checksum, build_cost, image, file_length, text_length) |
| ): |
| self._checksum = checksum |
| self._build_cost = build_cost |
| self._image = image |
| self._file_length = file_length |
| self._text_length = text_length |
| |
| def __GetBuildResult(self): |
| return ( |
| self._checksum, |
| self._build_cost, |
| self._image, |
| self._file_length, |
| self._text_length, |
| ) |
| |
| def GetTestResult(self): |
| return self._exe_cost |
| |
| def __SetTestResult(self, exe_cost): |
| self._exe_cost = exe_cost |
| |
| def LogSteeringCost(self): |
| """Log the performance results for the task. |
| |
| This method is called by the steering stage and this method writes the |
| results out to a file. The results include the build and the test results. |
| """ |
| |
| steering_log = "%s/%s/steering.txt" % self._log_path |
| |
| _CreateDirectory(steering_log) |
| |
| with open(steering_log, "w") as out_file: |
| # Include the build and the test results. |
| steering_result = ( |
| self._flag_set, |
| self._checksum, |
| self._build_cost, |
| self._image, |
| self._file_length, |
| self._text_length, |
| self._exe_cost, |
| ) |
| |
| # Write out the result in the comma-separated format (CSV). |
| out_file.write("%s,%s,%s,%s,%s,%s,%s\n" % steering_result) |
| |
| def __LogBuildCost(self, log): |
| """Log the build results for the task. |
| |
| The build results include the compilation time of the build, the result |
| image, the checksum, the file length and the text length of the image. |
| The file length of the image includes the length of the file of the image. |
| The text length only includes the length of the text section of the image. |
| |
| Args: |
| log: The build log of this task. |
| """ |
| |
| build_result_log = "%s/%s/build.txt" % self._log_path |
| |
| _CreateDirectory(build_result_log) |
| |
| with open(build_result_log, "w") as out_file: |
| build_result = ( |
| self._flag_set, |
| self._build_cost, |
| self._image, |
| self._checksum, |
| self._file_length, |
| self._text_length, |
| ) |
| |
| # Write out the result in the comma-separated format (CSV). |
| out_file.write("%s,%s,%s,%s,%s,%s\n" % build_result) |
| |
| # The build information about running the build. |
| build_run_log = "%s/%s/build_log.txt" % self._log_path |
| _CreateDirectory(build_run_log) |
| |
| with open(build_run_log, "w") as out_log_file: |
| # Write out the execution information. |
| out_log_file.write("%s" % log) |
| |
| def __LogTestCost(self, log): |
| """Log the test results for the task. |
| |
| The test results include the runtime execution time of the test. |
| |
| Args: |
| log: The test log of this task. |
| """ |
| |
| test_log = "%s/%s/test.txt" % self._log_path |
| |
| _CreateDirectory(test_log) |
| |
| with open(test_log, "w") as out_file: |
| test_result = (self._flag_set, self._checksum, self._exe_cost) |
| |
| # Write out the result in the comma-separated format (CSV). |
| out_file.write("%s,%s,%s\n" % test_result) |
| |
| # The execution information about running the test. |
| test_run_log = "%s/%s/test_log.txt" % self._log_path |
| |
| _CreateDirectory(test_run_log) |
| |
| with open(test_run_log, "w") as out_log_file: |
| # Append the test log information. |
| out_log_file.write("%s" % log) |
| |
| def IsImproved(self, other): |
| """Compare the current task with another task. |
| |
| Args: |
| other: The other task against which the current task is compared. |
| |
| Returns: |
| True if this task has improvement upon the other task. |
| """ |
| |
| # The execution costs must have been initiated. |
| assert self._exe_cost is not None |
| assert other.GetTestResult() is not None |
| |
| return self._exe_cost < other.GetTestResult() |