Source code for automan.automation

from __future__ import print_function

from fnmatch import fnmatch
import glob
import json
import os
import shlex
import shutil
import sys
import time
import traceback

from .jobs import Job


[docs]class Task(object): """Basic task to run. Subclass this to do whatever is needed. This class is very similar to luigi's Task class. """ def __init__(self, depends=None): # Depends is a list and available for all tasks. self.depends = depends if depends is not None else []
[docs] def complete(self): """Should return True/False indicating success of task. If the task was just executed (in this invocation) but failed, raise any Exception that is a subclass of Exception as this signals an error to the task execution engine. If the task was executed in an earlier invocation of the automation, then just return True/False so as to be able to re-run the simulation. """ return all([os.path.exists(x) for x in self.output()])
[docs] def output(self): """Return list of output paths. """ return []
[docs] def run(self, scheduler): """Run the task, using the given scheduler. Using the scheduler is optional but recommended for any long-running tasks. It is safe to raise an exception immediately when running the task but for long running tasks, the exception will not matter and the `complete` method should do. """ pass
[docs] def requires(self): """Return iterable of tasks this task requires. It is important that one either return tasks that are idempotent or return the same instance as this method is called repeatedly. """ return self.depends
[docs]class WrapperTask(Task): """A task that wraps other tasks and is done when all its requirements are done. """
[docs] def complete(self): return all(r.complete() for r in self.requires())
[docs]class TaskRunner(object): """Run given tasks using the given scheduler. """ def __init__(self, tasks, scheduler): """Constructor. **Parameters** tasks: iterable of `Task` instances. scheduler: `automan.jobs.Scheduler` instance """ self.scheduler = scheduler self.todo = [] self.task_status = dict() self.task_outputs = set() self.repeat_tasks = set() for task in tasks: self.add_task(task) # #### Private protocol ############################################## def _check_error_in_running_tasks(self): running = self._get_tasks_with_status('running') for task in running: if self._check_status_of_task(task) == 'error': return True return False def _check_status_of_requires(self, task): status = [self._check_status_of_task(t) for t in task.requires()] if 'error' in status: return 'error' if all(x is True for x in status): return 'done' else: return 'running' def _check_status_of_task(self, task): status = self.task_status.get(task) if status == 'not started': return False elif status == 'done': return True else: complete = False try: complete = task.complete() self.task_status[task] = 'done' if complete else 'running' except Exception: complete = 'error' self.task_status[task] = 'error' return complete def _get_tasks_with_status(self, status): return [ t for t, s in self.task_status.items() if s == status and t not in self.repeat_tasks ] def _is_output_registered(self, task): # Note, this has a side-effect of registering the task's output # when called. output = task.output() output_str = str(output) if output and output_str in self.task_outputs: self.repeat_tasks.add(task) return True else: if output: self.task_outputs.add(output_str) return False def _run(self, task): try: print("\nRunning task %s..." % task) self.task_status[task] = 'running' task.run(self.scheduler) status = 'running' except Exception: traceback.print_exc() status = 'error' self.task_status[task] = 'error' return status def _show_remaining_tasks(self, replace_line=False): start, end = ('\r', '') if replace_line else ('', '\n') running = self._get_tasks_with_status('running') print("{start}{pending} tasks pending and {running} tasks running". format( start=start, pending=len(self.todo), running=len(running) ), end=end) sys.stdout.flush() def _wait_for_running_tasks(self, wait): print("\nWaiting for already running tasks...") running = self._get_tasks_with_status('running') while len(running) > 0: for t in running: self._check_status_of_task(t) time.sleep(wait) running = self._get_tasks_with_status('running') errors = self._get_tasks_with_status('error') n_err = len(errors) print("{n_err} jobs had errors.".format(n_err=n_err)) return n_err # #### Public protocol ##############################################
[docs] def add_task(self, task): if task in self.task_status or self._is_output_registered(task): # This task is already added or another task produces exactly # the same output, so do nothing. return if not task.complete(): self.todo.append(task) self.task_status[task] = 'not started' for req in task.requires(): self.add_task(req) else: self.task_status[task] = 'done'
[docs] def run(self, wait=5): '''Run the tasks that were given. Wait for the given amount of time to poll for completed tasks. Returns the number of tasks that had errors. ''' self._show_remaining_tasks() status = 'running' while len(self.todo) > 0 and status != 'error': to_remove = [] for i in range(len(self.todo) - 1, -1, -1): task = self.todo[i] status = self._check_status_of_requires(task) if self._check_error_in_running_tasks(): status = 'error' if status == 'error': break elif status == 'done': to_remove.append(task) status = self._run(task) for task in to_remove: self.todo.remove(task) if len(self.todo) > 0: self._show_remaining_tasks(replace_line=True) time.sleep(wait) n_errors = self._wait_for_running_tasks(wait) if n_errors == 0: print("Finished!") else: print("Please fix the issues and re-run.") return n_errors
[docs]class CommandTask(Task): """Convenience class to run a command via the framework. The class provides a method to run the simulation and also check if the simulation is completed. The command should ideally produce all of its outputs inside an output directory that is specified. """ def __init__(self, command, output_dir, job_info=None, depends=None): """Constructor **Parameters** command: str or list: command to run; $output_dir is substituted. output_dir: str : path of output directory. job_info: dict: dictionary of job information. depends: list: list of tasks this depends on. """ super().__init__(depends=depends) if isinstance(command, str): self.command = shlex.split(command) else: self.command = command self.command = [x.replace('$output_dir', output_dir) for x in self.command] self.output_dir = output_dir self.job_info = job_info if job_info is not None else {} self.job_proxy = None self._copy_proc = None # This is a sentinel set to true when the job is finished # the data is copied to a local machine and cleaned on the remote. self._finished = False # This file will be created if the job exited with an error. self._error_status_file = os.path.join( self.output_dir, 'command_exited_with_error' ) self._job = None def __str__(self): return ('%s, output in: %s ' % (self.__class__.__name__, self.output_dir)) # #### Public protocol ###########################################
[docs] def complete(self): """Should return True/False indicating success of task. """ job_proxy = self.job_proxy if job_proxy is None: return self._is_done() elif self._finished: if os.path.exists(self._error_status_file): raise RuntimeError( 'Error in task with output in %s.' % self.output_dir ) return True else: return self._copy_output_and_check_status()
[docs] def run(self, scheduler): # Remove the error status file if it exists and we are going to run. if os.path.exists(self._error_status_file): os.remove(self._error_status_file) self.job_proxy = scheduler.submit(self.job)
[docs] def clean(self): """Clean out any generated results. This completely removes the output directory. """ if os.path.exists(self.output_dir): shutil.rmtree(self.output_dir)
[docs] def output(self): """Return list of output paths. """ return [self.output_dir]
[docs] def requires(self): return self.depends
# #### Private protocol ########################################### @property def job(self): if self._job is None: self._job = Job( command=self.command, output_dir=self.output_dir, **self.job_info ) return self._job def _is_done(self): """Returns True if the simulation completed. """ if (not os.path.exists(self.output_dir)) \ or os.path.exists(self._error_status_file): return False else: return self.job.status() == 'done' def _check_if_copy_complete(self): proc = self._copy_proc if proc is None: # Local job so no copy needed. return True else: if proc.poll() is None: return False else: if self.job_proxy is not None: self.job_proxy.clean() self._finished = True return True def _copy_output_and_check_status(self): jp = self.job_proxy status = jp.status() if status == 'done': if self._copy_proc is None: self._copy_proc = jp.copy_output('.') return self._check_if_copy_complete() elif status == 'error': cmd = ' '.join(self.command) msg = '\n***************** ERROR *********************\n' msg += 'On host %s Job %s failed!' % (jp.worker.host, cmd) print(msg) print(jp.get_stderr()) proc = jp.copy_output('.') if proc is not None: proc.wait() jp.clean() print('***************** ERROR **********************') with open(self._error_status_file, 'w') as fp: fp.write('') self._finished = True raise RuntimeError(msg) return False
[docs]class PySPHTask(CommandTask): """Convenience class to run a PySPH simulation via an automation framework. This task automatically adds the output directory specification for pysph so users to not need to add it. """ def __init__(self, command, output_dir, job_info=None, depends=None): """Constructor **Parameters** command: str or list: command to run; $output_dir is substituted. output_dir: str : path of output directory. job_info: dict: dictionary of job information. depends: list: list of tasks this depends on. """ super(PySPHTask, self).__init__(command, output_dir, job_info, depends) self.command += ['-d', output_dir] # #### Private protocol ########################################### def _is_done(self): """Returns True if the simulation completed. """ if not os.path.exists(self.output_dir): return False job_status = self.job.status() if job_status == 'error': # If job information exists, it trumps everything else # as it stores the process exit status which is usually # a much better indicator of the job status. return False else: info_fname = self._get_info_filename() if not info_fname or not os.path.exists(info_fname): return False with open(info_fname) as fp: d = json.load(fp) return d.get('completed') def _get_info_filename(self): files = glob.glob(os.path.join(self.output_dir, '*.info')) if len(files) > 0: return files[0] else: return None
[docs]class FileCommandTask(CommandTask): """Convenience class to run a command which produces as output one or more files. The difference from the CommandTask is that this does not place its outputs in a separate directory. """ def __init__(self, command, files, job_info=None, depends=None): """Constructor **Parameters** command: str or list: command to run; $output_dir is substituted. output_dir: str : path of output directory. files: list(str): relative paths of output files. job_info: dict: dictionary of job information. depends: list: list of tasks this depends on. """ self.files = files output_dir = os.path.join(files[0] + '.job_info') super().__init__( command, output_dir, job_info=job_info, depends=depends )
[docs] def clean(self): """Clean out any generated results. This completely removes the output directory. """ if os.path.exists(self.output_dir): shutil.rmtree(self.output_dir) for f in self.files: if os.path.exists(f): os.remove(f)
[docs] def output(self): """Return list of output paths. """ return self.files
[docs]class Problem(object): """This class represents a numerical problem or computational problem of interest that needs to be solved. The class helps one run a variety of commands (or simulations), and then assemble/compare the results from those in the `run` method. This is perhaps easily understood with an example. Let us say one wishes to run the elliptical drop example problem with the standard SPH and TVF and compare the results and their convergence properties while also keep track of the computational time. To do this one will have to run several simulations, then collect and process the results. This is achieved by subclassing this class and implementing the following methods: - `get_name(self)`: returns a string of the name of the problem. All results and simulations are collected inside a directory with this name. - `get_commands(self)`: returns a sequence of (directory_name, command_string, job_info, depends) tuples. These are to be executed before the `run` method is called. - `get_requires(self)`: returns a sequence of (name, task) tuples. These are to be exeuted before the `run` method is called. - `run(self)`: Processes the completed simulations to make plots etc. See the `EllipticalDrop` example class below to see a full implementation. """ # The Task class to create for the cases, change to suit your needs. task_cls = CommandTask def __init__(self, simulation_dir, output_dir): """Constructor. **Parameters** simulation_dir : str : directory where simulation output goes. output_dir : str : directory where outputs from `run` go. """ self.out_dir = output_dir self.sim_dir = simulation_dir # Setup the simulation instances in the cases. self.cases = None self.setup() def _make_depends(self, depends): if not depends: return [] deps = [] for x in depends: if isinstance(x, Task): deps.append(x) elif isinstance(x, Simulation): if x.depends: my_depends = self._make_depends(x.depends) else: my_depends = None task = self.task_cls( x.command, self.input_path(x.name), x.job_info, depends=my_depends ) deps.append(task) else: raise RuntimeError( 'Invalid dependency: {0} for problem {1}'.format( x, self ) ) return deps # #### Public protocol ###########################################
[docs] def input_path(self, *args): """Given any arguments, relative to the simulation dir, return the absolute path. """ return os.path.join(self.sim_dir, self.get_name(), *args)
simulation_path = input_path
[docs] def output_path(self, *args): """Given any arguments relative to the output_dir return the absolute path. """ return os.path.join(self.out_dir, self.get_name(), *args)
[docs] def setup(self): """Called by init, so add any initialization here. """ pass
[docs] def make_output_dir(self): """Convenience to make the output directory if needed. """ base = self.output_path() if not os.path.exists(base): os.makedirs(base)
[docs] def get_name(self): """Return the name of this problem, this name is used as a directory for the simulation and the outputs. """ # Return a sane default instead of forcing the user to do this. return self.__class__.__name__
[docs] def get_commands(self): """Return a sequence of (name, command_string, job_info_dict) or (name, command_string, job_info_dict, depends). The name represents the command being run and is used as a subdirectory for generated output. The command_string is the command that needs to be run. The job_info_dict is a dictionary with any additional info to be used by the job, these are additional arguments to the `automan.jobs.Job` class. It may be None if nothing special need be passed. The depends is any dependencies this simulation has in terms of other simulations/tasks. """ if self.cases is not None: return [ (x.name, x.command, x.job_info, x.depends) for x in self.cases ] else: return []
[docs] def get_requires(self): """Return a sequence of tuples of form (name, task). The name represents the command being run and is used as a subdirectory for generated output. The task is a `automan.automation.Task` instance. """ base = self.get_name() result = [] for cmd_info in self.get_commands(): name, cmd, job_info = cmd_info[:3] deps = cmd_info[3] if len(cmd_info) == 4 else [] sim_output_dir = self.input_path(name) depends = self._make_depends(deps) task = self.task_cls( cmd, sim_output_dir, job_info, depends=depends ) task_name = '%s.%s' % (base, name) result.append((task_name, task)) return result
[docs] def get_outputs(self): """Get a list of outputs generated by this problem. By default it returns the output directory (as a single element of a list). """ return [self.output_path()]
[docs] def run(self): """Run any analysis code for the simulations completed. This is usually run after the simulation commands are completed. """ pass
[docs] def clean(self): """Cleanup any generated output from the analysis code. This does not clean the output of any nested commands. """ for path in self.get_outputs(): if os.path.exists(path): if os.path.isdir(path): shutil.rmtree(path) elif os.path.isfile(path): os.remove(path)
[docs]class PySPHProblem(Problem): task_cls = PySPHTask
[docs]def key_to_option(key): """Convert a dictionary key to a valid command line option. This simply replaces underscores with dashes. """ return key.replace('_', '-')
[docs]def kwargs_to_command_line(kwargs): """Convert a dictionary of keyword arguments to a list of command-line options. If the value of the key is None, no value is passed. **Examples** >>> sorted(kwargs_to_command_line(dict(some_arg=1, something_else=None))) ['--some-arg=1', '--something-else'] """ cmd_line = [] for key, value in kwargs.items(): option = key_to_option(key) if value is None: arg = "--{option}".format(option=option) else: arg = "--{option}={value}".format( option=option, value=str(value) ) cmd_line.append(arg) return cmd_line
[docs]class Simulation(object): """A convenient class to abstract code for a particular simulation. Simulation objects are typically created by ``Problem`` instances in order to abstract and simulate repetitive code for a particular simulation. For example if one were comparing the elliptical_drop example, one could instantiate a Simulation object as follows:: >>> s = Simlation('outputs/sph', 'pysph run elliptical_drop') One can pass any additional command line arguments as follows:: >>> s = Simlation( ... 'outputs/sph', 'pysph run elliptical_drop', timestep=0.005 ... ) >>> s.command 'pysph run elliptical_drop --timestep=0.001' >>> s.input_path('results.npz') 'outputs/sph/results.npz' The extra parameters can be used to filter and compare different simulations. One can define additional plot methods for a particular subclass and use these to easily plot results for different cases. One can also pass any additional parameters to the `automan.jobs.Job` class via the job_info kwarg so as to run the command suitably. For example:: >>> s = Simlation('outputs/sph', 'pysph run elliptical_drop', ... job_info=dict(n_thread=4)) The object has other methods that are convenient when comparing plots. Along with the ``compare_cases``, ``filter_cases`` and ``filter_by_name`` this is an extremely powerful way to automate and compare results. """ def __init__(self, root, base_command, job_info=None, depends=None, **kw): """Constructor **Parameters** root: str Path to simulation output directory. base_command: str Base command to run. job_info: dict Extra arguments to the `automan.jobs.Job` class. depends: list List of other simulations/tasks this simulation depends on. **kw: dict Additional parameters to pass to command. """ self.root = root self.name = os.path.basename(root) self.base_command = base_command self.job_info = job_info self.depends = depends if depends is not None else [] self.params = dict(kw) self._results = None
[docs] def input_path(self, *args): """Given any arguments, relative to the simulation dir, return the absolute path. """ return os.path.join(self.root, *args)
@property def command(self): return self.base_command + ' ' + self.get_command_line_args() @property def data(self): if self._results is None: import numpy self._results = numpy.load(self.input_path('results.npz')) return self._results
[docs] def get_labels(self, labels): render = self.render_parameter if isinstance(labels, str): return render(labels) else: s = [render(x) for x in labels] s = [x for x in s if len(x) > 0] return r', '.join(s)
[docs] def kwargs_to_command_line(self, kwargs): return kwargs_to_command_line(kwargs)
[docs] def get_command_line_args(self): return ' '.join(self.kwargs_to_command_line(self.params))
[docs] def render_parameter(self, param): """Return string to be used for labels for given parameter. """ if param not in self.params: return '' value = self.params[param] if value is None: return r'%s' % param else: return r'%s=%s' % (param, self.params[param])
############################################################################ # Convenient classes that can be used to easily automate a collection # of problems.
[docs]class SolveProblem(Task): """Solves a particular `Problem`. This runs all the commands that the problem requires and then runs the problem instance's run method. The match argument is a string which when provided helps run only a subset of the requirements for the problem. The force argument specifies that the problem should be cleaned, so as to re-run any post-processing. """ def __init__(self, problem, match='', force=False, depends=None): super().__init__(depends=depends) self.problem = problem self.match = match self.force = force if self.force: self.problem.clean() self._requires = [ self._make_task(task) for name, task in self.problem.get_requires() if len(match) == 0 or fnmatch(name, match) ] def _make_task(self, obj): if isinstance(obj, Task): return obj elif isinstance(obj, Problem): return SolveProblem( problem=obj, match=self.match, force=self.force ) elif isinstance(obj, type) and issubclass(obj, Problem): problem = obj(self.problem.sim_dir, self.problem.out_dir) return SolveProblem( problem=problem, match=self.match, force=self.force ) else: raise RuntimeError( 'Unknown requirement: {0}, for problem: {1}.'.format( obj, self.problem ) ) def __str__(self): return 'Problem named %s' % self.problem.get_name()
[docs] def complete(self): if len(self.match) == 0: return super(SolveProblem, self).complete() else: return all(r.complete() for r in self.requires())
[docs] def output(self): return self.problem.get_outputs()
[docs] def run(self, scheduler): if len(self.match) == 0: self.problem.run()
[docs] def requires(self): return self._requires + self.depends
[docs]class RunAll(WrapperTask): """Solves a given collection of problems. """ def __init__(self, simulation_dir, output_dir, problem_classes, force=False, match='', depends=None): super().__init__(depends=depends) self.simulation_dir = simulation_dir self.output_dir = output_dir self.force = force self.match = match self.problems = self._make_problems(problem_classes) self._requires = self._get_requires() # #### Private protocol ############################################### def _get_requires(self): return [ SolveProblem(problem=x, match=self.match, force=self.force) for x in self.problems ] def _make_problems(self, problem_classes): problems = [] for klass in problem_classes: problem = klass(self.simulation_dir, self.output_dir) problems.append(problem) return problems # #### Public protocol ################################################
[docs] def requires(self): return self._requires + self.depends
[docs]class Automator(object): """Main class to automate a collection of problems. This processess command line options and runs all tasks with a scheduler that is configured using the ``config.json`` file if it is present. Here is typical usage:: >>> all_problems = [EllipticalDrop] >>> automator = Automator('outputs', 'figures', all_problems) >>> automator.run() The class also creates a `automan.cluster_manager.ClusterManager` instance and integrates the cluster management features as well. This allows a user to automate their results across a collection of remote machines accessible only by ssh. """ def __init__(self, simulation_dir, output_dir, all_problems, cluster_manager_factory=None): """Constructor. **Parameters** simulation_dir : str Root directory to generate simulation results in. output_dir: str Root directory where outputs will be generated by Problem instances. all_problems: sequence of `Problem` classes. Sequence of problem classes to automate. cluster_manager_factory: callable Callable should return `cluster_manager.ClusterManager` instance. None will use the default one. """ self.simulation_dir = simulation_dir self.output_dir = output_dir self.all_problems = all_problems self.named_tasks = {} self.tasks = [] self.post_proc_tasks = [] self.runner = None self.cluster_manager = None self.runall_task = None self._args = None if cluster_manager_factory is None: from automan.cluster_manager import ClusterManager self.cluster_manager_factory = ClusterManager else: self.cluster_manager_factory = cluster_manager_factory # #### Public Protocol ########################################
[docs] def add_task(self, task, name=None, post_proc=False): """Add a task or a problem instance to also execute. If the `name` is specified then it is a treated as a named task wherein it must be only invoked explicitly via the command line when asked. If `post_proc` is True then the task is given an additional dependency if possible such that the task is run after the `RunAll` task is completed. **Parameters** task: Task or Problem instance: Task or Problem to add. name: str: name of the task (optional). post_proc: bool: Add a dependency to the task with the RunAll task. """ if isinstance(task, type) and issubclass(task, Problem): p = task( simulation_dir=self.simulation_dir, output_dir=self.output_dir ) _task = SolveProblem(p) elif isinstance(task, Problem): _task = SolveProblem(task) elif isinstance(task, Task): _task = task else: raise ValueError( 'Invalid task: must be Problem class/instance or Task.' ) if name is not None: self.named_tasks[name] = _task else: self.tasks.append(_task) if post_proc: self.post_proc_tasks.append(_task)
[docs] def run(self, argv=None): """Start the automation. """ self._setup(argv) self._setup_tasks() self.runner.run()
# #### Private Protocol ######################################## def _check_positional_arguments(self, problems): names = [c.__name__ for c in self.all_problems] + ['all'] lower_names = [x.lower() for x in names] lower_names.extend(list(self.named_tasks.keys())) for p in problems: if p.lower() not in lower_names: print("ERROR: %s not a valid problem/task!" % p) print("Valid names are %s" % ', '.join(names)) self.parser.exit(1) def _get_exclude_paths(self): """Returns a list of exclude paths suitable for passing on to rsync to exclude syncing some directories on remote machines. """ paths = [] for path in [self.simulation_dir, self.output_dir]: if not path.endswith('/'): paths.append(path + '/') return paths def _parse_args(self, argv): '''Parse command line arguments. Override this when the CLI arguments are customized. ''' self._args = self.parser.parse_args(argv) return self._args def _select_problem_classes(self, problems): if 'all' in problems: return self.all_problems else: lower_names = [x.lower() for x in problems] return [cls for cls in self.all_problems if cls.__name__.lower() in lower_names] def _setup(self, argv): if self.runner is None: self._setup_argparse() args = self._parse_args(argv) self._check_positional_arguments(args.problem) self.cluster_manager = self.cluster_manager_factory( config_fname=args.config, exclude_paths=self._get_exclude_paths() ) from .cluster_manager import BootstrapError if len(args.host) > 0: try: self.cluster_manager.add_worker( args.host, args.home, args.nfs ) except BootstrapError: pass return elif len(args.host) == 0 and args.update_remote: self.cluster_manager.update(not args.no_rebuild) elif len(args.rm_remote_output) > 0: self.cluster_manager.delete( self.simulation_dir, args.rm_remote_output) problem_classes = self._select_problem_classes(args.problem) task = RunAll( simulation_dir=self.simulation_dir, output_dir=self.output_dir, problem_classes=problem_classes, force=args.force, match=args.match ) self.runall_task = task self.scheduler = self.cluster_manager.create_scheduler() self.runner = TaskRunner([task], self.scheduler) def _setup_argparse(self): import argparse desc = "Automation script to run simulations." parser = argparse.ArgumentParser( description=desc ) all_problem_names = [c.__name__ for c in self.all_problems] all_problem_names += list(self.named_tasks.keys()) + ['all'] parser.add_argument( 'problem', nargs='*', default=["all"], help="Specifies problem/task to run as a string " "(case-insensitive), valid names are %s. " "Defaults to running all of the problems." % all_problem_names ) parser.add_argument( '-a', '--add-node', action="store", dest="host", type=str, default='', help="Add a new remote worker." ) parser.add_argument( '-c', '--config', action="store", dest="config", default="config.json", help="Configuration file to use." ) parser.add_argument( '--home', action="store", dest="home", type=str, default='', help='Home directory of the remote worker (to be used with -a)' ) parser.add_argument( '--nfs', action="store_true", dest="nfs", default=False, help=('Does the remote remote worker share the filesystem ' '(to be used with -a)') ) parser.add_argument( '-f', '--force', action="store_true", default=False, dest='force', help='Redo the plots even if they were already made.' ) parser.add_argument( '-m', '--match', action="store", type=str, default='', dest='match', help="Name of the problem to run (uses fnmatch)" ) parser.add_argument( '--no-rebuild', action="store_true", dest="no_rebuild", default=False, help="Do not rebuild the sources on update, just update the files." ) parser.add_argument( '-u', '--update-remote', action='store_true', dest='update_remote', default=False, help='Update remote worker machines.' ) parser.add_argument( '--rm-remote-output', nargs='*', action="store", dest="rm_remote_output", type=str, default='', help="remove output folder from the mentioned machines use" "'all' to remove from all host (except localhost and host where " "nfs is true)" ) self.parser = parser def _setup_tasks(self): for task in self.post_proc_tasks: task.depends.append(self.runall_task) # Add generic tasks. for task in self.tasks: self.runner.add_task(task) # Add named tasks only if specifically requested on CLI. for name, task in self.named_tasks.items(): if name in self._args.problem: self.runner.add_task(task) # Reset the tasks so we can use the automator interactively. self.post_proc_tasks = [] self.tasks = [] self.named_tasks = {}