-
Notifications
You must be signed in to change notification settings - Fork 10
Description
Currently, the workflow executes all the pipelines, scores the performance, and returns the results. However, if there are many algorithms and many datasets, then this might take a long time.
It would be benificial to do some checkpointing in the workflow, for example every 100 jobs are the results-so-far saved, and then the workflow is continued. How often to be saved can be passed as an argument to the constructor of the Workflow. In this regard, it might also be good to save the results automatically in the workflow (maybe decided on a hyperparameter), instead of only returning the results, to ensure a unified format.
In addition, some methods to decide which jobs to still execute can help to restart the process if some problem occurred. In addition, some methods can be created to obtain all the jobs with an error in the workflow.