Reputation: 65
Currently, we are investigating how to effectively incorporate databricks latest feature for orchestration of tasks - Multi-task Jobs.
The default behaviour is that a downstream task would not be executed if the previous one has failed for some reason.
So the question is: Is it currently possible to have an onComplete status (similar to those in Azure Data Factory or SQL Server Integration services-SSIS) that regardsless of the task success or failure we can continue with the workflow and execute the next tasks.
Upvotes: 2
Views: 757
Reputation: 87119
Right now it's not possible, but you can just surround your code with try/catch, so if error is thrown then catch block won't propagate exception further, and task won't be marked as failed.
Upvotes: 2