Bug 4274 - Parallel step instances should not stop when another step fails
Parallel step instances should not stop when another step fails
Product: jbatch
Classification: Unclassified
Component: source
All All
: P5 normal
: ---
Assigned To: cvignola
Depends on:
  Show dependency treegraph
Reported: 2012-11-09 20:16 UTC by mminella
Modified: 2013-01-16 16:07 UTC (History)
2 users (show)

See Also:


Note You need to log in before you can comment on or make changes to this bug.
Description mminella 2012-11-09 20:16:02 UTC
Section 5.5 indicates that when parallel steps in a job are executing and one fails, the other should be marked as STOPPED.  Why would the other step be marked as stopped if it is independent of the step that throws an error?  Why not let it run to completion?
Comment 1 waynexlund 2012-11-14 23:15:28 UTC
I agree with this assessment. The job would default to FAILED but the rest of the parallel steps may still succeed.
Comment 2 cvignola 2012-11-19 12:16:46 UTC
I've had multiple clients ask to stop partitioned execution completely and immediately if any one partition fails so they can trigger off of job failure to promptly commence problem resolution.  So that influenced me here.  You can see room for policy.  But I don't want to complicate matters.  I agree the parallel execution units should be allowed to complete.  There are sufficient listeners and partition callbacks for the user to interpose and issue a stop if they want to pre-empt processing.
Comment 3 cvignola 2013-01-16 16:07:20 UTC
the updated spec states active partitions are allowed to complete before the job is failed