RDD.
barrier
Marks the current stage as a barrier stage, where Spark must launch all tasks together. In case of a task failure, instead of only restarting the failed task, Spark will abort the entire stage and relaunch all tasks for this stage. The barrier execution mode feature is experimental and it only handles limited scenarios. Please read the linked SPIP and design docs to understand the limitations and future plans.
New in version 2.4.0.
RDDBarrier
instance that provides actions within a barrier stage.
See also
pyspark.BarrierTaskContext
Notes
For additional information see
SPIP: Barrier Execution Mode
Design Doc
This API is experimental