pyspark.TaskContext.taskAttemptId#

TaskContext.taskAttemptId()[source]#

An ID that is unique to this task attempt (within the same SparkContext, no two task attempts will share the same attempt ID). This is roughly equivalent to Hadoop’s TaskAttemptID.

Returns
int

current task attempt id.