Merge two accumulable objects together
Merge two accumulable objects together
Normally, a user will not want to use this version, but will instead call +=
.
the other R
that will get merged with this
Add more data to this accumulator / accumulable
Add more data to this accumulator / accumulable
the data to add
Add more data to this accumulator / accumulable
Add more data to this accumulator / accumulable
the data to add
ID of this accumulator; for internal use only.
Get the current value of this accumulator from within a task.
Get the current value of this accumulator from within a task.
This is NOT the global value of the accumulator. To get the global value after a
completed operation on the dataset, call value
.
The typical use of this method is to directly mutate the local value, eg., to add an element to a Set.
Merge two accumulable objects together
Merge two accumulable objects together
Normally, a user will not want to use this version, but will instead call add
.
the other R
that will get merged with this
human-readable name for use in Spark's web UI
Set the accumulator's value.
Set the accumulator's value. For internal use only.
Access the accumulator's current value; only allowed on driver.
Set the accumulator's value; only allowed on driver.
A data type that can be accumulated, i.e. has a commutative and associative "add" operation, but where the result type,
R
, may be different from the element type being added,T
.You must define how to add data, and how to merge two of these together. For some data types, such as a counter, these might be the same operation. In that case, you can use the simpler org.apache.spark.Accumulator. They won't always be the same, though -- e.g., imagine you are accumulating a set. You will add items to the set, and you will union two sets together.
Operations are not thread-safe.
the full accumulated data (result type)
partial data that can be added in
(Since version 2.0.0) use AccumulatorV2