зеркало из https://github.com/microsoft/spark.git
Коммит
e3ae98b54e
|
@ -49,7 +49,16 @@ class Accumulable[R, T] (
|
|||
else throw new UnsupportedOperationException("Can't read accumulator value in task")
|
||||
}
|
||||
|
||||
private[spark] def localValue = value_
|
||||
/**
|
||||
* Get the current value of this accumulator from within a task.
|
||||
*
|
||||
* This is NOT the global value of the accumulator. To get the global value after a
|
||||
* completed operation on the dataset, call `value`.
|
||||
*
|
||||
* The typical use of this method is to directly mutate the local value, eg., to add
|
||||
* an element to a Set.
|
||||
*/
|
||||
def localValue = value_
|
||||
|
||||
def value_= (r: R) {
|
||||
if (!deserialized) value_ = r
|
||||
|
|
|
@ -112,4 +112,20 @@ class AccumulatorSuite extends FunSuite with ShouldMatchers with BeforeAndAfter
|
|||
sc.stop()
|
||||
}
|
||||
}
|
||||
|
||||
test ("localValue readable in tasks") {
|
||||
import SetAccum._
|
||||
val maxI = 1000
|
||||
for (nThreads <- List(1, 10)) { //test single & multi-threaded
|
||||
val sc = new SparkContext("local[" + nThreads + "]", "test")
|
||||
val acc: Accumulable[mutable.Set[Any], Any] = sc.accumulable(new mutable.HashSet[Any]())
|
||||
val groupedInts = (1 to (maxI/20)).map {x => (20 * (x - 1) to 20 * x).toSet}
|
||||
val d = sc.parallelize(groupedInts)
|
||||
d.foreach {
|
||||
x => acc.localValue ++= x
|
||||
}
|
||||
acc.value should be ( (0 to maxI).toSet)
|
||||
}
|
||||
}
|
||||
|
||||
}
|
||||
|
|
Загрузка…
Ссылка в новой задаче