From my understanding, the benchmark was done using Structured Streaming
that is still based on micro batching.
There are not throughput numbers for the new "Continuous Processing"
model Spark want to introduce. Only some latency numbers. Also note,
that the new "Continuous Processing" will not give exactly-once
semantics but only at-least-once (at least initially). Thus, there is
some tradeoff to make using "Continuous Processing" once it's available.
-Matthias
On 06/18/2017 03:51 PM, nragon wrote: