Is there a Flink DataSet equivalent to Spark's RDD.persist?

classic Classic list List threaded Threaded
2 messages Options
Reply | Threaded
Open this post in threaded view
|

Is there a Flink DataSet equivalent to Spark's RDD.persist?

Frank Grimes
Hi,

I'm trying to port an existing Spark job to Flink and have gotten stuck on the same issue brought up here:


Is there some way to accomplish this same thing in Flink?
i.e. avoid re-computing a particular DataSet when multiple different subsequent transformations are required on it.

I've even tried explicitly writing out the DataSet to avoid the re-computation but still taking an I/O hit for the initial write to HDFS and subsequent re-reading of it in the following stages. 
While it does yield a performance improvement over no caching at all, it doesn't match the performance I get with RDD.persist in Spark.

Thanks,

Frank Grimes
Reply | Threaded
Open this post in threaded view
|

Re: Is there a Flink DataSet equivalent to Spark's RDD.persist?

Andrey Zagrebin-3
Hi Frank,

This feature is currently under discussion. You can follow it in this issue:
https://issues.apache.org/jira/browse/FLINK-11199

Best,
Andrey

On Thu, Feb 21, 2019 at 7:41 PM Frank Grimes <[hidden email]> wrote:
Hi,

I'm trying to port an existing Spark job to Flink and have gotten stuck on the same issue brought up here:


Is there some way to accomplish this same thing in Flink?
i.e. avoid re-computing a particular DataSet when multiple different subsequent transformations are required on it.

I've even tried explicitly writing out the DataSet to avoid the re-computation but still taking an I/O hit for the initial write to HDFS and subsequent re-reading of it in the following stages. 
While it does yield a performance improvement over no caching at all, it doesn't match the performance I get with RDD.persist in Spark.

Thanks,

Frank Grimes