Hi.
We would like to do some batch analytics on our data set stored in Cassandra and are looking for an efficient way to load data from a single table. Not by key, but random 15%, 50% or 100%
Data bricks has create an efficient way to load Cassandra data into Apache Spark and they are doing it by reading from the underlying SS tables to load in parallel.
Do we have something similarly in Flink, or how is the most efficient way to load all, or many random data from a single Cassandra table into Flink?
Any suggestions and/or recommendations is highly appreciated.
Thanks in advance
Lasse Nedergaard