Hi to all,
I'm building a recommendation system to my application.
I have a set of logs (that contains the user info, the hour, the button that was clicked ect...) that arrive to my Flink by kafka, then I save every log in a HDFS (HADOOP), but know I have a problem, I want to apply ML to (all) my data.
I think in 2 scenarios:
First : Transform my DataStream in a DataSet and perform the ML task. It is possible?
Second : Preform a task in flink that get the data from Hadoop and perform the ML task.
What is the best way to do it?
I already check the IncrementalLearningSkeleton but I didn't understand how to apply that to an actual real case. Is there some complex example that I could look?
Another thing that I would like to ask is how to perform the second scenario, where I need to perform this task every hour, what it is the best way to do it?
Thanks,
Fábio Dias.