Hello – I’m working on an anomaly detector for some time series monitoring data. I’ve setup an example project with Flink that reads from Kafka to get the monitoring data. Unfortunately, I’m not sure what to do next. The goal is to perform some clustering on the metric values that Flink is receiving and detect when the value[s] are anomalous. I’ve got a DataStream<MetricDefinition> that I think needs to go through some pre-processing, like transforming it into a vector, but I’m not sure how to proceed. The pojo[MetricDefinition] looks like this – Can anyone point me in the right direction? Thanks! Jeremy D. Branham
Technology Architect - Sprint
#gettingbettereveryday This e-mail may contain Sprint proprietary information intended for the sole use of the recipient(s). Any use by others is prohibited. If you are not the intended recipient, please contact the sender and delete all copies of the message. |
This post was updated on .
Hello Jeremy,
it looks like what you are looking for is map (1 in, 1 out) / flatmap (1 in, 0-n out) for preprocessing on a single element basis as well as windows for looking at related MetricDefinition elements calculating some result. I suggest you look into Windows (https://ci.apache.org/projects/flink/flink-docs-release-1.3/dev/windows.html) and basic transformations (https://ci.apache.org/projects/flink/flink-docs-release-1.3/dev/datastream_api.html#datastream-transformations). EDIT: I must have missed the ML part of your question :) I have no idea about ML but my advice for Flink (the non-ML part) is still applicable :) Regards, Jonas |
Free forum by Nabble | Edit this page |