Hi All Does flink provide any ability to streamline logs being generated from a pipeline. How can we keep the logs from two pipelines separate so that its easy to debug the pipeline execution (something dynamic to automatically partition the logs per pipeline) Regards Sumit Chawla |
Hi Sumit, What exactly do you mean by pipeline? Are you talking about cases were multiple jobs are running concurrently on the same TaskManager, or are you referring to parallel instances of a Flink job? On Wed, Jul 13, 2016 at 9:49 PM, Chawla,Sumit <[hidden email]> wrote:
|
Hi Robert I actually mean both. Scenarios where multiple jobs are running on cluster, and same job could be running on multiple task managers. How can we make sure that each job logs to a different file so that Logs are not mixed, and its easy to debug a particular job. Something like Hadoop Yarn, where each attempt of a task produces a different log file. Regards Sumit Chawla On Thu, Jul 14, 2016 at 6:11 AM, Robert Metzger <[hidden email]> wrote:
|
Hi, I'm afraid that's not possible right now. The preferred way of running would be to have a Yarn cluster per job, that way you can isolate the logs. Cheers, Aljoscha On Thu, 14 Jul 2016 at 09:49 Chawla,Sumit <[hidden email]> wrote:
|
Free forum by Nabble | Edit this page |