Hi everyone,
I am running some Flink experiments with Peel benchmark http://peel-framework.org/ and I am struggling with exceptions: the environment is a 25-nodes cluster, 16 cores per nodes. The dataset is ~80GiB and is located on Hdfs 2.7.1. At the beginning I tried with 400 as degree of parallelism and with the following configuration: jobmanager.rpc.address = ${runtime.hostname} akka.log.lifecycle.events = ON akka.ask.timeout = 300s jobmanager.rpc.port = 6002 jobmanager.heap.mb = 1024 jobmanager.web.port = 6004 taskmanager.heap.mb = 28672 taskmanager.memory.fraction = 0.7 taskmanager.network.numberOfBuffers = 32768 taskmanager.network.bufferSizeInBytes = 16384 taskmanager.tmp.dirs = "/data/1/peel/flink/tmp:/data/2/peel/flink/tmp:/data/3/peel/flink/tmp:/data/4/peel/flink/tmp" taskmanager.debug.memory.startLogThread = true the following exception will raise Caused by: java.io.IOException: Insufficient number of network buffers: required 350, but only 317 available. The total number of network buffers is currently set to 32768. You can increase this number by setting the configuration key 'taskmanager.network.numberOfBuffers'. at org.apache.flink.runtime.io.network.buffer.NetworkBufferPool.createBufferPool(NetworkBufferPool.java:196) at org.apache.flink.runtime.io.network.NetworkEnvironment.registerTask(NetworkEnvironment.java:327) at org.apache.flink.runtime.taskmanager.Task.run(Task.java:469) at java.lang.Thread.run(Thread.java:745) So I tried different solutions, both with increasing numberOfBuffers (Max value tried 98304) or decreasing the degreeOfParallelism (Min value tried 300) and testing those configs with a dummy dataset seems to solve the number of buffers issue. But In each case with the 80GiB dataset now I struggle with a new exception; the following with a degree of parallelism = 300 and numberOfBuffers = 32768 org.apache.flink.client.program.ProgramInvocationException: The program execution failed: Job execution failed. at org.apache.flink.client.program.Client.runBlocking(Client.java:381) at org.apache.flink.client.program.Client.runBlocking(Client.java:355) at org.apache.flink.client.program.Client.runBlocking(Client.java:315) at org.apache.flink.client.program.ContextEnvironment.execute(ContextEnvironment.java:60) at org.apache.flink.api.scala.ExecutionEnvironment.execute(ExecutionEnvironment.scala:652) at dima.tu.berlin.benchmark.flink.mlr.FlinkSLRTrainCommon$.main(FlinkSLRTrainCommon.scala:110) at dima.tu.berlin.benchmark.flink.mlr.FlinkSLRTrainCommon.main(FlinkSLRTrainCommon.scala) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:505) at org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:403) at org.apache.flink.client.program.Client.runBlocking(Client.java:248) at org.apache.flink.client.CliFrontend.executeProgramBlocking(CliFrontend.java:866) at org.apache.flink.client.CliFrontend.run(CliFrontend.java:333) at org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:1192) at org.apache.flink.client.CliFrontend.main(CliFrontend.java:1243) Caused by: org.apache.flink.runtime.client.JobExecutionException: Job execution failed. at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply$mcV$sp(JobManager.scala:717) at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply(JobManager.scala:663) at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply(JobManager.scala:663) at scala.concurrent.impl.Future$PromiseCompletingRunnable.liftedTree1$1(Future.scala:24) at scala.concurrent.impl.Future$PromiseCompletingRunnable.run(Future.scala:24) at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:41) at akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:401) at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253) at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346) at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) Caused by: java.lang.RuntimeException: Emitting the record caused an I/O exception: Channel to path '/data/3/peel/flink/tmp/flink-io-3a97d62c-ada4-44f1-ab72-dd018386f9aa/79305169d69f7e8b361a175af87353fa.channel' could not be opened. at org.apache.flink.runtime.operators.shipping.OutputCollector.collect(OutputCollector.java:69) at org.apache.flink.runtime.operators.chaining.ChainedMapDriver.collect(ChainedMapDriver.java:78) at org.apache.flink.runtime.operators.MapDriver.run(MapDriver.java:97) at org.apache.flink.runtime.operators.BatchTask.run(BatchTask.java:480) at org.apache.flink.runtime.operators.BatchTask.invoke(BatchTask.java:345) at org.apache.flink.runtime.taskmanager.Task.run(Task.java:559) at java.lang.Thread.run(Thread.java:745) Caused by: java.io.IOException: Channel to path '/data/3/peel/flink/tmp/flink-io-3a97d62c-ada4-44f1-ab72-dd018386f9aa/79305169d69f7e8b361a175af87353fa.channel' could not be opened. at org.apache.flink.runtime.io.disk.iomanager.AbstractFileIOChannel.<init>(AbstractFileIOChannel.java:61) at org.apache.flink.runtime.io.disk.iomanager.AsynchronousFileIOChannel.<init>(AsynchronousFileIOChannel.java:86) at org.apache.flink.runtime.io.disk.iomanager.AsynchronousBufferFileWriter.<init>(AsynchronousBufferFileWriter.java:31) at org.apache.flink.runtime.io.disk.iomanager.IOManagerAsync.createBufferFileWriter(IOManagerAsync.java:257) at org.apache.flink.runtime.io.network.partition.SpillableSubpartition.releaseMemory(SpillableSubpartition.java:151) at org.apache.flink.runtime.io.network.partition.ResultPartition.releaseMemory(ResultPartition.java:366) at org.apache.flink.runtime.io.network.buffer.LocalBufferPool.requestBuffer(LocalBufferPool.java:159) at org.apache.flink.runtime.io.network.buffer.LocalBufferPool.requestBufferBlocking(LocalBufferPool.java:133) at org.apache.flink.runtime.io.network.api.writer.RecordWriter.emit(RecordWriter.java:92) at org.apache.flink.runtime.operators.shipping.OutputCollector.collect(OutputCollector.java:65) ... 6 more Caused by: java.io.FileNotFoundException: /data/3/peel/flink/tmp/flink-io-3a97d62c-ada4-44f1-ab72-dd018386f9aa/79305169d69f7e8b361a175af87353fa.channel (No such file or directory) at java.io.RandomAccessFile.open0(Native Method) at java.io.RandomAccessFile.open(RandomAccessFile.java:316) at java.io.RandomAccessFile.<init>(RandomAccessFile.java:243) at java.io.RandomAccessFile.<init>(RandomAccessFile.java:124) at org.apache.flink.runtime.io.disk.iomanager.AbstractFileIOChannel.<init>(AbstractFileIOChannel.java:57) ... 15 more here the related jobmanager full log. I can't figure out a solution. Thank you and have a nice day. Andrea Spina Guest student at DIMA, TU Berlin N.Tessera: 74598 MAT: 89369 Ingegneria Informatica [LM] (D.M. 270) |
Hi Andrea,
The number of network buffers should be sufficient. Actually, assuming you have 16 task slots on each of the 25 nodes, it should be enough to have 16^2 * 25 * 4 = 14400 network buffers. See https://ci.apache.org/projects/flink/flink-docs-master/setup/config.html#background So we have to investigate a little more. Which version of Flink are you using? Cheers, Max On Tue, Jun 28, 2016 at 11:57 AM, ANDREA SPINA <[hidden email]> wrote: > Hi everyone, > > I am running some Flink experiments with Peel benchmark > http://peel-framework.org/ and I am struggling with exceptions: the > environment is a 25-nodes cluster, 16 cores per nodes. The dataset is ~80GiB > and is located on Hdfs 2.7.1. > > At the beginning I tried with 400 as degree of parallelism and with the > following configuration: > > jobmanager.rpc.address = ${runtime.hostname} > akka.log.lifecycle.events = ON > akka.ask.timeout = 300s > jobmanager.rpc.port = 6002 > > jobmanager.heap.mb = 1024 > jobmanager.web.port = 6004 > > taskmanager.heap.mb = 28672 > taskmanager.memory.fraction = 0.7 > taskmanager.network.numberOfBuffers = 32768 > taskmanager.network.bufferSizeInBytes = 16384 > taskmanager.tmp.dirs = > "/data/1/peel/flink/tmp:/data/2/peel/flink/tmp:/data/3/peel/flink/tmp:/data/4/peel/flink/tmp" > taskmanager.debug.memory.startLogThread = true > > the following exception will raise > > Caused by: java.io.IOException: Insufficient number of network buffers: > required 350, but only 317 available. The total number of network buffers is > currently set to 32768. You can increase this number by setting the > configuration key 'taskmanager.network.numberOfBuffers'. > at > org.apache.flink.runtime.io.network.buffer.NetworkBufferPool.createBufferPool(NetworkBufferPool.java:196) > at > org.apache.flink.runtime.io.network.NetworkEnvironment.registerTask(NetworkEnvironment.java:327) > at org.apache.flink.runtime.taskmanager.Task.run(Task.java:469) > at java.lang.Thread.run(Thread.java:745) > > So I tried different solutions, both with increasing numberOfBuffers (Max > value tried 98304) or decreasing the degreeOfParallelism (Min value tried > 300) and testing those configs with a dummy dataset seems to solve the > number of buffers issue. > But In each case with the 80GiB dataset now I struggle with a new exception; > the following with a degree of parallelism = 300 and numberOfBuffers = 32768 > > org.apache.flink.client.program.ProgramInvocationException: The program > execution failed: Job execution failed. > at org.apache.flink.client.program.Client.runBlocking(Client.java:381) > at org.apache.flink.client.program.Client.runBlocking(Client.java:355) > at org.apache.flink.client.program.Client.runBlocking(Client.java:315) > at > org.apache.flink.client.program.ContextEnvironment.execute(ContextEnvironment.java:60) > at > org.apache.flink.api.scala.ExecutionEnvironment.execute(ExecutionEnvironment.scala:652) > at > dima.tu.berlin.benchmark.flink.mlr.FlinkSLRTrainCommon$.main(FlinkSLRTrainCommon.scala:110) > at > dima.tu.berlin.benchmark.flink.mlr.FlinkSLRTrainCommon.main(FlinkSLRTrainCommon.scala) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at > org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:505) > at > org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:403) > at org.apache.flink.client.program.Client.runBlocking(Client.java:248) > at > org.apache.flink.client.CliFrontend.executeProgramBlocking(CliFrontend.java:866) > at org.apache.flink.client.CliFrontend.run(CliFrontend.java:333) > at > org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:1192) > at org.apache.flink.client.CliFrontend.main(CliFrontend.java:1243) > Caused by: org.apache.flink.runtime.client.JobExecutionException: Job > execution failed. > at > org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply$mcV$sp(JobManager.scala:717) > at > org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply(JobManager.scala:663) > at > org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply(JobManager.scala:663) > at > scala.concurrent.impl.Future$PromiseCompletingRunnable.liftedTree1$1(Future.scala:24) > at > scala.concurrent.impl.Future$PromiseCompletingRunnable.run(Future.scala:24) > at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:41) > at > akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:401) > at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) > at > scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253) > at > scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346) > at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) > at > scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) > Caused by: java.lang.RuntimeException: Emitting the record caused an I/O > exception: Channel to path > '/data/3/peel/flink/tmp/flink-io-3a97d62c-ada4-44f1-ab72-dd018386f9aa/79305169d69f7e8b361a175af87353fa.channel' > could not be opened. > at > org.apache.flink.runtime.operators.shipping.OutputCollector.collect(OutputCollector.java:69) > at > org.apache.flink.runtime.operators.chaining.ChainedMapDriver.collect(ChainedMapDriver.java:78) > at org.apache.flink.runtime.operators.MapDriver.run(MapDriver.java:97) > at org.apache.flink.runtime.operators.BatchTask.run(BatchTask.java:480) > at org.apache.flink.runtime.operators.BatchTask.invoke(BatchTask.java:345) > at org.apache.flink.runtime.taskmanager.Task.run(Task.java:559) > at java.lang.Thread.run(Thread.java:745) > Caused by: java.io.IOException: Channel to path > '/data/3/peel/flink/tmp/flink-io-3a97d62c-ada4-44f1-ab72-dd018386f9aa/79305169d69f7e8b361a175af87353fa.channel' > could not be opened. > at > org.apache.flink.runtime.io.disk.iomanager.AbstractFileIOChannel.<init>(AbstractFileIOChannel.java:61) > at > org.apache.flink.runtime.io.disk.iomanager.AsynchronousFileIOChannel.<init>(AsynchronousFileIOChannel.java:86) > at > org.apache.flink.runtime.io.disk.iomanager.AsynchronousBufferFileWriter.<init>(AsynchronousBufferFileWriter.java:31) > at > org.apache.flink.runtime.io.disk.iomanager.IOManagerAsync.createBufferFileWriter(IOManagerAsync.java:257) > at > org.apache.flink.runtime.io.network.partition.SpillableSubpartition.releaseMemory(SpillableSubpartition.java:151) > at > org.apache.flink.runtime.io.network.partition.ResultPartition.releaseMemory(ResultPartition.java:366) > at > org.apache.flink.runtime.io.network.buffer.LocalBufferPool.requestBuffer(LocalBufferPool.java:159) > at > org.apache.flink.runtime.io.network.buffer.LocalBufferPool.requestBufferBlocking(LocalBufferPool.java:133) > at > org.apache.flink.runtime.io.network.api.writer.RecordWriter.emit(RecordWriter.java:92) > at > org.apache.flink.runtime.operators.shipping.OutputCollector.collect(OutputCollector.java:65) > ... 6 more > Caused by: java.io.FileNotFoundException: > /data/3/peel/flink/tmp/flink-io-3a97d62c-ada4-44f1-ab72-dd018386f9aa/79305169d69f7e8b361a175af87353fa.channel > (No such file or directory) > at java.io.RandomAccessFile.open0(Native Method) > at java.io.RandomAccessFile.open(RandomAccessFile.java:316) > at java.io.RandomAccessFile.<init>(RandomAccessFile.java:243) > at java.io.RandomAccessFile.<init>(RandomAccessFile.java:124) > at > org.apache.flink.runtime.io.disk.iomanager.AbstractFileIOChannel.<init>(AbstractFileIOChannel.java:57) > ... 15 more > > here the related jobmanager full log. I can't figure out a solution. > > Thank you and have a nice day. > > -- > Andrea Spina > Guest student at DIMA, TU Berlin > N.Tessera: 74598 > MAT: 89369 > Ingegneria Informatica [LM] (D.M. 270) |
Hi Max, thank you for the fast reply and sorry: I use flink-1.0.3. Yes I tested on dummy dataset with numOfBuffers = 16384 and decreasing the parallelism degree and this solution solved the first exception. Anyway on the 80GiB dataset I struggle with the second exception. Regards, Andrea 2016-06-28 12:08 GMT+02:00 Maximilian Michels <[hidden email]>: Hi Andrea, Andrea Spina N.Tessera: 74598 MAT: 89369 Ingegneria Informatica [LM] (D.M. 270) |
Hi, the problem was solved after I figured out there was an istance of Flink TaskManager running on a node of the cluster. Thank you, Andrea 2016-06-28 12:17 GMT+02:00 ANDREA SPINA <[hidden email]>:
Andrea Spina N.Tessera: 74598 MAT: 89369 Ingegneria Informatica [LM] (D.M. 270) |
Good to hear! :)
On Wed, Jun 29, 2016 at 12:08 PM, ANDREA SPINA <[hidden email]> wrote: > Hi, > > the problem was solved after I figured out there was an istance of Flink > TaskManager running on a node of the cluster. > Thank you, > Andrea > > 2016-06-28 12:17 GMT+02:00 ANDREA SPINA <[hidden email]>: >> >> Hi Max, >> thank you for the fast reply and sorry: I use flink-1.0.3. >> Yes I tested on dummy dataset with numOfBuffers = 16384 and decreasing the >> parallelism degree and this solution solved the first exception. Anyway on >> the 80GiB dataset I struggle with the second exception. >> >> Regards, >> Andrea >> >> 2016-06-28 12:08 GMT+02:00 Maximilian Michels <[hidden email]>: >>> >>> Hi Andrea, >>> >>> The number of network buffers should be sufficient. Actually, assuming >>> you have 16 task slots on each of the 25 nodes, it should be enough to >>> have 16^2 * 25 * 4 = 14400 network buffers. >>> >>> See >>> https://ci.apache.org/projects/flink/flink-docs-master/setup/config.html#background >>> >>> So we have to investigate a little more. Which version of Flink are you >>> using? >>> >>> Cheers, >>> Max >>> >>> On Tue, Jun 28, 2016 at 11:57 AM, ANDREA SPINA >>> <[hidden email]> wrote: >>> > Hi everyone, >>> > >>> > I am running some Flink experiments with Peel benchmark >>> > http://peel-framework.org/ and I am struggling with exceptions: the >>> > environment is a 25-nodes cluster, 16 cores per nodes. The dataset is >>> > ~80GiB >>> > and is located on Hdfs 2.7.1. >>> > >>> > At the beginning I tried with 400 as degree of parallelism and with the >>> > following configuration: >>> > >>> > jobmanager.rpc.address = ${runtime.hostname} >>> > akka.log.lifecycle.events = ON >>> > akka.ask.timeout = 300s >>> > jobmanager.rpc.port = 6002 >>> > >>> > jobmanager.heap.mb = 1024 >>> > jobmanager.web.port = 6004 >>> > >>> > taskmanager.heap.mb = 28672 >>> > taskmanager.memory.fraction = 0.7 >>> > taskmanager.network.numberOfBuffers = 32768 >>> > taskmanager.network.bufferSizeInBytes = 16384 >>> > taskmanager.tmp.dirs = >>> > >>> > "/data/1/peel/flink/tmp:/data/2/peel/flink/tmp:/data/3/peel/flink/tmp:/data/4/peel/flink/tmp" >>> > taskmanager.debug.memory.startLogThread = true >>> > >>> > the following exception will raise >>> > >>> > Caused by: java.io.IOException: Insufficient number of network buffers: >>> > required 350, but only 317 available. The total number of network >>> > buffers is >>> > currently set to 32768. You can increase this number by setting the >>> > configuration key 'taskmanager.network.numberOfBuffers'. >>> > at >>> > >>> > org.apache.flink.runtime.io.network.buffer.NetworkBufferPool.createBufferPool(NetworkBufferPool.java:196) >>> > at >>> > >>> > org.apache.flink.runtime.io.network.NetworkEnvironment.registerTask(NetworkEnvironment.java:327) >>> > at org.apache.flink.runtime.taskmanager.Task.run(Task.java:469) >>> > at java.lang.Thread.run(Thread.java:745) >>> > >>> > So I tried different solutions, both with increasing numberOfBuffers >>> > (Max >>> > value tried 98304) or decreasing the degreeOfParallelism (Min value >>> > tried >>> > 300) and testing those configs with a dummy dataset seems to solve the >>> > number of buffers issue. >>> > But In each case with the 80GiB dataset now I struggle with a new >>> > exception; >>> > the following with a degree of parallelism = 300 and numberOfBuffers = >>> > 32768 >>> > >>> > org.apache.flink.client.program.ProgramInvocationException: The program >>> > execution failed: Job execution failed. >>> > at org.apache.flink.client.program.Client.runBlocking(Client.java:381) >>> > at org.apache.flink.client.program.Client.runBlocking(Client.java:355) >>> > at org.apache.flink.client.program.Client.runBlocking(Client.java:315) >>> > at >>> > >>> > org.apache.flink.client.program.ContextEnvironment.execute(ContextEnvironment.java:60) >>> > at >>> > >>> > org.apache.flink.api.scala.ExecutionEnvironment.execute(ExecutionEnvironment.scala:652) >>> > at >>> > >>> > dima.tu.berlin.benchmark.flink.mlr.FlinkSLRTrainCommon$.main(FlinkSLRTrainCommon.scala:110) >>> > at >>> > >>> > dima.tu.berlin.benchmark.flink.mlr.FlinkSLRTrainCommon.main(FlinkSLRTrainCommon.scala) >>> > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>> > at >>> > >>> > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >>> > at >>> > >>> > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>> > at java.lang.reflect.Method.invoke(Method.java:498) >>> > at >>> > >>> > org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:505) >>> > at >>> > >>> > org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:403) >>> > at org.apache.flink.client.program.Client.runBlocking(Client.java:248) >>> > at >>> > >>> > org.apache.flink.client.CliFrontend.executeProgramBlocking(CliFrontend.java:866) >>> > at org.apache.flink.client.CliFrontend.run(CliFrontend.java:333) >>> > at >>> > >>> > org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:1192) >>> > at org.apache.flink.client.CliFrontend.main(CliFrontend.java:1243) >>> > Caused by: org.apache.flink.runtime.client.JobExecutionException: Job >>> > execution failed. >>> > at >>> > >>> > org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply$mcV$sp(JobManager.scala:717) >>> > at >>> > >>> > org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply(JobManager.scala:663) >>> > at >>> > >>> > org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply(JobManager.scala:663) >>> > at >>> > >>> > scala.concurrent.impl.Future$PromiseCompletingRunnable.liftedTree1$1(Future.scala:24) >>> > at >>> > >>> > scala.concurrent.impl.Future$PromiseCompletingRunnable.run(Future.scala:24) >>> > at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:41) >>> > at >>> > >>> > akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:401) >>> > at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) >>> > at >>> > >>> > scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253) >>> > at >>> > >>> > scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346) >>> > at >>> > scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) >>> > at >>> > >>> > scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) >>> > Caused by: java.lang.RuntimeException: Emitting the record caused an >>> > I/O >>> > exception: Channel to path >>> > >>> > '/data/3/peel/flink/tmp/flink-io-3a97d62c-ada4-44f1-ab72-dd018386f9aa/79305169d69f7e8b361a175af87353fa.channel' >>> > could not be opened. >>> > at >>> > >>> > org.apache.flink.runtime.operators.shipping.OutputCollector.collect(OutputCollector.java:69) >>> > at >>> > >>> > org.apache.flink.runtime.operators.chaining.ChainedMapDriver.collect(ChainedMapDriver.java:78) >>> > at org.apache.flink.runtime.operators.MapDriver.run(MapDriver.java:97) >>> > at org.apache.flink.runtime.operators.BatchTask.run(BatchTask.java:480) >>> > at >>> > org.apache.flink.runtime.operators.BatchTask.invoke(BatchTask.java:345) >>> > at org.apache.flink.runtime.taskmanager.Task.run(Task.java:559) >>> > at java.lang.Thread.run(Thread.java:745) >>> > Caused by: java.io.IOException: Channel to path >>> > >>> > '/data/3/peel/flink/tmp/flink-io-3a97d62c-ada4-44f1-ab72-dd018386f9aa/79305169d69f7e8b361a175af87353fa.channel' >>> > could not be opened. >>> > at >>> > >>> > org.apache.flink.runtime.io.disk.iomanager.AbstractFileIOChannel.<init>(AbstractFileIOChannel.java:61) >>> > at >>> > >>> > org.apache.flink.runtime.io.disk.iomanager.AsynchronousFileIOChannel.<init>(AsynchronousFileIOChannel.java:86) >>> > at >>> > >>> > org.apache.flink.runtime.io.disk.iomanager.AsynchronousBufferFileWriter.<init>(AsynchronousBufferFileWriter.java:31) >>> > at >>> > >>> > org.apache.flink.runtime.io.disk.iomanager.IOManagerAsync.createBufferFileWriter(IOManagerAsync.java:257) >>> > at >>> > >>> > org.apache.flink.runtime.io.network.partition.SpillableSubpartition.releaseMemory(SpillableSubpartition.java:151) >>> > at >>> > >>> > org.apache.flink.runtime.io.network.partition.ResultPartition.releaseMemory(ResultPartition.java:366) >>> > at >>> > >>> > org.apache.flink.runtime.io.network.buffer.LocalBufferPool.requestBuffer(LocalBufferPool.java:159) >>> > at >>> > >>> > org.apache.flink.runtime.io.network.buffer.LocalBufferPool.requestBufferBlocking(LocalBufferPool.java:133) >>> > at >>> > >>> > org.apache.flink.runtime.io.network.api.writer.RecordWriter.emit(RecordWriter.java:92) >>> > at >>> > >>> > org.apache.flink.runtime.operators.shipping.OutputCollector.collect(OutputCollector.java:65) >>> > ... 6 more >>> > Caused by: java.io.FileNotFoundException: >>> > >>> > /data/3/peel/flink/tmp/flink-io-3a97d62c-ada4-44f1-ab72-dd018386f9aa/79305169d69f7e8b361a175af87353fa.channel >>> > (No such file or directory) >>> > at java.io.RandomAccessFile.open0(Native Method) >>> > at java.io.RandomAccessFile.open(RandomAccessFile.java:316) >>> > at java.io.RandomAccessFile.<init>(RandomAccessFile.java:243) >>> > at java.io.RandomAccessFile.<init>(RandomAccessFile.java:124) >>> > at >>> > >>> > org.apache.flink.runtime.io.disk.iomanager.AbstractFileIOChannel.<init>(AbstractFileIOChannel.java:57) >>> > ... 15 more >>> > >>> > here the related jobmanager full log. I can't figure out a solution. >>> > >>> > Thank you and have a nice day. >>> > >>> > -- >>> > Andrea Spina >>> > Guest student at DIMA, TU Berlin >>> > N.Tessera: 74598 >>> > MAT: 89369 >>> > Ingegneria Informatica [LM] (D.M. 270) >> >> >> >> >> -- >> Andrea Spina >> N.Tessera: 74598 >> MAT: 89369 >> Ingegneria Informatica [LM] (D.M. 270) > > > > > -- > Andrea Spina > N.Tessera: 74598 > MAT: 89369 > Ingegneria Informatica [LM] (D.M. 270) |
Free forum by Nabble | Edit this page |