Sorry, I forgot to mention the environment.
We use Flink 1.9.1 on a cloudera cdh6. 3.1 cluster (with Hadoop 3.0.0 but using Flink shaded 2.8.3-7. Might this be a problem? As it seems to arise from kryo, I doubt it)
Our flink is configured as default. Our job uses FsStateBackend and exactly once processing with Kafka source and sink.
Best regards
Theo
-------- Ursprüngliche Nachricht --------
Betreff: ArrayIndexOutOfBoundException on checkpoint creation
Von: Theo Diefenthal
An: user
Cc:
Hi,
We have a pipeline with a custom ProcessFunction and state (see [1], implemented as suggested by Fabian with a ValueState<LinkedList<MyPOJO> and ValueState<HashSet<String>>)
The behavior of that function works fine in our unittests and with low load in our test environment (100.000 records per minute). On the production environment, we observe reproduceable crashes like the attached one.
Any idea on why this out of bound could be caused? Every time we read the state and modify it, we are certain that an .update() was called:
2019-11-26T11:26:55+01:00 host19 java.lang.Exception: Could not materialize checkpoint 7 for operator our_operator) (4/8).
2019-11-26T11:26:55+01:00 host19 at org.apache.flink.streaming.runtime.tasks.StreamTask$AsyncCheckpointRunnable.handleExecutionException(StreamTask.java:1100)
2019-11-26T11:26:55+01:00 host19 at org.apache.flink.streaming.runtime.tasks.StreamTask$AsyncCheckpointRunnable.run(StreamTask.java:1042)
2019-11-26T11:26:55+01:00 host19 at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
2019-11-26T11:26:55+01:00 host19 at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
2019-11-26T11:26:55+01:00 host19 at java.lang.Thread.run(Thread.java:745)
2019-11-26T11:26:55+01:00 host19 Caused by: java.util.concurrent.ExecutionException: java.lang.ArrayIndexOutOfBoundsException: 67108864
2019-11-26T11:26:55+01:00 host19 at java.util.concurrent.FutureTask.report(FutureTask.java:122)
2019-11-26T11:26:55+01:00 host19 at java.util.concurrent.FutureTask.get(FutureTask.java:192)
2019-11-26T11:26:55+01:00 host19 at org.apache.flink.runtime.concurrent.FutureUtils.runIfNotDoneAndGet(FutureUtils.java:450)
2019-11-26T11:26:55+01:00 host19 at org.apache.flink.streaming.api.operators.OperatorSnapshotFinalizer.<init>(OperatorSnapshotFinalizer.java:47)
2019-11-26T11:26:55+01:00 host19 at org.apache.flink.streaming.runtime.tasks.StreamTask$AsyncCheckpointRunnable.run(StreamTask.java:1011)
2019-11-26T11:26:55+01:00 host19 ... 3 more
2019-11-26T11:26:55+01:00 host19 Caused by: java.lang.ArrayIndexOutOfBoundsException: 67108864
2019-11-26T11:26:55+01:00 host19 at com.esotericsoftware.kryo.util.IdentityObjectIntMap.clear(IdentityObjectIntMap.java:364)
2019-11-26T11:26:55+01:00 host19 at com.esotericsoftware.kryo.util.MapReferenceResolver.reset(MapReferenceResolver.java:47)
2019-11-26T11:26:55+01:00 host19 at com.esotericsoftware.kryo.Kryo.reset(Kryo.java:836)
2019-11-26T11:26:55+01:00 host19 at com.esotericsoftware.kryo.Kryo.writeClassAndObject(Kryo.java:601)
2019-11-26T11:26:55+01:00 host19 at org.apache.flink.api.java.typeutils.runtime.kryo.KryoSerializer.serialize(KryoSerializer.java:305)
2019-11-26T11:26:55+01:00 host19 at org.apache.flink.runtime.state.heap.CopyOnWriteStateMapSnapshot.writeState(CopyOnWriteStateMapSnapshot.java:116)
2019-11-26T11:26:55+01:00 host19 at org.apache.flink.runtime.state.heap.AbstractStateTableSnapshot.writeStateInKeyGroup(AbstractStateTableSnapshot.java:121)
2019-11-26T11:26:55+01:00 host19 at org.apache.flink.runtime.state.heap.CopyOnWriteStateTableSnapshot.writeStateInKeyGroup(CopyOnWriteStateTableSnapshot.java:37)
2019-11-26T11:26:55+01:00 host19 at org.apache.flink.runtime.state.heap.HeapSnapshotStrategy$1.callInternal(HeapSnapshotStrategy.java:191)
2019-11-26T11:26:55+01:00 host19 at org.apache.flink.runtime.state.heap.HeapSnapshotStrategy$1.callInternal(HeapSnapshotStrategy.java:158)
2019-11-26T11:26:55+01:00 host19 at org.apache.flink.runtime.state.AsyncSnapshotCallable.call(AsyncSnapshotCallable.java:75)
2019-11-26T11:26:55+01:00 host19 at java.util.concurrent.FutureTask.run(FutureTask.java:266)
2019-11-26T11:26:55+01:00 host19 at org.apache.flink.runtime.concurrent.FutureUtils.runIfNotDoneAndGet(FutureUtils.java:447)
2019-11-26T11:26:55+01:00 host19 ... 5 more
2019-11-26T11:26:55+01:00 host18 WARN org.apache.hadoop.hdfs.DataStreamer - DataStreamer Exception
2019-11-26T11:26:55+01:00 host18 java.io.FileNotFoundException: File does not exist: /.../STATE/CHECKPOINTS/0a2e111b3a800aae0d3b49f33e0db6f3/chk-7/3da2a0a4-f5ef-4e8c-bc1a-9fe892cb0b18 (inode 577546140) Holder DFSClient_NONMAPREDUCE_-1714419242_95 does not have any open files.
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:2782)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.analyzeFileState(FSDirWriteFileOp.java:599)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.validateAddBlock(FSDirWriteFileOp.java:171)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2661)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:872)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:550)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:523)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:991)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:869)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:815)
2019-11-26T11:26:55+01:00 host18 at java.security.AccessController.doPrivileged(Native Method)
2019-11-26T11:26:55+01:00 host18 at javax.security.auth.Subject.doAs(Subject.java:422)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1875)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2675)
2019-11-26T11:26:55+01:00 host18
2019-11-26T11:26:55+01:00 host18 at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
2019-11-26T11:26:55+01:00 host18 at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
2019-11-26T11:26:55+01:00 host18 at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
2019-11-26T11:26:55+01:00 host18 at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:121)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:88)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.DataStreamer.locateFollowingBlock(DataStreamer.java:1842)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.DataStreamer.nextBlockOutputStream(DataStreamer.java:1638)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:704)
2019-11-26T11:26:55+01:00 host18 Caused by: org.apache.hadoop.ipc.RemoteException(java.io.FileNotFoundException): File does not exist: /.../STATE/INGESTOR/CHECKPOINTS/0a2e111b3a800aae0d3b49f33e0db6f3/chk-7/3da2a0a4-f5ef-4e8c-bc1a-9fe892cb0b18 (inode 577546140) Holder DFSClient_NONMAPREDUCE_-1714419242_95 does not have any open files.
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:2782)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.analyzeFileState(FSDirWriteFileOp.java:599)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.validateAddBlock(FSDirWriteFileOp.java:171)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2661)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:872)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:550)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:523)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:991)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:869)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:815)
2019-11-26T11:26:55+01:00 host18 at java.security.AccessController.doPrivileged(Native Method)
2019-11-26T11:26:55+01:00 host18 at javax.security.auth.Subject.doAs(Subject.java:422)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1875)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2675)
2019-11-26T11:26:55+01:00 host18
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1489)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.Client.call(Client.java:1435)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.Client.call(Client.java:1345)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:227)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:116)
2019-11-26T11:26:55+01:00 host18 at com.sun.proxy.$Proxy19.addBlock(Unknown Source)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:444)
2019-11-26T11:26:55+01:00 host18 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
2019-11-26T11:26:55+01:00 host18 at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
2019-11-26T11:26:55+01:00 host18 at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
2019-11-26T11:26:55+01:00 host18 at java.lang.reflect.Method.invoke(Method.java:498)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:409)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:163)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:155)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:95)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:346)
2019-11-26T11:26:55+01:00 host18 at com.sun.proxy.$Proxy20.addBlock(Unknown Source)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.DataStreamer.locateFollowingBlock(DataStreamer.java:1838)
2019-11-26T11:26:55+01:00 host18 ... 2 more
2019-11-26T11:26:55+01:00 host18 WARN org.apache.flink.runtime.state.filesystem.FsCheckpointStreamFactory - Could not close the state stream for hdfs:/.../STATE/INGESTOR/CHECKPOINTS/0a2e111b3a800aae0d3b49f33e0db6f3/chk-7/3da2a0a4-f5ef-4e8c-bc1a-9fe892cb0b18.
2019-11-26T11:26:55+01:00 host18 java.io.FileNotFoundException: File does not exist: /.../STATE/INGESTOR/CHECKPOINTS/0a2e111b3a800aae0d3b49f33e0db6f3/chk-7/3da2a0a4-f5ef-4e8c-bc1a-9fe892cb0b18 (inode 577546140) Holder DFSClient_NONMAPREDUCE_-1714419242_95 does not have any open files.
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:2782)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.analyzeFileState(FSDirWriteFileOp.java:599)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.validateAddBlock(FSDirWriteFileOp.java:171)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2661)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:872)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:550)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:523)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:991)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:869)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:815)
2019-11-26T11:26:55+01:00 host18 at java.security.AccessController.doPrivileged(Native Method)
2019-11-26T11:26:55+01:00 host18 at javax.security.auth.Subject.doAs(Subject.java:422)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1875)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2675)
2019-11-26T11:26:55+01:00 host18
2019-11-26T11:26:55+01:00 host18 at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
2019-11-26T11:26:55+01:00 host18 at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
2019-11-26T11:26:55+01:00 host18 at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
2019-11-26T11:26:55+01:00 host18 at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:121)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:88)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.DataStreamer.locateFollowingBlock(DataStreamer.java:1842)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.DataStreamer.nextBlockOutputStream(DataStreamer.java:1638)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:704)
2019-11-26T11:26:55+01:00 host18 Caused by: org.apache.hadoop.ipc.RemoteException(java.io.FileNotFoundException): File does not exist: /.../STATE/INGESTOR/CHECKPOINTS/0a2e111b3a800aae0d3b49f33e0db6f3/chk-7/3da2a0a4-f5ef-4e8c-bc1a-9fe892cb0b18 (inode 577546140) Holder DFSClient_NONMAPREDUCE_-1714419242_95 does not have any open files.
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:2782)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.analyzeFileState(FSDirWriteFileOp.java:599)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.validateAddBlock(FSDirWriteFileOp.java:171)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2661)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:872)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:550)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:523)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:991)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:869)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:815)
2019-11-26T11:26:55+01:00 host18 at java.security.AccessController.doPrivileged(Native Method)
2019-11-26T11:26:55+01:00 host18 at javax.security.auth.Subject.doAs(Subject.java:422)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1875)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2675)
2019-11-26T11:26:55+01:00 host18
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1489)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.Client.call(Client.java:1435)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.Client.call(Client.java:1345)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:227)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:116)
2019-11-26T11:26:55+01:00 host18 at com.sun.proxy.$Proxy19.addBlock(Unknown Source)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:444)
2019-11-26T11:26:55+01:00 host18 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
2019-11-26T11:26:55+01:00 host18 at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
2019-11-26T11:26:55+01:00 host18 at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
2019-11-26T11:26:55+01:00 host18 at java.lang.reflect.Method.invoke(Method.java:498)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:409)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:163)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:155)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:95)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:346)
2019-11-26T11:26:55+01:00 host18 at com.sun.proxy.$Proxy20.addBlock(Unknown Source)
2019-11-26T11:26:55+01:00 host18 at org.apache.hadoop.hdfs.DataStreamer.locateFollowingBlock(DataStreamer.java:1838)
2019-11-26T11:26:55+01:00 host18 ... 2 more
2019-11-26T11:29:50+01:00 host19 INFO org.apache.flink.yarn.YarnResourceManager - The heartbeat of TaskManager with id container_e137_1574681304655_2382_01_000007 timed out.
Best regards
Theo
[1] http://mail-archives.apache.org/mod_mbox/flink-user/201909.mbox/%
[hidden email]%3E