Hi,
I get many (multiple times per minute) errors in my Namenode HDFS logfile: 2016-10-11 17:17:07,596 INFO ipc.Server (Server.java:logException(2401)) - IPC Server handler 295 on 8020, call org.apache.hadoop.hdfs.protocol.ClientProtocol.delete from datanode1:34872 Call#2361 Retry#0 org.apache.hadoop.fs.PathIsNotEmptyDirectoryException: `/flink/recovery is non empty': Directory is not empty at org.apache.hadoop.hdfs.server.namenode.FSDirDeleteOp.delete(FSDirDeleteOp.java:89) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.delete(FSNamesystem.java:3829) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.delete(NameNodeRpcServer.java:1071) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.delete(ClientNamenodeProtocolServerSideTranslatorPB.java:619) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:640) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:982) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2313) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2309) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1724) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2307) That is the directory I configured for Jobmanager HA. I deleted it before starting the YARN session but that did not help. The folder gets created by Flink without problems. I'm using latest Flink Master (Commit: 6731ec1) and build it for Hadoop 2.7.3. Any idea is highly appreciated. Thanks a lot! |
Hi! I think to some extend this is expected. There is some cleanup code that deletes files and then issues parent directory remove requests. It relies on the fact that the parent directory is only removed if it is empty (after the last file was deleted). Is this a problem right now, or just a confusing behavior? Greetings, Stephan On Tue, Oct 11, 2016 at 5:25 PM, static-max <[hidden email]> wrote:
|
Hi Stephan, it's not a problem, but makes finding other errors on my NameNode complicated as I have this error message every minute. Can't we just delete the directory recursively? Regards, Max 2016-10-11 17:59 GMT+02:00 Stephan Ewen <[hidden email]>:
|
Update: I deleted the /flink/recovery folder on HDFS and even then I get the same Exception after the next checkpoint. 2016-11-21 21:51 GMT+01:00 static-max <[hidden email]>:
|
This could be related to
https://issues.apache.org/jira/browse/FLINK-5063 where some issues related to the cleanup of checkpointing files were fixed. -Max On Mon, Nov 21, 2016 at 10:05 PM, static-max <[hidden email]> wrote: > Update: I deleted the /flink/recovery folder on HDFS and even then I get the > same Exception after the next checkpoint. > > 2016-11-21 21:51 GMT+01:00 static-max <[hidden email]>: >> >> Hi Stephan, >> >> it's not a problem, but makes finding other errors on my NameNode >> complicated as I have this error message every minute. >> Can't we just delete the directory recursively? >> >> Regards, >> Max >> >> 2016-10-11 17:59 GMT+02:00 Stephan Ewen <[hidden email]>: >>> >>> Hi! >>> >>> I think to some extend this is expected. There is some cleanup code that >>> deletes files and then issues parent directory remove requests. It relies >>> on the fact that the parent directory is only removed if it is empty (after >>> the last file was deleted). >>> >>> Is this a problem right now, or just a confusing behavior? >>> >>> Greetings, >>> Stephan >>> >>> >>> On Tue, Oct 11, 2016 at 5:25 PM, static-max <[hidden email]> >>> wrote: >>>> >>>> Hi, >>>> >>>> I get many (multiple times per minute) errors in my Namenode HDFS >>>> logfile: >>>> >>>> 2016-10-11 17:17:07,596 INFO ipc.Server >>>> (Server.java:logException(2401)) - IPC Server handler 295 on 8020, call >>>> org.apache.hadoop.hdfs.protocol.ClientProtocol.delete from datanode1:34872 >>>> Call#2361 Retry#0 >>>> org.apache.hadoop.fs.PathIsNotEmptyDirectoryException: `/flink/recovery >>>> is non empty': Directory is not empty >>>> at >>>> org.apache.hadoop.hdfs.server.namenode.FSDirDeleteOp.delete(FSDirDeleteOp.java:89) >>>> at >>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.delete(FSNamesystem.java:3829) >>>> at >>>> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.delete(NameNodeRpcServer.java:1071) >>>> at >>>> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.delete(ClientNamenodeProtocolServerSideTranslatorPB.java:619) >>>> at >>>> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) >>>> at >>>> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:640) >>>> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:982) >>>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2313) >>>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2309) >>>> at java.security.AccessController.doPrivileged(Native Method) >>>> at javax.security.auth.Subject.doAs(Subject.java:422) >>>> at >>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1724) >>>> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2307) >>>> >>>> That is the directory I configured for Jobmanager HA. I deleted it >>>> before starting the YARN session but that did not help. The folder gets >>>> created by Flink without problems. >>>> >>>> I'm using latest Flink Master (Commit: 6731ec1) and build it for Hadoop >>>> 2.7.3. >>>> >>>> Any idea is highly appreciated. Thanks a lot! >>> >>> >> > |
Free forum by Nabble | Edit this page |