Hi, I am running issues when trying to move from HDFS to S3 using Flink 1.6. I am getting an exception from Hadoop code: IOException("Resource " + sCopy + Digging into this, I found there was one commit made by Nico trying to fix this issue in 2018. However, the fix did not work for my case, as the fs.setTimes() method was not implemented in the hadoop-aws S3AFilesystem I am using. And it seems S3 does not allow you to override the last modified time for an object. I am able to make an workaround the other way round: reading the timestamp from S3 and override the local resource. Just wonder if any one has seen similar issues, or he/she can actually make it work by using different implementation of S3AFilesystem? Thanks! Best, Yan |
Hi, Yan. we have met this problem too when using aliyun-pangu and have commented in FLINK-8801 but no response yet. I think most file systems including s3/s3a/s3n/azure/aliyun-oss etc can encounter this problem since they doesn’t implement FileSystem#setTimes but the PR in FLINK-8801 think they does. We have made a similar workaround for this problem. Tao Yang
smime.p7s (3K) Download Attachment |
Hi Yantao, Thanks, I have also commented in the original JIRA. https://issues.apache.org/jira/browse/FLINK-8801?focusedCommentId=16807691&page=com.atlassian.jira.plugin.system.issuetabpanels%3Acomment-tabpanel @Nico @Till Do you mind review if an alternative fix would be needed? If so, I can create a new JIRA. Thanks, Yan
Best, Yan |
Hi Yan and Tao Yang, thanks for raising this issue. Let's continue the discussion on the ticket in order to figure out a proper solution. Cheers, Till On Fri, Apr 5, 2019 at 11:23 PM Yan Yan <[hidden email]> wrote:
|
Free forum by Nabble | Edit this page |