Hello, I’m trying to set up my flink native Kubernetes cluster with High availability. Here’s the relevant config: |
Sorry, fat finger send before I finished writing…. Hello, I’m trying to set up my flink native Kubernetes cluster with High availability. Here’s the relevant config: kubernetes.service-account: flink-service-account high-availability: org.apache.flink.kubernetes.highavailability.KubernetesHaServicesFactory high-availability.storageDir: s3://corvana-target-file-upload-k8s-usw2.dev.sugar.build/flink/recovery I’m getting an error accessing the bucket. 2021-06-08 14:33:42,189 DEBUG com.amazonaws.services.s3.AmazonS3Client [] - Bucket region cache doesn't have an entry for corvana-target-file-upload-k8s-usw2.dev.sugar.build.
Trying to get bucket region from Amazon S3. 2021-06-08 14:33:42,193 DEBUG com.amazonaws.util.json.Jackson [] - Failed to parse JSON string. com.fasterxml.jackson.databind.exc.MismatchedInputException: No content to map due to end-of-input at [Source: (String)""; line: 1, column: 0] at com.fasterxml.jackson.databind.exc.MismatchedInputException.from(MismatchedInputException.java:59) ~[flink-s3-fs-presto-1.13.0.jar:1.13.0] Is there an additional config I need for specifying the region for the bucket? I’ve been searching the doc and haven’t found anything like that. From:
Kurtis Walker <[hidden email]> Hello, I’m trying to set up my flink native Kubernetes cluster with High availability. Here’s the relevant config: |
It seems to be a S3 issue. And I am not sure it is the root cause. Could you please share more details of the JobManager log? Or could you verify that the Flink cluster could access the S3 bucket successfully(e.g. store the checkpoint) when HA is disabled? Best, Yang Kurtis Walker <[hidden email]> 于2021年6月8日周二 下午11:00写道:
|
I'd try several things
try accessing the bucket from CLI first locally
If it does not work
please check your credentials under ~/.aws/credentials file + ~/.aws/config = since the AWS clients read the credentials from these files by default(unless other credentials are set)
If everything works well:
Please provide more code snippet.
I recently ran Flink job on Application cluster in EKS. the job also reads files from S3. (Without HA)
Tamir
From: Yang Wang <[hidden email]>
Sent: Wednesday, June 9, 2021 11:29 AM To: Kurtis Walker <[hidden email]> Cc: user <[hidden email]> Subject: Re: Using s3 bucket for high availability
It seems to be a S3 issue. And I am not sure it is the root cause. Could you please share more details of the JobManager log?
Or could you verify that the Flink cluster could access the S3 bucket successfully(e.g. store the checkpoint) when HA is disabled?
Best,
Yang
Kurtis Walker <[hidden email]> 于2021年6月8日周二 下午11:00写道:
|
Thank you, I figured it out. My IAM policy was missing some actions. Seems I needed to give it “*” for it to work. From:
Tamir Sagi <[hidden email]> EXTERNAL EMAIL I'd try several things try accessing the bucket from CLI first locally If it does not work please check your credentials under ~/.aws/credentials file + ~/.aws/config = since the AWS clients read the credentials from these files by default(unless other credentials are set) If everything works well:
Please provide more code snippet. I recently ran Flink job on Application cluster in EKS. the job also reads files from S3. (Without HA) Tamir From: Yang Wang <[hidden email]>
It seems to be a S3 issue. And I am not sure it is the root cause. Could you please share more details of the JobManager log?
Or could you verify that the Flink cluster could access the S3 bucket successfully(e.g. store the checkpoint) when HA is disabled? Best, Yang Kurtis Walker <[hidden email]>
于2021年6月8日周二
下午11:00写道:
|
Free forum by Nabble | Edit this page |