Hello everybody,
I set up a a Flink (1.12.1) and Hadoop (3.2.1) cluster on two machines. The job should store the checkpoints on HDFS like so: ```java StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment(); env.enableCheckpointing(15000, CheckpointingMode.EXACTLY_ONCE); env.setStateBackend(new FsStateBackend("hdfs://node-1:9000/flink")); ``` Unfortunately, the JobManager throws ``` org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: Could not find a file system implementation for scheme 'hdfs'. The scheme is not directly supported by Flink and no Hadoop file system to support this scheme could be loaded. For a full list of supported file systems, please see https://ci.apache.org/projects/flink/flink-docs-stable/ops/filesystems/. // ... Caused by: org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: Hadoop is not in the classpath/dependencies. ``` and I don't understand why. `echo $HADOOP_CLASSPATH` returns the path of Hadoop libraries with wildcards. Flink's JobManger prints the classpath which includes specific packages from these Hadoop libraries. Besides that, Flink creates the state directories on HDFS, but no content. Thank you for any advice, Matthias |
I have the same problem ...
-- Sent from: http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/ |
In reply to this post by Matthias Seiler
I downloaded the lib (last version) from here:
https://repo.maven.apache.org/maven2/org/apache/flink/flink-shaded-hadoop-2-uber/2.8.3-7.0/ and put it in the flink_home/lib directory. It helped. -- Sent from: http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/ |
In reply to this post by Matthias Seiler
Hey Matthias, Maybe the classpath contains hadoop libraries, but not the HDFS libraries? The "DistributedFileSystem" class needs to be accessible to the classloader. Can you check if that class is available? Best, Robert On Thu, Mar 25, 2021 at 11:10 AM Matthias Seiler <[hidden email]> wrote: Hello everybody, |
Thank you all for the replies! On 3/26/21 10:20 AM, Robert Metzger
wrote:
|
This looks related to HDFS-12920; where
Hadoop 2.X tries to read a duration from hdfs-default.xml
expecting plain numbers, but in 3.x they also contain time units.
On 3/30/2021 9:37 AM, Matthias Seiler
wrote:
|
Free forum by Nabble | Edit this page |