hi everyone!
can flink submit job which read some custom file distributed by hdfs DistributedCache.
like spark can do that with the follow command:
bin/spark-submit --master yarn --deploy-mode cluster --files /opt/its007-datacollection-conf.properties#its007-datacollection-conf.properties ...
then spark driver can read `its007-datacollection-conf.properties` file in work directory.
thanks!