How to run Apache Beam Python pipelines on Flink cluster in Kubernetes? How to run Apache Beam Python pipelines on Flink cluster in Kubernetes? kubernetes kubernetes

How to run Apache Beam Python pipelines on Flink cluster in Kubernetes?


By default (as of this writing), Beam stages runtime dependencies ("artifacts") to a certain directory (/tmp/staged by default) that needs to be accessible to both the job server (in your case, the client) and the Beam worker.

You can get around this by setting the --flink_submit_uber_jar pipeline option. When --flink_submit_uber_jar is set, Beam wraps all your dependencies in a jar that is submitted to Flink.