尝试在 Airflow 中获取 Spark 连接 ID 时出错

Error while try getting spark connection id in Airflow

我在气流 DAG 中的 SparkSubmitOperator 如下所示,下面是我在气流 UI 中的 connection_id 'spark_local' when I try running my DAG am getting this error ,如果我在这里遗漏任何东西,任何人都可以帮助我?

SparkSubmitOperator (task_id='spark_task', application = 'gs://xxx/xxx.jar', conf = {"spark.driver.allowMultipleContexts":True, "spark.blacklist.enabled":False}, conn_id='spark_local', java_class = 'xxx', jars=["gs://xxx/*"], application_args=["xxx "xxx" "xxx,] )

基于 TypeError ,来自文档 https://airflow.apache.org/docs/apache-airflow/1.10.12/_api/airflow/contrib/operators/spark_submit_operator/index.html,它说 jars 参数应该是一个字符串。您能否尝试将 jars="gs://xxx/*" 作为字符串来查看它是否有效?