Apache_beam--python --error: the following arguments are required: --output-path

Apache_beam--python --error: the following arguments are required: --output-path

当 运行 我的代码在 python 或协作者出现以下错误时,我从 apache beam 安装所有库,有人在一瞬间给出了这个错误或者知道它..

用法:aaa_users_py.py [-h] [--运行ner RUNNER] [--streaming] [--resource_hint RESOURCE_HINTS] [--beam_services BEAM_SERVICES] [--type_check_strictness {ALL_REQUIRED,DEFAULT_TO_ANY}] [--type_check_additional TYPE_CHECK_ADDITIONAL]
[--no_pipeline_type_check] [--runtime_type_check] [--performance_runtime_type_check] [--allow_non_deterministic_key_coders] [--allow_unsafe_triggers] [--no_direct_runner_use_stacked_bundle] [--direct_runner_bundle_repeat DIRECT_RUNNER_BUNDLE_REPEAT]
[--direct_num_workers DIRECT_NUM_WORKERS] [--direct_running_mode {in_memory,multi_threading,multi_processing}] [--dataflow_endpoint DATAFLOW_ENDPOINT] [--项目项目] [--job_name JOB_NAME] [--staging_location STAGING_LOCATION] [--temp_location TEMP_LOCATION] [--region REGION] [--service_account_email SERVICE_ACCOUNT_EMAIL] [--no_auth] [--template_location TEMPLATE_LOCATION]
[--标签标签] [--更新] [--transform_name_mapping TRANSFORM_NAME_MAPPING] [--enable_streaming_engine] [--dataflow_kms_key DATAFLOW_KMS_KEY] [--create_from_snapshot CREATE_FROM_SNAPSHOT] [--flexrs_goal {COST_OPTIMIZED,SPEED_OPTIMIZED}] [--dataflow_service_option DATAFLOW_SERVICE_OPTIONS] [--enable_hot_key_logging] [--enable_artifact_caching] [--impersonate_service_account IMPERSONATE_SERVICE_ACCOUNT] [--hdfs_host HDFS_HOST] [--hdfs_port HDFS_PORT] [--hdfs_user HDFS_USER] [--hdfs_full_urls] [--num_workers NUM_WORKERS] [--max_num_workers MAX_NUM_WORKERS] [--autoscaling_algorithm {NONE,THROUGHPUT_BASED}] [--worker_machine_type MACHINE_TYPE] [--disk_size_gb DISK_SIZE_GB] [--worker_disk_type DISK_TYPE] [--worker_region WORKER_REGION] [--worker_zone WORKER_ZONE] [--zone ZONE] [--network NETWORK] [--subnetwork SUBNETWORK] [--worker_harness_container_image WORKER_HARNESS_CONTAINER_IMAGE] [--sdk_container_image SDK_CONTAINER_IMAGE] [--sdk_harness_container_image_overrides SDK_HARNESS_CONTAINER_IMAGE_OVERRIDES] [--use_public_ips]
[--no_use_public_ips] [--min_cpu_platform MIN_CPU_PLATFORM] [--dataflow_worker_jar DATAFLOW_WORKER_JAR]
[--dataflow_job_file DATAFLOW_JOB_FILE] [--实验实验] [--number_of_worker_harness_threads NUMBER_OF_WORKER_HARNESS_THREADS] [--profile_cpu] [--profile_memory] [--profile_location PROFILE_LOCATION] [--profile_sample_rate PROFILE_SAMPLE_RATE]
[--requirements_file REQUIREMENTS_FILE] [--requirements_cache REQUIREMENTS_CACHE] [--requirements_cache_only_sources] [--setup_file SETUP_FILE] [--beam_plugin BEAM_PLUGINS] [--pickle_library {cloudpickle,default,dill}] [--save_main_session] [--sdk_location SDK_LOCATION]
[--extra_package EXTRA_PACKAGES] [--prebuild_sdk_container_engine PREBUILD_SDK_CONTAINER_ENGINE] [--prebuild_sdk_container_base_image PREBUILD_SDK_CONTAINER_BASE_IMAGE] [--cloud_build_machine_type CLOUD_BUILD_MACHINE_TYPE] [--docker_registry_push_url DOCKER_REGISTRY_PUSH_URL] [--job_endpoint JOB_ENDPOINT] [--artifact_endpoint ARTIFACT_ENDPOINT] [--job_server_timeout JOB_SERVER_TIMEOUT] [--environment_type ENVIRONMENT_TYPE] [--environment_config ENVIRONMENT_CONFIG] [--environment_option ENVIRONMENT_OPTIONS] [--sdk_worker_parallelism SDK_WORKER_PARALLELISM] [--environment_cache_millis ENVIRONMENT_CACHE_MILLIS] [--output_executable_path OUTPUT_EXECUTABLE_PATH]
[--artifacts_dir ARTIFACTS_DIR] [--job_port JOB_PORT] [--artifact_port ARTIFACT_PORT] [--expansion_port EXPANSION_PORT] [--job_server_java_launcher JOB_SERVER_JAVA_LAUNCHER] [--job_server_jvm_properties JOB_SERVER_JVM_PROPERTIES] [--flink_master FLINK_MASTER] [--flink_version {1.12,1.13,1.14}] [--flink_job_server_jar FLINK_JOB_SERVER_JAR] [--flink_submit_uber_jar] [--spark_master_url SPARK_MASTER_URL] [--spark_job_server_jar SPARK_JOB_SERVER_JAR] [--spark_submit_uber_jar] [--spark_rest_url SPARK_REST_URL] [--spark_version {2,3}] [--on_success_matcher ON_SUCCESS_MATCHER]
[--dry_run DRY_RUN] [--wait_until_finish_duration WAIT_UNTIL_FINISH_DURATION] [--pubsub_root_url PUBSUBROOTURL] [--s3_access_key_id S3_ACCESS_KEY_ID] [--s3_secret_access_key S3_SECRET_ACCESS_KEY] [--s3_session_token S3_SESSION_TOKEN] [--s3_endpoint_url S3_ENDPOINT_URL] [--s3_region_name S3_REGION_NAME] [--s3_api_version S3_API_VERSION]
[--s3_verify S3_VERIFY] [--s3_disable_ssl] [--输入文件INPUT_FILE] --输出路径OUTPUT_PATH aaa_users_py.py:错误:需要以下参数:--output-path

这可能意味着在此 Python 脚本中构建的管道有一个必需的自定义管道选项,其中包括一个名为 --output-path 的字段。将其视为通过从 --input-path 到 --output-path 的 ETL 数据生成作业的“模板”,您必须在将其作为作业提交之前告诉管道在哪里读取和写入。