Apache_beam--python --error: the following arguments are required: --output-path
Apache_beam--python --error: the following arguments are required: --output-path
当 运行 我的代码在 python 或协作者出现以下错误时,我从 apache beam 安装所有库,有人在一瞬间给出了这个错误或者知道它..
用法:aaa_users_py.py [-h] [--运行ner RUNNER] [--streaming] [--resource_hint RESOURCE_HINTS] [--beam_services BEAM_SERVICES]
[--type_check_strictness {ALL_REQUIRED,DEFAULT_TO_ANY}] [--type_check_additional TYPE_CHECK_ADDITIONAL]
[--no_pipeline_type_check] [--runtime_type_check] [--performance_runtime_type_check]
[--allow_non_deterministic_key_coders] [--allow_unsafe_triggers]
[--no_direct_runner_use_stacked_bundle] [--direct_runner_bundle_repeat DIRECT_RUNNER_BUNDLE_REPEAT]
[--direct_num_workers DIRECT_NUM_WORKERS]
[--direct_running_mode {in_memory,multi_threading,multi_processing}]
[--dataflow_endpoint DATAFLOW_ENDPOINT] [--项目项目] [--job_name JOB_NAME]
[--staging_location STAGING_LOCATION] [--temp_location TEMP_LOCATION] [--region REGION]
[--service_account_email SERVICE_ACCOUNT_EMAIL] [--no_auth] [--template_location TEMPLATE_LOCATION]
[--标签标签] [--更新] [--transform_name_mapping TRANSFORM_NAME_MAPPING]
[--enable_streaming_engine] [--dataflow_kms_key DATAFLOW_KMS_KEY]
[--create_from_snapshot CREATE_FROM_SNAPSHOT] [--flexrs_goal {COST_OPTIMIZED,SPEED_OPTIMIZED}]
[--dataflow_service_option DATAFLOW_SERVICE_OPTIONS] [--enable_hot_key_logging]
[--enable_artifact_caching] [--impersonate_service_account IMPERSONATE_SERVICE_ACCOUNT]
[--hdfs_host HDFS_HOST] [--hdfs_port HDFS_PORT] [--hdfs_user HDFS_USER] [--hdfs_full_urls]
[--num_workers NUM_WORKERS] [--max_num_workers MAX_NUM_WORKERS]
[--autoscaling_algorithm {NONE,THROUGHPUT_BASED}] [--worker_machine_type MACHINE_TYPE]
[--disk_size_gb DISK_SIZE_GB] [--worker_disk_type DISK_TYPE] [--worker_region WORKER_REGION]
[--worker_zone WORKER_ZONE] [--zone ZONE] [--network NETWORK] [--subnetwork SUBNETWORK]
[--worker_harness_container_image WORKER_HARNESS_CONTAINER_IMAGE]
[--sdk_container_image SDK_CONTAINER_IMAGE]
[--sdk_harness_container_image_overrides SDK_HARNESS_CONTAINER_IMAGE_OVERRIDES] [--use_public_ips]
[--no_use_public_ips] [--min_cpu_platform MIN_CPU_PLATFORM] [--dataflow_worker_jar DATAFLOW_WORKER_JAR]
[--dataflow_job_file DATAFLOW_JOB_FILE] [--实验实验]
[--number_of_worker_harness_threads NUMBER_OF_WORKER_HARNESS_THREADS] [--profile_cpu]
[--profile_memory] [--profile_location PROFILE_LOCATION] [--profile_sample_rate PROFILE_SAMPLE_RATE]
[--requirements_file REQUIREMENTS_FILE] [--requirements_cache REQUIREMENTS_CACHE]
[--requirements_cache_only_sources] [--setup_file SETUP_FILE] [--beam_plugin BEAM_PLUGINS]
[--pickle_library {cloudpickle,default,dill}] [--save_main_session] [--sdk_location SDK_LOCATION]
[--extra_package EXTRA_PACKAGES] [--prebuild_sdk_container_engine PREBUILD_SDK_CONTAINER_ENGINE]
[--prebuild_sdk_container_base_image PREBUILD_SDK_CONTAINER_BASE_IMAGE]
[--cloud_build_machine_type CLOUD_BUILD_MACHINE_TYPE]
[--docker_registry_push_url DOCKER_REGISTRY_PUSH_URL] [--job_endpoint JOB_ENDPOINT]
[--artifact_endpoint ARTIFACT_ENDPOINT] [--job_server_timeout JOB_SERVER_TIMEOUT]
[--environment_type ENVIRONMENT_TYPE] [--environment_config ENVIRONMENT_CONFIG]
[--environment_option ENVIRONMENT_OPTIONS] [--sdk_worker_parallelism SDK_WORKER_PARALLELISM]
[--environment_cache_millis ENVIRONMENT_CACHE_MILLIS] [--output_executable_path OUTPUT_EXECUTABLE_PATH]
[--artifacts_dir ARTIFACTS_DIR] [--job_port JOB_PORT] [--artifact_port ARTIFACT_PORT]
[--expansion_port EXPANSION_PORT] [--job_server_java_launcher JOB_SERVER_JAVA_LAUNCHER]
[--job_server_jvm_properties JOB_SERVER_JVM_PROPERTIES] [--flink_master FLINK_MASTER]
[--flink_version {1.12,1.13,1.14}] [--flink_job_server_jar FLINK_JOB_SERVER_JAR]
[--flink_submit_uber_jar] [--spark_master_url SPARK_MASTER_URL]
[--spark_job_server_jar SPARK_JOB_SERVER_JAR] [--spark_submit_uber_jar]
[--spark_rest_url SPARK_REST_URL] [--spark_version {2,3}] [--on_success_matcher ON_SUCCESS_MATCHER]
[--dry_run DRY_RUN] [--wait_until_finish_duration WAIT_UNTIL_FINISH_DURATION]
[--pubsub_root_url PUBSUBROOTURL] [--s3_access_key_id S3_ACCESS_KEY_ID]
[--s3_secret_access_key S3_SECRET_ACCESS_KEY] [--s3_session_token S3_SESSION_TOKEN]
[--s3_endpoint_url S3_ENDPOINT_URL] [--s3_region_name S3_REGION_NAME] [--s3_api_version S3_API_VERSION]
[--s3_verify S3_VERIFY] [--s3_disable_ssl] [--输入文件INPUT_FILE] --输出路径OUTPUT_PATH
aaa_users_py.py:错误:需要以下参数:--output-path
这可能意味着在此 Python 脚本中构建的管道有一个必需的自定义管道选项,其中包括一个名为 --output-path 的字段。将其视为通过从 --input-path 到 --output-path 的 ETL 数据生成作业的“模板”,您必须在将其作为作业提交之前告诉管道在哪里读取和写入。
当 运行 我的代码在 python 或协作者出现以下错误时,我从 apache beam 安装所有库,有人在一瞬间给出了这个错误或者知道它..
用法:aaa_users_py.py [-h] [--运行ner RUNNER] [--streaming] [--resource_hint RESOURCE_HINTS] [--beam_services BEAM_SERVICES]
[--type_check_strictness {ALL_REQUIRED,DEFAULT_TO_ANY}] [--type_check_additional TYPE_CHECK_ADDITIONAL]
[--no_pipeline_type_check] [--runtime_type_check] [--performance_runtime_type_check]
[--allow_non_deterministic_key_coders] [--allow_unsafe_triggers]
[--no_direct_runner_use_stacked_bundle] [--direct_runner_bundle_repeat DIRECT_RUNNER_BUNDLE_REPEAT]
[--direct_num_workers DIRECT_NUM_WORKERS]
[--direct_running_mode {in_memory,multi_threading,multi_processing}]
[--dataflow_endpoint DATAFLOW_ENDPOINT] [--项目项目] [--job_name JOB_NAME]
[--staging_location STAGING_LOCATION] [--temp_location TEMP_LOCATION] [--region REGION]
[--service_account_email SERVICE_ACCOUNT_EMAIL] [--no_auth] [--template_location TEMPLATE_LOCATION]
[--标签标签] [--更新] [--transform_name_mapping TRANSFORM_NAME_MAPPING]
[--enable_streaming_engine] [--dataflow_kms_key DATAFLOW_KMS_KEY]
[--create_from_snapshot CREATE_FROM_SNAPSHOT] [--flexrs_goal {COST_OPTIMIZED,SPEED_OPTIMIZED}]
[--dataflow_service_option DATAFLOW_SERVICE_OPTIONS] [--enable_hot_key_logging]
[--enable_artifact_caching] [--impersonate_service_account IMPERSONATE_SERVICE_ACCOUNT]
[--hdfs_host HDFS_HOST] [--hdfs_port HDFS_PORT] [--hdfs_user HDFS_USER] [--hdfs_full_urls]
[--num_workers NUM_WORKERS] [--max_num_workers MAX_NUM_WORKERS]
[--autoscaling_algorithm {NONE,THROUGHPUT_BASED}] [--worker_machine_type MACHINE_TYPE]
[--disk_size_gb DISK_SIZE_GB] [--worker_disk_type DISK_TYPE] [--worker_region WORKER_REGION]
[--worker_zone WORKER_ZONE] [--zone ZONE] [--network NETWORK] [--subnetwork SUBNETWORK]
[--worker_harness_container_image WORKER_HARNESS_CONTAINER_IMAGE]
[--sdk_container_image SDK_CONTAINER_IMAGE]
[--sdk_harness_container_image_overrides SDK_HARNESS_CONTAINER_IMAGE_OVERRIDES] [--use_public_ips]
[--no_use_public_ips] [--min_cpu_platform MIN_CPU_PLATFORM] [--dataflow_worker_jar DATAFLOW_WORKER_JAR]
[--dataflow_job_file DATAFLOW_JOB_FILE] [--实验实验]
[--number_of_worker_harness_threads NUMBER_OF_WORKER_HARNESS_THREADS] [--profile_cpu]
[--profile_memory] [--profile_location PROFILE_LOCATION] [--profile_sample_rate PROFILE_SAMPLE_RATE]
[--requirements_file REQUIREMENTS_FILE] [--requirements_cache REQUIREMENTS_CACHE]
[--requirements_cache_only_sources] [--setup_file SETUP_FILE] [--beam_plugin BEAM_PLUGINS]
[--pickle_library {cloudpickle,default,dill}] [--save_main_session] [--sdk_location SDK_LOCATION]
[--extra_package EXTRA_PACKAGES] [--prebuild_sdk_container_engine PREBUILD_SDK_CONTAINER_ENGINE]
[--prebuild_sdk_container_base_image PREBUILD_SDK_CONTAINER_BASE_IMAGE]
[--cloud_build_machine_type CLOUD_BUILD_MACHINE_TYPE]
[--docker_registry_push_url DOCKER_REGISTRY_PUSH_URL] [--job_endpoint JOB_ENDPOINT]
[--artifact_endpoint ARTIFACT_ENDPOINT] [--job_server_timeout JOB_SERVER_TIMEOUT]
[--environment_type ENVIRONMENT_TYPE] [--environment_config ENVIRONMENT_CONFIG]
[--environment_option ENVIRONMENT_OPTIONS] [--sdk_worker_parallelism SDK_WORKER_PARALLELISM]
[--environment_cache_millis ENVIRONMENT_CACHE_MILLIS] [--output_executable_path OUTPUT_EXECUTABLE_PATH]
[--artifacts_dir ARTIFACTS_DIR] [--job_port JOB_PORT] [--artifact_port ARTIFACT_PORT]
[--expansion_port EXPANSION_PORT] [--job_server_java_launcher JOB_SERVER_JAVA_LAUNCHER]
[--job_server_jvm_properties JOB_SERVER_JVM_PROPERTIES] [--flink_master FLINK_MASTER]
[--flink_version {1.12,1.13,1.14}] [--flink_job_server_jar FLINK_JOB_SERVER_JAR]
[--flink_submit_uber_jar] [--spark_master_url SPARK_MASTER_URL]
[--spark_job_server_jar SPARK_JOB_SERVER_JAR] [--spark_submit_uber_jar]
[--spark_rest_url SPARK_REST_URL] [--spark_version {2,3}] [--on_success_matcher ON_SUCCESS_MATCHER]
[--dry_run DRY_RUN] [--wait_until_finish_duration WAIT_UNTIL_FINISH_DURATION]
[--pubsub_root_url PUBSUBROOTURL] [--s3_access_key_id S3_ACCESS_KEY_ID]
[--s3_secret_access_key S3_SECRET_ACCESS_KEY] [--s3_session_token S3_SESSION_TOKEN]
[--s3_endpoint_url S3_ENDPOINT_URL] [--s3_region_name S3_REGION_NAME] [--s3_api_version S3_API_VERSION]
[--s3_verify S3_VERIFY] [--s3_disable_ssl] [--输入文件INPUT_FILE] --输出路径OUTPUT_PATH
aaa_users_py.py:错误:需要以下参数:--output-path
这可能意味着在此 Python 脚本中构建的管道有一个必需的自定义管道选项,其中包括一个名为 --output-path 的字段。将其视为通过从 --input-path 到 --output-path 的 ETL 数据生成作业的“模板”,您必须在将其作为作业提交之前告诉管道在哪里读取和写入。