As per the documentation, I am trying to use the gcloud command to setup the GCSLoader dataflow job, I am using the below command:
gcloud dataflow jobs run snowplowdemo-collectorcloudstorageloader-job1
–gcs-location gs://sp-hosted-assets/4-storage/snowplow-google-cloud-storage-loader/0.3.1/SnowplowGoogleCloudStorageLoaderTemplate-0.3.1
–project=project_name
–region=us-east1
–worker-zone=us-east1-a
–num-workers=1
–service-account-email=service_account_email
–parameters
inputSubscription=subscription_name ,
outputDirectory=gs://bucket_name/BadRecords/,
outputFilenamePrefix=output,
shardTemplate=-W-P-SSSSS-of-NNNNN,
outputFilenameSuffix=.txt,
windowDuration=5,
compression=none,
numShards=1
The dataflow goes in the failed state with the error:
The supplied parameters for autoscaling and the worker pool size are incorrect. Causes: Streaming autoscaling requires maxNumWorkers to be set.
Can you please suggest on the additional parameter that needs to be set.