venf2k / google-cloud-sdk

Automatically exported from code.google.com/p/google-cloud-sdk
0 stars 0 forks source link

Dataflow can't start VM #736

Closed GoogleCodeExporter closed 8 years ago

GoogleCodeExporter commented 8 years ago
What steps will reproduce the problem?

I'm trying to run the Starter Pipeline example project from Eclipse 
(BlockingDataflowPipelineRunner option).

What is the expected output? What do you see instead?

I see on the web console: 'Workflow failed"

Apr 14, 2016, 3:50:53 PM
(af9d15fab3feeeb7): Workflow failed. Causes: (d7d9df60916d7d07): Unable to 
create VMs., (d7d9df60916d7304): Error: Message: Invalid value for field 
'resource.networkInterfaces[0].network': 
'https://www.googleapis.com/compute/v1/projects/sirtrackwms/global/networks/defa
ult'. The referenced network resource cannot be found. HTTP Code: 400

Please provide any additional information below.

Eclipse Logs:

Apr 14, 2016 3:47:47 PM 
com.google.cloud.dataflow.sdk.runners.DataflowPipelineRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from 
the classpath: will stage 62 files. Enable logging at DEBUG level to see which 
files will be staged.
Apr 14, 2016 3:47:47 PM com.google.cloud.dataflow.sdk.Pipeline applyInternal
WARNING: Transform AnonymousParDo2 does not have a stable unique name. This 
will prevent updating of pipelines.
Apr 14, 2016 3:47:47 PM 
com.google.cloud.dataflow.sdk.runners.DataflowPipelineRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing 
implications related to Google Compute Engine usage and other Google Cloud 
Services.
Apr 14, 2016 3:47:47 PM com.google.cloud.dataflow.sdk.util.PackageUtil 
stageClasspathElements
INFO: Uploading 62 files from PipelineOptions.filesToStage to staging location 
to prepare for execution.
Apr 14, 2016 3:50:47 PM com.google.cloud.dataflow.sdk.util.PackageUtil 
stageClasspathElements
INFO: Uploading PipelineOptions.filesToStage complete: 62 files newly uploaded, 
0 files cached
Dataflow SDK version: 1.5.0
Apr 14, 2016 3:50:50 PM 
com.google.cloud.dataflow.sdk.runners.DataflowPipelineRunner run
INFO: To access the Dataflow monitoring console, please navigate to 
https://console.developers.google.com/project/sirtrackwms/dataflow/job/2016-04-1
3_20_50_49-2612736595740349101
Submitted job: 2016-04-13_20_50_49-2612736595740349101
Apr 14, 2016 3:50:50 PM 
com.google.cloud.dataflow.sdk.runners.DataflowPipelineRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud alpha dataflow jobs --project=sirtrackwms cancel 
2016-04-13_20_50_49-2612736595740349101
2016-04-14T03:50:52.737Z: Detail:  (8c7783c1389a5b2a): Checking required Cloud 
APIs are enabled.
2016-04-14T03:50:52.765Z: Detail:  (8c7783c1389a5249): Expanding GroupByKey 
operations into optimizable parts.
2016-04-14T03:50:52.767Z: Detail:  (8c7783c1389a5217): Lifting 
ValueCombiningMappingFns into MergeBucketsMappingFns
2016-04-14T03:50:52.769Z: Detail:  (8c7783c1389a51e5): Annotating graph with 
Autotuner information.
2016-04-14T03:50:52.778Z: Detail:  (8c7783c1389a5968): Fusing adjacent ParDo, 
Read, Write, and Flatten operations
2016-04-14T03:50:52.783Z: Detail:  (8c7783c1389a5904): Fusing consumer 
AnonymousParDo2 into AnonymousParDo
2016-04-14T03:50:52.791Z: Basic:  (8c7783c1389a583c): Worker configuration: 
n1-standard-1 in us-central1-f.
2016-04-14T03:50:52.970Z: Detail:  (8c7783c1389a5fbf): Adding StepResource 
setup and teardown to workflow graph.
2016-04-14T03:50:53.026Z: Basic:  (d20ec902213aa896): Starting 3 workers...
2016-04-14T03:50:53.102Z: Basic:  S01: (547d1b51475f9554): Executing operation 
AnonymousParDo+AnonymousParDo2
2016-04-14T03:50:53.654Z: Error:   (af9d15fab3feeeb7): Workflow failed. Causes: 
(d7d9df60916d7d07): Unable to create VMs., (d7d9df60916d7304): Error:
 Message: Invalid value for field 'resource.networkInterfaces[0].network': 'https://www.googleapis.com/compute/v1/projects/sirtrackwms/global/networks/default'. The referenced network resource cannot be found.
 HTTP Code: 400
2016-04-14T03:50:53.669Z: Detail:  (479290d2ab0eaf86): Cleaning up.
2016-04-14T03:50:53.791Z: Basic:  (479290d2ab0ea822): Stopping worker pool...
2016-04-14T03:50:53.799Z: Basic:  (479290d2ab0eabd3): Worker pool stopped.

Original issue reported on code.google.com by theb...@emanueleziglioli.it on 14 Apr 2016 at 8:54

GoogleCodeExporter commented 8 years ago
We'll take a look on our end. Thanks!

Original comment by z...@google.com on 14 Apr 2016 at 9:01

GoogleCodeExporter commented 8 years ago
I got a response from the Dataflow team:

----------------------------------------------------------------------------
That particular error looks related to some questions we've had on 
stackoverflow. The most likely cause is that the user has deleted the default 
network in their project.

For example: 
http://stackoverflow.com/questions/35590882/existing-jobs-stalled-new-jobs-faili
ng-network-issue
----------------------------------------------------------------------------
Can you try re-creating the default network?

Original comment by z...@google.com on 14 Apr 2016 at 9:25

GoogleCodeExporter commented 8 years ago
Thank you, I have no idea what the Default Network is, and surely I haven't 
deleted it. Perhaps it never existed because I was in the Beta program?
The SO answer is poor, no links to any documentation about this so called 
"Default Network"

Original comment by theb...@emanueleziglioli.it on 14 Apr 2016 at 9:30

GoogleCodeExporter commented 8 years ago
This is a better answer: 
http://stackoverflow.com/questions/36185553/why-is-my-gcp-dataflow-pipeline-unab
le-to-create-vms

Original comment by theb...@emanueleziglioli.it on 14 Apr 2016 at 9:31

GoogleCodeExporter commented 8 years ago

>gcloud compute networks create default
Created 
[https://www.googleapis.com/compute/v1/projects/sirtrackwms/global/networks/defa
ult].
NAME    MODE IPV4_RANGE GATEWAY_IPV4
default auto

Instances on this network will not be reachable until firewall rules
are created. As an example, you can allow all internal traffic between
instances as well as SSH, RDP, and ICMP by running:

$ gcloud compute firewall-rules create <FIREWALL_NAME> --network default 
--allow tcp,udp,icmp --source-ranges <IP_RANGE>

$ gcloud compute firewall-rules create <FIREWALL_NAME> --network default 
--allow tcp:22,tcp:3389,icmp

>gcloud compute routes list
NAME                           NETWORK DEST_RANGE    NEXT_HOP                 
PRIORITY
default-route-0c6e8fb938ae8c9c default 10.132.0.0/20                          
1000
default-route-2c9380c310aa7f33 default 10.142.0.0/20                          
1000
default-route-a20b34994e459131 default 10.140.0.0/20                          
1000
default-route-a248c479e53ff07f default 0.0.0.0/0     default-internet-gateway 
1000
default-route-f7912f56bfa88337 default 10.128.0.0/20                          
1000

I'll try now

Original comment by theb...@emanueleziglioli.it on 14 Apr 2016 at 9:40

GoogleCodeExporter commented 8 years ago
It worked, awesome!

Apr 15, 2016 9:44:12 AM 
com.google.cloud.dataflow.sdk.runners.BlockingDataflowPipelineRunner run
INFO: Job finished with status DONE

Original comment by theb...@emanueleziglioli.it on 14 Apr 2016 at 9:45

GoogleCodeExporter commented 8 years ago
Glad that worked!

Original comment by z...@google.com on 14 Apr 2016 at 10:35