Skip to content

adam-stasiak/dataproc-initialization-actions

This branch is 5 commits ahead of, 693 commits behind GoogleCloudDataproc/initialization-actions:master.

Folders and files

NameName
Last commit message
Last commit date

Latest commit

author
Adam Stasiak
Feb 23, 2018
0da8830 · Feb 23, 2018
Feb 5, 2018
Feb 8, 2018
Jan 29, 2018
Apr 6, 2017
Feb 23, 2018
Jan 29, 2018
Mar 27, 2017
Feb 12, 2018
Sep 28, 2017
Jan 29, 2018
Jan 29, 2018
Feb 12, 2018
Sep 12, 2016
Oct 13, 2017
Sep 27, 2016
Jan 29, 2018
Jan 29, 2018
Jun 27, 2016
Jan 29, 2018
Jan 29, 2018
Jan 29, 2018
Oct 12, 2015
Sep 29, 2017
Oct 12, 2015

Repository files navigation

Cloud Dataproc Initialization Actions

When creating a Google Cloud Dataproc cluster, you can specify initialization actions in executables and/or scripts that Cloud Dataproc will run on all nodes in your Cloud Dataproc cluster immediately after the cluster is set up.

How initialization actions are used

Initialization actions are stored in a Google Cloud Storage bucket and can be passed as a parameter to the gcloud command or the clusters.create API when creating a Cloud Dataproc cluster. For example, to specify an initialization action when creating a cluster with the gcloud command, you can run:

gcloud dataproc clusters create CLUSTER-NAME
[--initialization-actions [GCS_URI,...]]
[--initialization-action-timeout TIMEOUT]

For convenience, copies of initialization actions in this repository are stored in the following Cloud Storage bucket, which is publicly accessible:

gs://dataproc-initialization-actions

The folder structure of this Cloud Storage bucket mirrors this repository. You should be able to use this Cloud Storage bucket (and the initialization scripts within it) for your clusters.

Why these samples are provided

These samples are provided to show how various packages and components can be installed on Cloud Dataproc clusters. You should understand how these samples work before running them on your clusters. The initialization actions provided in this repository are provided without support and you use them at your own risk.

Actions provided

This repository presently offers the following actions for use with Cloud Dataproc clusters.

Initialization actions on single node clusters

Single Node clusters have dataproc-role set to Master and dataproc-worker-count set to 0. Most of the initialization actions in this repository should work out of the box, as they run only on the master. Examples include notebooks (such as Apache Zeppelin) and libraries (such as Apache Tez). Actions that run on all nodes of the cluster (such as cloud-sql-proxy) similarly work out of the box.

Some initialization actions are known not to work on Single Node clusters. All of these expect to have daemons on multiple nodes.

  • Apache Drill
  • Apache Flink
  • Apache Kafka
  • Presto
  • Apache Zookeeper

Feel free to send pull requests or file issues if you have a good use case for running one of these actions on a Single Node cluster.

For more information

For more information, review the Cloud Dataproc documentation. You can also pose questions to the Stack Overflow community with the tag google-cloud-dataproc. See our other Google Cloud Platform github repos for sample applications and scaffolding for other frameworks and use cases.

Contributing changes

Licensing

About

Run in all nodes of your cluster before the cluster starts - lets you customize your cluster

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Shell 98.1%
  • Python 1.9%