Uploaded image for project: 'Beam'
  1. Beam
  2. BEAM-14449

Support cluster provisioning when using Flink on Dataproc

Details

    • Improvement
    • Status: Resolved
    • P2
    • Resolution: Fixed
    • None
    • 2.40.0
    • runner-py-interactive
    • None

    Description

      Provide the capability for the user to explicitly provision a cluster.

      Current implementation provisions each cluster at the location specified by GoogleCloudOptions using 3 worker nodes. There is no explicit API to configure the number or shape of workers.

      We could use the WorkerOptions to allow customers to explicitly provision a cluster and expose an explicit API (with UX in notebook extension) for customers to change the size of a cluster connected with their notebook (until we have an auto scaling solution with Dataproc for Flink).

      The API looks like this when configuring the workers for a dataproc cluster when creating it:

      An example request setting the masterConfig and workerConfig:

      Attachments

        Issue Links

          Activity

            People

              ningk Ning
              ningk Ning
              Votes:
              0 Vote for this issue
              Watchers:
              1 Start watching this issue

              Dates

                Created:
                Updated:
                Resolved:

                Time Tracking

                  Estimated:
                  Original Estimate - Not Specified
                  Not Specified
                  Remaining:
                  Remaining Estimate - 0h
                  0h
                  Logged:
                  Time Spent - 2h 20m
                  2h 20m