Skip to content

Increasing Stackit Nodepool Maximum Requires Replacement Of Cluster #1364

@enricobertolotti

Description

@enricobertolotti

Description

Increasing the maximum property of a node_pool in stackit_ske_cluster leads to terraform requiring replacement of the cluster.

Steps to reproduce

resource "stackit_ske_cluster" "cluster" {
  project_id = var.project_id
  name       = "cluster-01"

  kubernetes_version_min = "1.35"

  node_pools = [
    {
      name         = "nodepool-01"
      machine_type = "c2i.2" 
      minimum      = 1
      maximum      = 1

      availability_zones = ["eu01-1"]
      volume_size = 20 
    }
  ]

  network = {
    control_plane = {
      access_scope = "PUBLIC"
    }
  }
}
  1. Terraform apply
  2. Change maximum = 2
  3. Terraform apply

Actual behavior

  # stackit_ske_cluster.cluster must be replaced
-/+ resource "stackit_ske_cluster" "cluster" {
      ~ egress_address_ranges   = [
          - ".../32",
        ] -> (known after apply)
      ~ id                      = "...,eu01,cluster-01" -> (known after apply)
      ~ kubernetes_version_used = "1.35.3" -> (known after apply)
        name                    = "cluster-01"
      ~ network                 = {
          + id            = (known after apply) # forces replacement
            # (1 unchanged attribute hidden)
        }
      ~ node_pools              = [
          ~ {
              ~ labels                  = {} -> (known after apply)
              ~ max_surge               = 1 -> (known after apply)
              ~ max_unavailable         = 0 -> (known after apply)
              ~ maximum                 = 1 -> 2
                name                    = "nodepool-01"
              ~ os_version_used         = "4459.2.3" -> (known after apply)
                # (8 unchanged attributes hidden)
            },
        ]
      ~ pod_address_ranges      = [
          - ".../13",
        ] -> (known after apply)
        # (6 unchanged attributes hidden)
    }

Expected behavior

I would expect the node pool to grow without having to provision a new cluster.

Environment

  • OS: macOS
  • Terraform version (see terraform --version): v1.14.8
  • Version of the STACKIT Terraform provider: v0.90.0

Additional information

Could this have anything to do with network changes due to the recent outage?

Feel free to add any additional information here.

Metadata

Metadata

Assignees

No one assigned

    Labels

    bugSomething isn't working

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions