-
Notifications
You must be signed in to change notification settings - Fork 879
gcp: Add nodepool for tests #7943
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
/hold |
Ran Plan for dir: Show OutputTerraform used the selected providers to generate the following execution
plan. Resource actions are indicated with the following symbols:
+ create
Terraform will perform the following actions:
# module.prow_build_nodepool_c4_highmem_8_localssd.google_container_node_pool.node_pool will be created
+ resource "google_container_node_pool" "node_pool" {
+ cluster = (sensitive value)
+ id = (known after apply)
+ initial_node_count = 1
+ instance_group_urls = (known after apply)
+ location = (sensitive value)
+ managed_instance_group_urls = (known after apply)
+ max_pods_per_node = (known after apply)
+ name = (known after apply)
+ name_prefix = "pool6-"
+ node_count = (known after apply)
+ node_locations = [
+ "us-central1-b",
+ "us-central1-c",
+ "us-central1-f",
]
+ operation = (known after apply)
+ project = "k8s-infra-prow-build"
+ version = (known after apply)
+ autoscaling {
+ location_policy = (known after apply)
+ max_node_count = 80
+ min_node_count = 1
}
+ management {
+ auto_repair = true
+ auto_upgrade = true
}
+ network_config (known after apply)
+ node_config {
+ disk_size_gb = 100
+ disk_type = "hyperdisk-balanced"
+ effective_taints = (known after apply)
+ guest_accelerator = (known after apply)
+ image_type = "COS_CONTAINERD"
+ labels = (known after apply)
+ local_ssd_count = (known after apply)
+ logging_variant = (known after apply)
+ machine_type = "c4-highmem-8"
+ metadata = {
+ "disable-legacy-endpoints" = "true"
}
+ min_cpu_platform = (known after apply)
+ oauth_scopes = [
+ "https://www.googleapis.com/auth/cloud-platform",
]
+ preemptible = false
+ service_account = "gke-nodes-prow-build@k8s-infra-prow-build.iam.gserviceaccount.com"
+ spot = false
+ confidential_nodes (known after apply)
+ gcfs_config (known after apply)
+ kubelet_config (known after apply)
+ shielded_instance_config (known after apply)
+ taint {
+ effect = "NO_SCHEDULE"
+ key = "dedicated"
+ value = "sig-testing"
}
+ workload_metadata_config {
+ mode = "GKE_METADATA"
}
}
+ upgrade_settings (known after apply)
}
Plan: 1 to add, 0 to change, 0 to destroy.
Plan: 1 to add, 0 to change, 0 to destroy.
|
4c22d88
to
993f75a
Compare
Ran Plan for dir: Show OutputTerraform used the selected providers to generate the following execution
plan. Resource actions are indicated with the following symbols:
+ create
Terraform will perform the following actions:
# module.prow_build_nodepool_c4_highmem_8_localssd.google_container_node_pool.node_pool will be created
+ resource "google_container_node_pool" "node_pool" {
+ cluster = (sensitive value)
+ id = (known after apply)
+ initial_node_count = 1
+ instance_group_urls = (known after apply)
+ location = (sensitive value)
+ managed_instance_group_urls = (known after apply)
+ max_pods_per_node = (known after apply)
+ name = (known after apply)
+ name_prefix = "pool6-"
+ node_count = (known after apply)
+ node_locations = [
+ "us-central1-b",
+ "us-central1-c",
+ "us-central1-f",
]
+ operation = (known after apply)
+ project = "k8s-infra-prow-build"
+ version = (known after apply)
+ autoscaling {
+ location_policy = (known after apply)
+ max_node_count = 80
+ min_node_count = 1
}
+ management {
+ auto_repair = true
+ auto_upgrade = true
}
+ network_config (known after apply)
+ node_config {
+ disk_size_gb = 500
+ disk_type = "hyperdisk-balanced"
+ effective_taints = (known after apply)
+ guest_accelerator = (known after apply)
+ image_type = "COS_CONTAINERD"
+ labels = (known after apply)
+ local_ssd_count = (known after apply)
+ logging_variant = (known after apply)
+ machine_type = "c4-highmem-8"
+ metadata = {
+ "disable-legacy-endpoints" = "true"
}
+ min_cpu_platform = (known after apply)
+ oauth_scopes = [
+ "https://www.googleapis.com/auth/cloud-platform",
]
+ preemptible = false
+ service_account = "gke-nodes-prow-build@k8s-infra-prow-build.iam.gserviceaccount.com"
+ spot = false
+ confidential_nodes (known after apply)
+ gcfs_config (known after apply)
+ kubelet_config (known after apply)
+ shielded_instance_config (known after apply)
+ taint {
+ effect = "NO_SCHEDULE"
+ key = "dedicated"
+ value = "sig-testing"
}
+ workload_metadata_config {
+ mode = "GKE_METADATA"
}
}
+ upgrade_settings (known after apply)
}
Plan: 1 to add, 0 to change, 0 to destroy.
Plan: 1 to add, 0 to change, 0 to destroy.
|
993f75a
to
a29a7c6
Compare
Ran Plan for dir: Show OutputTerraform used the selected providers to generate the following execution
plan. Resource actions are indicated with the following symbols:
+ create
Terraform will perform the following actions:
# module.prow_build_nodepool_c4_highmem_8_localssd.google_container_node_pool.node_pool will be created
+ resource "google_container_node_pool" "node_pool" {
+ cluster = (sensitive value)
+ id = (known after apply)
+ initial_node_count = 1
+ instance_group_urls = (known after apply)
+ location = (sensitive value)
+ managed_instance_group_urls = (known after apply)
+ max_pods_per_node = (known after apply)
+ name = (known after apply)
+ name_prefix = "sig-testing-pool6-"
+ node_count = (known after apply)
+ node_locations = [
+ "us-central1-b",
+ "us-central1-c",
+ "us-central1-f",
]
+ operation = (known after apply)
+ project = "k8s-infra-prow-build"
+ version = (known after apply)
+ autoscaling {
+ location_policy = (known after apply)
+ max_node_count = 80
+ min_node_count = 1
}
+ management {
+ auto_repair = true
+ auto_upgrade = true
}
+ network_config (known after apply)
+ node_config {
+ disk_size_gb = 500
+ disk_type = "hyperdisk-balanced"
+ effective_taints = (known after apply)
+ guest_accelerator = (known after apply)
+ image_type = "COS_CONTAINERD"
+ labels = (known after apply)
+ local_ssd_count = (known after apply)
+ logging_variant = (known after apply)
+ machine_type = "c4-highmem-8"
+ metadata = {
+ "disable-legacy-endpoints" = "true"
}
+ min_cpu_platform = (known after apply)
+ oauth_scopes = [
+ "https://www.googleapis.com/auth/cloud-platform",
]
+ preemptible = false
+ service_account = "gke-nodes-prow-build@k8s-infra-prow-build.iam.gserviceaccount.com"
+ spot = false
+ confidential_nodes (known after apply)
+ gcfs_config (known after apply)
+ kubelet_config (known after apply)
+ shielded_instance_config (known after apply)
+ taint {
+ effect = "NO_SCHEDULE"
+ key = "dedicated"
+ value = "sig-testing"
}
+ workload_metadata_config {
+ mode = "GKE_METADATA"
}
}
+ upgrade_settings (known after apply)
}
Plan: 1 to add, 0 to change, 0 to destroy.
Plan: 1 to add, 0 to change, 0 to destroy.
|
atlantis apply |
Ran Apply for dir: Apply Error Show Output
|
Related to: - kubernetes#2438 Setup a new nodepool with taints so we can schedule specific tests on it for evaluation before we move all the test to a new nodepool. This nodepool will also use COS and cgroups v2. Signed-off-by: Arnaud Meukam <[email protected]>
a29a7c6
to
e135ed4
Compare
Ran Plan for dir: Show OutputTerraform used the selected providers to generate the following execution
plan. Resource actions are indicated with the following symbols:
+ create
Terraform will perform the following actions:
# module.prow_build_nodepool_c4_highmem_8_localssd.google_container_node_pool.node_pool will be created
+ resource "google_container_node_pool" "node_pool" {
+ cluster = (sensitive value)
+ id = (known after apply)
+ initial_node_count = 1
+ instance_group_urls = (known after apply)
+ location = (sensitive value)
+ managed_instance_group_urls = (known after apply)
+ max_pods_per_node = (known after apply)
+ name = (known after apply)
+ name_prefix = "pool6-"
+ node_count = (known after apply)
+ node_locations = [
+ "us-central1-b",
+ "us-central1-c",
+ "us-central1-f",
]
+ operation = (known after apply)
+ project = "k8s-infra-prow-build"
+ version = (known after apply)
+ autoscaling {
+ location_policy = (known after apply)
+ max_node_count = 80
+ min_node_count = 1
}
+ management {
+ auto_repair = true
+ auto_upgrade = true
}
+ network_config (known after apply)
+ node_config {
+ disk_size_gb = 500
+ disk_type = "hyperdisk-balanced"
+ effective_taints = (known after apply)
+ guest_accelerator = (known after apply)
+ image_type = "COS_CONTAINERD"
+ labels = (known after apply)
+ local_ssd_count = (known after apply)
+ logging_variant = (known after apply)
+ machine_type = "c4-highmem-8"
+ metadata = {
+ "disable-legacy-endpoints" = "true"
}
+ min_cpu_platform = (known after apply)
+ oauth_scopes = [
+ "https://www.googleapis.com/auth/cloud-platform",
]
+ preemptible = false
+ service_account = "gke-nodes-prow-build@k8s-infra-prow-build.iam.gserviceaccount.com"
+ spot = false
+ confidential_nodes (known after apply)
+ gcfs_config (known after apply)
+ kubelet_config (known after apply)
+ shielded_instance_config (known after apply)
+ taint {
+ effect = "NO_SCHEDULE"
+ key = "dedicated"
+ value = "sig-testing"
}
+ workload_metadata_config {
+ mode = "GKE_METADATA"
}
}
+ upgrade_settings (known after apply)
}
Plan: 1 to add, 0 to change, 0 to destroy.
Plan: 1 to add, 0 to change, 0 to destroy.
|
atlantis apply |
Ran Apply for dir: Show Outputmodule.prow_build_nodepool_c4_highmem_8_localssd.google_container_node_pool.node_pool: Creating...
module.prow_build_nodepool_c4_highmem_8_localssd.google_container_node_pool.node_pool: Still creating... [10s elapsed]
module.prow_build_nodepool_c4_highmem_8_localssd.google_container_node_pool.node_pool: Still creating... [20s elapsed]
module.prow_build_nodepool_c4_highmem_8_localssd.google_container_node_pool.node_pool: Still creating... [30s elapsed]
module.prow_build_nodepool_c4_highmem_8_localssd.google_container_node_pool.node_pool: Still creating... [40s elapsed]
module.prow_build_nodepool_c4_highmem_8_localssd.google_container_node_pool.node_pool: Still creating... [50s elapsed]
module.prow_build_nodepool_c4_highmem_8_localssd.google_container_node_pool.node_pool: Still creating... [1m1s elapsed]
module.prow_build_nodepool_c4_highmem_8_localssd.google_container_node_pool.node_pool: Still creating... [1m11s elapsed]
module.prow_build_nodepool_c4_highmem_8_localssd.google_container_node_pool.node_pool: Still creating... [1m21s elapsed]
module.prow_build_nodepool_c4_highmem_8_localssd.google_container_node_pool.node_pool: Still creating... [1m31s elapsed]
module.prow_build_nodepool_c4_highmem_8_localssd.google_container_node_pool.node_pool: Still creating... [1m41s elapsed]
module.prow_build_nodepool_c4_highmem_8_localssd.google_container_node_pool.node_pool: Still creating... [1m51s elapsed]
module.prow_build_nodepool_c4_highmem_8_localssd.google_container_node_pool.node_pool: Still creating... [2m1s elapsed]
module.prow_build_nodepool_c4_highmem_8_localssd.google_container_node_pool.node_pool: Still creating... [2m11s elapsed]
module.prow_build_nodepool_c4_highmem_8_localssd.google_container_node_pool.node_pool: Still creating... [2m21s elapsed]
module.prow_build_nodepool_c4_highmem_8_localssd.google_container_node_pool.node_pool: Still creating... [2m31s elapsed]
module.prow_build_nodepool_c4_highmem_8_localssd.google_container_node_pool.node_pool: Still creating... [2m41s elapsed]
module.prow_build_nodepool_c4_highmem_8_localssd.google_container_node_pool.node_pool: Still creating... [2m51s elapsed]
module.prow_build_nodepool_c4_highmem_8_localssd.google_container_node_pool.node_pool: Still creating... [3m1s elapsed]
module.prow_build_nodepool_c4_highmem_8_localssd.google_container_node_pool.node_pool: Still creating... [3m11s elapsed]
module.prow_build_nodepool_c4_highmem_8_localssd.google_container_node_pool.node_pool: Still creating... [3m21s elapsed]
module.prow_build_nodepool_c4_highmem_8_localssd.google_container_node_pool.node_pool: Still creating... [3m31s elapsed]
module.prow_build_nodepool_c4_highmem_8_localssd.google_container_node_pool.node_pool: Still creating... [3m41s elapsed]
module.prow_build_nodepool_c4_highmem_8_localssd.google_container_node_pool.node_pool: Creation complete after 3m45s [id=projects/k8s-infra-prow-build/locations/us-central1/clusters/prow-build/nodePools/pool6-20250327232037500200000001]
Apply complete! Resources: 1 added, 0 changed, 0 destroyed. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
[APPROVALNOTIFIER] This PR is APPROVED This pull-request has been approved by: ameukam, upodroid The full list of commands accepted by this bot can be found here. The pull request process is described here
Needs approval from an approver in each of these files:
Approvers can indicate their approval by writing |
Canceling the hold as the infra change has been applied /hold cancel |
Locks and plans deleted for the projects and workspaces modified in this pull request:
|
Follow-up: - kubernetes/k8s.io#7943 move some prowjobs to a dedicated nodepool in order to evaluate new instances. Signed-off-by: Arnaud Meukam <[email protected]>
Follow-up of: - kubernetes/k8s.io#7943 Move ci-test-infra-continuous-test to a dedicated nodepool Signed-off-by: Arnaud Meukam <[email protected]>
Follow-up of: - kubernetes/k8s.io#7943 Move ci-test-infra-continuous-test to a dedicated nodepool Signed-off-by: Arnaud Meukam <[email protected]>
disk_size_gb = 500 | ||
disk_type = "hyperdisk-balanced" | ||
service_account = module.prow_build_cluster.cluster_node_sa.email | ||
taints = [{ key = "dedicated", value = "sig-testing", effect = "NO_SCHEDULE" }] |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
weirdly I don't see this taint on the actual cluster nodes?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
$gcloud container node-pools describe pool6-20250327232037500200000001 --project k8s-infra-prow-build --format='value(config.taints)' --location us-central1 --cluster prow-build
{'effect': 'NO_SCHEDULE', 'key': 'dedicated', 'value': 'sig-testing'}
Related to:
Setup a new nodepool with taints so we can schedule specific tests on it for evaluation before we move all the test to a new nodepool. This nodepool will also use COS and cgroups v2.