@@ -99,7 +99,7 @@ resource "azurerm_kubernetes_cluster" "jupyterhub" {
99
99
#
100
100
# Most changes to this node pool forces a replace operation on the entire
101
101
# cluster. This can be avoided with v3.47.0+ of this provider by declaring
102
- # temporary_name_for_rotation = "core-b ".
102
+ # temporary_name_for_rotation = "coreb ".
103
103
#
104
104
# ref: https://github.com/hashicorp/terraform-provider-azurerm/pull/20628
105
105
# ref: https://registry.terraform.io/providers/hashicorp/azurerm/latest/docs/resources/kubernetes_cluster#temporary_name_for_rotation.
@@ -108,9 +108,11 @@ resource "azurerm_kubernetes_cluster" "jupyterhub" {
108
108
name = var. core_node_pool . name
109
109
vm_size = var. core_node_pool . vm_size
110
110
os_disk_size_gb = var. core_node_pool . os_disk_size_gb
111
- enable_auto_scaling = true
112
- min_count = var. core_node_pool . min
113
- max_count = var. core_node_pool . max
111
+ enable_auto_scaling = var. core_node_pool . enable_auto_scaling
112
+ min_count = var. core_node_pool . enable_auto_scaling ? var. core_node_pool . min : null
113
+ max_count = var. core_node_pool . enable_auto_scaling ? var. core_node_pool . max : null
114
+ node_count = var. core_node_pool . node_count
115
+ kubelet_disk_type = var. core_node_pool . kubelet_disk_type
114
116
vnet_subnet_id = azurerm_subnet. node_subnet . id
115
117
node_labels = merge ({
116
118
" hub.jupyter.org/node-purpose" = " core" ,
@@ -147,13 +149,14 @@ resource "azurerm_kubernetes_cluster" "jupyterhub" {
147
149
148
150
149
151
resource "azurerm_kubernetes_cluster_node_pool" "user_pool" {
150
- for_each = var. notebook_nodes
152
+ for_each = var. user_node_pools
151
153
152
154
name = coalesce (each. value . name , each. key )
153
155
kubernetes_cluster_id = azurerm_kubernetes_cluster. jupyterhub . id
154
156
enable_auto_scaling = true
155
- os_disk_size_gb = 200
157
+ os_disk_size_gb = each . value . os_disk_size_gb
156
158
vnet_subnet_id = azurerm_subnet. node_subnet . id
159
+ kubelet_disk_type = each. value . kubelet_disk_type
157
160
158
161
orchestrator_version = each. value . kubernetes_version == " " ? var. kubernetes_version : each. value . kubernetes_version
159
162
@@ -173,14 +176,14 @@ resource "azurerm_kubernetes_cluster_node_pool" "user_pool" {
173
176
}
174
177
175
178
resource "azurerm_kubernetes_cluster_node_pool" "dask_pool" {
176
- # If dask_nodes is set, we use that. If it isn't, we use notebook_nodes .
177
- # This lets us set dask_nodes to an empty array to get no dask nodes
178
- for_each = var. dask_nodes
179
+ # If dask_node_pools is set, we use that. If it isn't, we use user_node_pools .
180
+ # This lets us set dask_node_pools to an empty array to get no dask nodes
181
+ for_each = var. dask_node_pools
179
182
180
183
name = " dask${ each . key } "
181
184
kubernetes_cluster_id = azurerm_kubernetes_cluster. jupyterhub . id
182
185
enable_auto_scaling = true
183
- os_disk_size_gb = 200
186
+ os_disk_size_gb = each . value . os_disk_size_gb
184
187
vnet_subnet_id = azurerm_subnet. node_subnet . id
185
188
186
189
orchestrator_version = each. value . kubernetes_version == " " ? var. kubernetes_version : each. value . kubernetes_version
0 commit comments