'Terraform Azure AKS internal load balancer
I deployed an Azure AKS cluster via the following terraform statements into an existing vnet. It worked, the AKS cluster is created with an Azure load balancer and an public IP address assigned to it. I need a setup with an internal Azure load balancer only. How do I have to change the terraform code to only get an internal Azure load balancer? Thanks
resource "azurerm_kubernetes_cluster" "aks" {
name = "${var.tags.department}-${var.tags.stage}-${var.tags.environment}_aks"
location = var.location
resource_group_name = azurerm_resource_group.aksrg.name
dns_prefix = lower("${var.tags.department}-${var.tags.stage}-${var.tags.environment}-aks")
private_link_enabled = true
node_resource_group = "${var.tags.department}-${var.tags.stage}-${var.tags.environment}_aks_nodes_rg"
linux_profile {
admin_username = "testadmin"
ssh_key {
key_data = file("/ssh/id_rsa.pub") #ssh-keygen
}
}
default_node_pool {
name = "default"
vm_size = "Standard_DS1_v2"
enable_auto_scaling = false
enable_node_public_ip = false
node_count = 1
vnet_subnet_id = azurerm_subnet.akssubnet.id
}
network_profile {
network_plugin = "azure"
service_cidr = "172.100.0.0/24"
dns_service_ip = "172.100.0.10"
docker_bridge_cidr = "172.101.0.1/16"
load_balancer_sku = "standard"
}
service_principal {
client_id = azurerm_azuread_service_principal.aks_sp.application_id
client_secret = azurerm_azuread_service_principal_password.aks_sp_pwd.value
}
addon_profile {
kube_dashboard {
enabled = true
}
}
role_based_access_control {
enabled = false
}
}
Solution 1:[1]
For your issue, you can see that you set the network_profile of the AKS cluster like this:
network_profile {
network_plugin = "azure"
service_cidr = "172.100.0.0/24"
dns_service_ip = "172.100.0.10"
docker_bridge_cidr = "172.101.0.1/16"
load_balancer_sku = "standard"
}
And you set the property load_balancer_sku, it will create a standard type Load Balancer for you. And this property only can use the value basic and standard. So it seems it's impossible to create the AKS cluster without a Load Balancer. And you can see the issue in Github. As it said, the newest CLI also cannot help you achieve that.
If you have more questions about this issue, please let me know. I'm glad if I can help you solve the problem.
Solution 2:[2]
This is configured in the Kubernetes LoadBalancer service using annotations. Specifically you need to add service.beta.kubernetes.io/azure-load-balancer-internal: "true" to the spec of your LoadBalancer service.
This is documented here https://docs.microsoft.com/en-us/azure/aks/load-balancer-standard#additional-customizations-via-kubernetes-annotations
Solution 3:[3]
This public IP is used for outbound connectivity. To deny creating this IP, you need to specify parameter outbound_type = "userDefinedRouting", and provide your network with route table attached.
Sources
This article follows the attribution requirements of Stack Overflow and is licensed under CC BY-SA 3.0.
Source: Stack Overflow
| Solution | Source |
|---|---|
| Solution 1 | Charles Xu |
| Solution 2 | Michael |
| Solution 3 | user18698976 |
