terraform-hcloud-kube-hetzner/main.tf
2022-04-13 15:59:03 +02:00

96 lines
2.8 KiB
HCL

resource "random_password" "k3s_token" {
length = 48
special = false
}
resource "hcloud_ssh_key" "k3s" {
name = var.cluster_name
public_key = local.ssh_public_key
}
resource "hcloud_network" "k3s" {
name = var.cluster_name
ip_range = local.network_ipv4_cidr
}
# We start from the end of the subnets cird array,
# as we would have fewer control plane nodepools, than angent ones.
resource "hcloud_network_subnet" "control_plane" {
count = length(var.control_plane_nodepools)
network_id = hcloud_network.k3s.id
type = "cloud"
network_zone = var.network_region
ip_range = local.network_ipv4_subnets[255 - count.index]
}
# Here we start at the beginning of the subnets cird array
resource "hcloud_network_subnet" "agent" {
count = length(var.agent_nodepools)
network_id = hcloud_network.k3s.id
type = "cloud"
network_zone = var.network_region
ip_range = local.network_ipv4_subnets[count.index]
}
resource "hcloud_firewall" "k3s" {
name = var.cluster_name
dynamic "rule" {
for_each = concat(local.base_firewall_rules, var.extra_firewall_rules)
content {
direction = rule.value.direction
protocol = rule.value.protocol
port = lookup(rule.value, "port", null)
destination_ips = lookup(rule.value, "destination_ips", [])
source_ips = lookup(rule.value, "source_ips", [])
}
}
}
resource "hcloud_placement_group" "control_plane" {
count = ceil(local.control_plane_count / 10)
name = "${var.cluster_name}-control-plane-${count.index + 1}"
type = "spread"
}
resource "hcloud_placement_group" "agent" {
count = ceil(local.agent_count / 10)
name = "${var.cluster_name}-agent-${count.index + 1}"
type = "spread"
}
data "hcloud_load_balancer" "traefik" {
count = local.is_single_node_cluster ? 0 : var.traefik_enabled == false ? 0 : 1
name = "${var.cluster_name}-traefik"
depends_on = [null_resource.kustomization]
}
resource "null_resource" "destroy_traefik_loadbalancer" {
# this only gets triggered before total destruction of the cluster, but when the necessary elements to run the commands are still available
triggers = {
kustomization_id = null_resource.kustomization.id
}
# Important when issuing terraform destroy, otherwise the LB will not let the network get deleted
provisioner "local-exec" {
when = destroy
command = <<-EOT
kubectl -n kube-system delete service traefik --kubeconfig ${path.module}/kubeconfig.yaml
EOT
on_failure = continue
}
depends_on = [
local_sensitive_file.kubeconfig,
null_resource.control_planes[0],
hcloud_network_subnet.control_plane,
hcloud_network_subnet.agent,
hcloud_placement_group.control_plane,
hcloud_placement_group.agent,
hcloud_network.k3s,
hcloud_firewall.k3s,
hcloud_ssh_key.k3s
]
}