mirror of
https://github.com/puppetmaster/typhoon.git
synced 2024-12-24 18:19:33 +01:00
df3f40bcce
* Set Kubelet cgroup driver to systemd when Flatcar Linux edge is chosen Note: Typhoon module status assumes use of the stable variant of an OS channel/stream. Its possible to use earlier variants and those are sometimes tested or developed against, but stable is the recommendation
179 lines
5.7 KiB
HCL
179 lines
5.7 KiB
HCL
# Discrete DNS records for each controller's private IPv4 for etcd usage
|
|
resource "azurerm_dns_a_record" "etcds" {
|
|
count = var.controller_count
|
|
resource_group_name = var.dns_zone_group
|
|
|
|
# DNS Zone name where record should be created
|
|
zone_name = var.dns_zone
|
|
|
|
# DNS record
|
|
name = format("%s-etcd%d", var.cluster_name, count.index)
|
|
ttl = 300
|
|
|
|
# private IPv4 address for etcd
|
|
records = [azurerm_network_interface.controllers.*.private_ip_address[count.index]]
|
|
}
|
|
|
|
locals {
|
|
# Container Linux derivative
|
|
# coreos-stable -> Container Linux Stable
|
|
# flatcar-stable -> Flatcar Linux Stable
|
|
flavor = split("-", var.os_image)[0]
|
|
channel = split("-", var.os_image)[1]
|
|
}
|
|
|
|
# Controller availability set to spread controllers
|
|
resource "azurerm_availability_set" "controllers" {
|
|
resource_group_name = azurerm_resource_group.cluster.name
|
|
|
|
name = "${var.cluster_name}-controllers"
|
|
location = var.region
|
|
platform_fault_domain_count = 2
|
|
platform_update_domain_count = 4
|
|
managed = true
|
|
}
|
|
|
|
# Controller instances
|
|
resource "azurerm_linux_virtual_machine" "controllers" {
|
|
count = var.controller_count
|
|
resource_group_name = azurerm_resource_group.cluster.name
|
|
|
|
name = "${var.cluster_name}-controller-${count.index}"
|
|
location = var.region
|
|
availability_set_id = azurerm_availability_set.controllers.id
|
|
|
|
size = var.controller_type
|
|
custom_data = base64encode(data.ct_config.controller-ignitions.*.rendered[count.index])
|
|
|
|
# storage
|
|
os_disk {
|
|
name = "${var.cluster_name}-controller-${count.index}"
|
|
caching = "None"
|
|
disk_size_gb = var.disk_size
|
|
storage_account_type = "Premium_LRS"
|
|
}
|
|
|
|
# CoreOS Container Linux or Flatcar Container Linux
|
|
source_image_reference {
|
|
publisher = local.flavor == "flatcar" ? "Kinvolk" : "CoreOS"
|
|
offer = local.flavor == "flatcar" ? "flatcar-container-linux-free" : "CoreOS"
|
|
sku = local.channel
|
|
version = "latest"
|
|
}
|
|
|
|
# Gross hack for Flatcar Linux
|
|
dynamic "plan" {
|
|
for_each = local.flavor == "flatcar" ? [1] : []
|
|
|
|
content {
|
|
name = local.channel
|
|
publisher = "kinvolk"
|
|
product = "flatcar-container-linux-free"
|
|
}
|
|
}
|
|
|
|
# network
|
|
network_interface_ids = [
|
|
azurerm_network_interface.controllers.*.id[count.index]
|
|
]
|
|
|
|
# Azure requires setting admin_ssh_key, though Ignition custom_data handles it too
|
|
admin_username = "core"
|
|
admin_ssh_key {
|
|
username = "core"
|
|
public_key = var.ssh_authorized_key
|
|
}
|
|
|
|
lifecycle {
|
|
ignore_changes = [
|
|
os_disk,
|
|
custom_data,
|
|
]
|
|
}
|
|
}
|
|
|
|
# Controller public IPv4 addresses
|
|
resource "azurerm_public_ip" "controllers" {
|
|
count = var.controller_count
|
|
resource_group_name = azurerm_resource_group.cluster.name
|
|
|
|
name = "${var.cluster_name}-controller-${count.index}"
|
|
location = azurerm_resource_group.cluster.location
|
|
sku = "Standard"
|
|
allocation_method = "Static"
|
|
}
|
|
|
|
# Controller NICs with public and private IPv4
|
|
resource "azurerm_network_interface" "controllers" {
|
|
count = var.controller_count
|
|
resource_group_name = azurerm_resource_group.cluster.name
|
|
|
|
name = "${var.cluster_name}-controller-${count.index}"
|
|
location = azurerm_resource_group.cluster.location
|
|
|
|
ip_configuration {
|
|
name = "ip0"
|
|
subnet_id = azurerm_subnet.controller.id
|
|
private_ip_address_allocation = "Dynamic"
|
|
# instance public IPv4
|
|
public_ip_address_id = azurerm_public_ip.controllers.*.id[count.index]
|
|
}
|
|
}
|
|
|
|
# Associate controller network interface with controller security group
|
|
resource "azurerm_network_interface_security_group_association" "controllers" {
|
|
count = var.controller_count
|
|
|
|
network_interface_id = azurerm_network_interface.controllers[count.index].id
|
|
network_security_group_id = azurerm_network_security_group.controller.id
|
|
}
|
|
|
|
# Associate controller network interface with controller backend address pool
|
|
resource "azurerm_network_interface_backend_address_pool_association" "controllers" {
|
|
count = var.controller_count
|
|
|
|
network_interface_id = azurerm_network_interface.controllers[count.index].id
|
|
ip_configuration_name = "ip0"
|
|
backend_address_pool_id = azurerm_lb_backend_address_pool.controller.id
|
|
}
|
|
|
|
# Controller Ignition configs
|
|
data "ct_config" "controller-ignitions" {
|
|
count = var.controller_count
|
|
content = data.template_file.controller-configs.*.rendered[count.index]
|
|
strict = true
|
|
snippets = var.controller_snippets
|
|
}
|
|
|
|
# Controller Container Linux configs
|
|
data "template_file" "controller-configs" {
|
|
count = var.controller_count
|
|
|
|
template = file("${path.module}/cl/controller.yaml")
|
|
|
|
vars = {
|
|
# Cannot use cyclic dependencies on controllers or their DNS records
|
|
etcd_name = "etcd${count.index}"
|
|
etcd_domain = "${var.cluster_name}-etcd${count.index}.${var.dns_zone}"
|
|
# etcd0=https://cluster-etcd0.example.com,etcd1=https://cluster-etcd1.example.com,...
|
|
etcd_initial_cluster = join(",", data.template_file.etcds.*.rendered)
|
|
cgroup_driver = local.flavor == "flatcar" && local.channel == "edge" ? "systemd" : "cgroupfs"
|
|
kubeconfig = indent(10, module.bootstrap.kubeconfig-kubelet)
|
|
ssh_authorized_key = var.ssh_authorized_key
|
|
cluster_dns_service_ip = cidrhost(var.service_cidr, 10)
|
|
cluster_domain_suffix = var.cluster_domain_suffix
|
|
}
|
|
}
|
|
|
|
data "template_file" "etcds" {
|
|
count = var.controller_count
|
|
template = "etcd$${index}=https://$${cluster_name}-etcd$${index}.$${dns_zone}:2380"
|
|
|
|
vars = {
|
|
index = count.index
|
|
cluster_name = var.cluster_name
|
|
dns_zone = var.dns_zone
|
|
}
|
|
}
|
|
|