Merge pull request #9 from HideyoshiNakazone/implements-k3s-cluster
Implements k3s cluster
This commit is contained in:
@@ -1,11 +1,23 @@
|
|||||||
terraform {
|
terraform {
|
||||||
required_providers {
|
required_providers {
|
||||||
aws = {
|
aws = {
|
||||||
source = "hashicorp/aws"
|
source = "hashicorp/aws"
|
||||||
version = "5.17.0"
|
version = "5.17.0"
|
||||||
configuration_aliases = [ aws.main ]
|
configuration_aliases = [ aws.main ]
|
||||||
|
}
|
||||||
|
tls = {
|
||||||
|
source = "hashicorp/tls"
|
||||||
|
version = "3.1.0"
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
# TERRAFORM SSH KEYS
|
||||||
|
|
||||||
|
resource "tls_private_key" "terraform_ssh_key" {
|
||||||
|
algorithm = "RSA"
|
||||||
|
rsa_bits = 4096
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
@@ -29,25 +41,12 @@ locals {
|
|||||||
6443,
|
6443,
|
||||||
10250
|
10250
|
||||||
]
|
]
|
||||||
ports_out = [
|
|
||||||
0,
|
|
||||||
]
|
|
||||||
}
|
}
|
||||||
|
|
||||||
resource "aws_security_group" "project_pool" {
|
resource "aws_security_group" "project_pool" {
|
||||||
name = "${var.project_name}_pool_security_group"
|
name = "${var.project_name}_pool_security_group"
|
||||||
description = "Security group for project pool"
|
description = "Security group for project pool"
|
||||||
|
|
||||||
dynamic "egress" {
|
|
||||||
for_each = toset(local.ports_out)
|
|
||||||
content {
|
|
||||||
from_port = egress.value
|
|
||||||
to_port = egress.value
|
|
||||||
protocol = "tcp"
|
|
||||||
cidr_blocks = ["0.0.0.0/0"]
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
dynamic "ingress" {
|
dynamic "ingress" {
|
||||||
for_each = toset(local.ports_in)
|
for_each = toset(local.ports_in)
|
||||||
content {
|
content {
|
||||||
@@ -57,6 +56,15 @@ resource "aws_security_group" "project_pool" {
|
|||||||
cidr_blocks = ["0.0.0.0/0"]
|
cidr_blocks = ["0.0.0.0/0"]
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
egress {
|
||||||
|
from_port = 0
|
||||||
|
to_port = 0
|
||||||
|
protocol = "-1"
|
||||||
|
cidr_blocks = ["0.0.0.0/0"]
|
||||||
|
ipv6_cidr_blocks = ["::/0"]
|
||||||
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
resource "aws_instance" "main" {
|
resource "aws_instance" "main" {
|
||||||
@@ -66,11 +74,27 @@ resource "aws_instance" "main" {
|
|||||||
|
|
||||||
key_name = aws_key_pair.ssh_key_main.key_name
|
key_name = aws_key_pair.ssh_key_main.key_name
|
||||||
|
|
||||||
user_data = templatefile("${path.module}/scripts/setup_main.sh", {
|
user_data = templatefile("${path.module}/scripts/setup_server.sh", {
|
||||||
extra_key = aws_key_pair.ssh_key_ci_cd.public_key
|
extra_key = aws_key_pair.ssh_key_ci_cd.public_key
|
||||||
k3s_token = var.k3s_token
|
terraform_key = tls_private_key.terraform_ssh_key.public_key_openssh
|
||||||
})
|
})
|
||||||
|
|
||||||
|
provisioner "remote-exec" {
|
||||||
|
connection {
|
||||||
|
type = "ssh"
|
||||||
|
user = "ubuntu"
|
||||||
|
agent = false
|
||||||
|
private_key = tls_private_key.terraform_ssh_key.private_key_pem
|
||||||
|
host = self.public_ip
|
||||||
|
}
|
||||||
|
|
||||||
|
inline = [
|
||||||
|
"echo 'curl -sfL https://get.k3s.io | K3S_TOKEN=\"${var.k3s_token}\" sh -' >> /home/ubuntu/setup.sh",
|
||||||
|
"chmod +x /home/ubuntu/setup.sh",
|
||||||
|
"exec /home/ubuntu/setup.sh | tee logs.txt"
|
||||||
|
]
|
||||||
|
}
|
||||||
|
|
||||||
tags = {
|
tags = {
|
||||||
Name = "${var.project_name}-main"
|
Name = "${var.project_name}-main"
|
||||||
}
|
}
|
||||||
@@ -84,12 +108,27 @@ resource "aws_instance" "worker" {
|
|||||||
|
|
||||||
key_name = aws_key_pair.ssh_key_main.key_name
|
key_name = aws_key_pair.ssh_key_main.key_name
|
||||||
|
|
||||||
user_data = templatefile("${path.module}/scripts/setup_worker.sh", {
|
user_data = templatefile("${path.module}/scripts/setup_server.sh", {
|
||||||
extra_key = aws_key_pair.ssh_key_ci_cd.public_key
|
extra_key = aws_key_pair.ssh_key_ci_cd.public_key
|
||||||
k3s_token = var.k3s_token
|
terraform_key = tls_private_key.terraform_ssh_key.public_key_openssh
|
||||||
k3s_cluster_ip = var.project_domain
|
|
||||||
})
|
})
|
||||||
|
|
||||||
|
provisioner "remote-exec" {
|
||||||
|
connection {
|
||||||
|
type = "ssh"
|
||||||
|
user = "ubuntu"
|
||||||
|
agent = false
|
||||||
|
private_key = tls_private_key.terraform_ssh_key.private_key_pem
|
||||||
|
host = self.public_ip
|
||||||
|
}
|
||||||
|
|
||||||
|
inline = [
|
||||||
|
"echo 'curl -sfL https://get.k3s.io | INSTALL_K3S_EXEC=\"agent\" K3S_TOKEN=\"${var.k3s_token}\" sh -s - --server ${var.project_domain}:6443' >> /home/ubuntu/setup.sh",
|
||||||
|
"chmod +x /home/ubuntu/setup.sh",
|
||||||
|
"exec /home/ubuntu/setup.sh | tee logs.txt"
|
||||||
|
]
|
||||||
|
}
|
||||||
|
|
||||||
tags = {
|
tags = {
|
||||||
Name = "${var.project_name}-worker-${count.index+1}"
|
Name = "${var.project_name}-worker-${count.index+1}"
|
||||||
}
|
}
|
||||||
@@ -101,3 +140,11 @@ resource "aws_instance" "worker" {
|
|||||||
output "pool_master_public_ip" {
|
output "pool_master_public_ip" {
|
||||||
value = aws_instance.main.public_ip
|
value = aws_instance.main.public_ip
|
||||||
}
|
}
|
||||||
|
|
||||||
|
output "pool_master_instance" {
|
||||||
|
value = aws_instance.main
|
||||||
|
}
|
||||||
|
|
||||||
|
output "pool_worker_instances" {
|
||||||
|
value = aws_instance.worker
|
||||||
|
}
|
||||||
@@ -1,12 +0,0 @@
|
|||||||
#!/bin/bash -xe
|
|
||||||
|
|
||||||
|
|
||||||
echo -e "\n${extra_key}" >> /home/ubuntu/.ssh/authorized_keys;
|
|
||||||
|
|
||||||
echo -e "export TERM='xterm-256color'" >> /home/ubuntu/.profile;
|
|
||||||
|
|
||||||
su ubuntu -i << EOF
|
|
||||||
# curl -sfL https://get.k3s.io | \
|
|
||||||
# K3S_TOKEN="${k3s_token}" sh -'
|
|
||||||
echo "HERE" >> /home/ubuntu/test.txt
|
|
||||||
EOF
|
|
||||||
9
instances/scripts/setup_server.sh
Normal file
9
instances/scripts/setup_server.sh
Normal file
@@ -0,0 +1,9 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
|
||||||
|
|
||||||
|
echo -e "${extra_key}" >> /home/ubuntu/.ssh/authorized_keys;
|
||||||
|
echo -e "${terraform_key}" >> /home/ubuntu/.ssh/authorized_keys;
|
||||||
|
|
||||||
|
echo "export TERM='xterm-256color';" > /home/ubuntu/.profile;
|
||||||
|
|
||||||
|
exit 0;
|
||||||
@@ -1,14 +0,0 @@
|
|||||||
#!/bin/bash -xe
|
|
||||||
|
|
||||||
|
|
||||||
echo -e "\n${extra_key}" >> /home/ubuntu/.ssh/authorized_keys;
|
|
||||||
|
|
||||||
echo "export TERM='xterm-256color'" > /home/ubuntu/.profile;
|
|
||||||
|
|
||||||
su ubuntu -i << EOF
|
|
||||||
# curl -sfL https://get.k3s.io | \
|
|
||||||
# INSTALL_K3S_EXEC="agent" \
|
|
||||||
# K3S_TOKEN="${k3s_token}" \
|
|
||||||
# sh -s - --server ${k3s_cluster_ip}
|
|
||||||
echo "HERE" >> /home/ubuntu/test.txt
|
|
||||||
EOF
|
|
||||||
38
k3s/k3s.tf
38
k3s/k3s.tf
@@ -1,38 +0,0 @@
|
|||||||
module "k3s" {
|
|
||||||
source = "xunleii/k3s/module"
|
|
||||||
version = "3.3.0"
|
|
||||||
k3s_version = "v1.21.4+k3s1"
|
|
||||||
|
|
||||||
cluster_domain = "civo_k3s"
|
|
||||||
|
|
||||||
drain_timeout = "60s"
|
|
||||||
managed_fields = ["label"]
|
|
||||||
generate_ca_certificates = true
|
|
||||||
|
|
||||||
global_flags = [for instance in civo_instance.node_instances : "--tls-san ${instance.public_ip}"]
|
|
||||||
|
|
||||||
servers = {
|
|
||||||
# The node name will be automatically provided by
|
|
||||||
# the module using the field name... any usage of
|
|
||||||
# --node-name in additional_flags will be ignored
|
|
||||||
|
|
||||||
for instance in civo_instance.node_instances :
|
|
||||||
instance.hostname => {
|
|
||||||
ip = instance.private_ip
|
|
||||||
connection = {
|
|
||||||
timeout = "60s"
|
|
||||||
type = "ssh"
|
|
||||||
host = instance.public_ip
|
|
||||||
password = instance.initial_password
|
|
||||||
user = "root"
|
|
||||||
}
|
|
||||||
|
|
||||||
labels = { "node.kubernetes.io/type" = "master" }
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
output "kube_config" {
|
|
||||||
value = module.k3s.kube_config
|
|
||||||
sensitive = true
|
|
||||||
}
|
|
||||||
Reference in New Issue
Block a user