Adds AWS Instances to Output
This commit is contained in:
@@ -1,11 +1,11 @@
|
||||
terraform {
|
||||
required_providers {
|
||||
aws = {
|
||||
source = "hashicorp/aws"
|
||||
version = "5.17.0"
|
||||
configuration_aliases = [ aws.main ]
|
||||
required_providers {
|
||||
aws = {
|
||||
source = "hashicorp/aws"
|
||||
version = "5.17.0"
|
||||
configuration_aliases = [ aws.main ]
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@@ -66,9 +66,8 @@ resource "aws_instance" "main" {
|
||||
|
||||
key_name = aws_key_pair.ssh_key_main.key_name
|
||||
|
||||
user_data = templatefile("${path.module}/scripts/setup_main.sh", {
|
||||
user_data = templatefile("${path.module}/scripts/setup_server.sh", {
|
||||
extra_key = aws_key_pair.ssh_key_ci_cd.public_key
|
||||
k3s_token = var.k3s_token
|
||||
})
|
||||
|
||||
tags = {
|
||||
@@ -84,10 +83,8 @@ resource "aws_instance" "worker" {
|
||||
|
||||
key_name = aws_key_pair.ssh_key_main.key_name
|
||||
|
||||
user_data = templatefile("${path.module}/scripts/setup_worker.sh", {
|
||||
user_data = templatefile("${path.module}/scripts/setup_server.sh", {
|
||||
extra_key = aws_key_pair.ssh_key_ci_cd.public_key
|
||||
k3s_token = var.k3s_token
|
||||
k3s_cluster_ip = var.project_domain
|
||||
})
|
||||
|
||||
tags = {
|
||||
@@ -101,3 +98,11 @@ resource "aws_instance" "worker" {
|
||||
output "pool_master_public_ip" {
|
||||
value = aws_instance.main.public_ip
|
||||
}
|
||||
|
||||
output "pool_master_instance" {
|
||||
value = aws_instance.main
|
||||
}
|
||||
|
||||
output "pool_worker_instances" {
|
||||
value = aws_instance.worker
|
||||
}
|
||||
@@ -1,12 +0,0 @@
|
||||
#!/bin/bash -xe
|
||||
|
||||
|
||||
echo -e "\n${extra_key}" >> /home/ubuntu/.ssh/authorized_keys;
|
||||
|
||||
echo -e "export TERM='xterm-256color'" >> /home/ubuntu/.profile;
|
||||
|
||||
su ubuntu -i << EOF
|
||||
# curl -sfL https://get.k3s.io | \
|
||||
# K3S_TOKEN="${k3s_token}" sh -'
|
||||
echo "HERE" >> /home/ubuntu/test.txt
|
||||
EOF
|
||||
8
instances/scripts/setup_server.sh
Normal file
8
instances/scripts/setup_server.sh
Normal file
@@ -0,0 +1,8 @@
|
||||
#!/bin/bash
|
||||
|
||||
|
||||
echo -e "\n${extra_key}" >> /home/ubuntu/.ssh/authorized_keys;
|
||||
|
||||
echo "export TERM='xterm-256color';" > /home/ubuntu/.profile;
|
||||
|
||||
exit 0;
|
||||
@@ -1,14 +0,0 @@
|
||||
#!/bin/bash -xe
|
||||
|
||||
|
||||
echo -e "\n${extra_key}" >> /home/ubuntu/.ssh/authorized_keys;
|
||||
|
||||
echo "export TERM='xterm-256color'" > /home/ubuntu/.profile;
|
||||
|
||||
su ubuntu -i << EOF
|
||||
# curl -sfL https://get.k3s.io | \
|
||||
# INSTALL_K3S_EXEC="agent" \
|
||||
# K3S_TOKEN="${k3s_token}" \
|
||||
# sh -s - --server ${k3s_cluster_ip}
|
||||
echo "HERE" >> /home/ubuntu/test.txt
|
||||
EOF
|
||||
38
k3s/k3s.tf
38
k3s/k3s.tf
@@ -1,38 +0,0 @@
|
||||
module "k3s" {
|
||||
source = "xunleii/k3s/module"
|
||||
version = "3.3.0"
|
||||
k3s_version = "v1.21.4+k3s1"
|
||||
|
||||
cluster_domain = "civo_k3s"
|
||||
|
||||
drain_timeout = "60s"
|
||||
managed_fields = ["label"]
|
||||
generate_ca_certificates = true
|
||||
|
||||
global_flags = [for instance in civo_instance.node_instances : "--tls-san ${instance.public_ip}"]
|
||||
|
||||
servers = {
|
||||
# The node name will be automatically provided by
|
||||
# the module using the field name... any usage of
|
||||
# --node-name in additional_flags will be ignored
|
||||
|
||||
for instance in civo_instance.node_instances :
|
||||
instance.hostname => {
|
||||
ip = instance.private_ip
|
||||
connection = {
|
||||
timeout = "60s"
|
||||
type = "ssh"
|
||||
host = instance.public_ip
|
||||
password = instance.initial_password
|
||||
user = "root"
|
||||
}
|
||||
|
||||
labels = { "node.kubernetes.io/type" = "master" }
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
output "kube_config" {
|
||||
value = module.k3s.kube_config
|
||||
sensitive = true
|
||||
}
|
||||
Reference in New Issue
Block a user