forked from catalin/fukuops
feat: remove ansible/
This commit is contained in:
parent
d84e66353b
commit
23ca6eb433
29 changed files with 21 additions and 568 deletions
|
|
@ -1,3 +0,0 @@
|
||||||
[defaults]
|
|
||||||
inventory = inventory
|
|
||||||
host_key_checking = False
|
|
||||||
|
|
@ -1,26 +0,0 @@
|
||||||
[nextclouds]
|
|
||||||
cloud.fuku
|
|
||||||
|
|
||||||
[nextclouds:vars]
|
|
||||||
ansible_user=root
|
|
||||||
|
|
||||||
[k3s_masters]
|
|
||||||
master1.ramiel.fuku
|
|
||||||
master2.ramiel.fuku
|
|
||||||
master3.ramiel.fuku
|
|
||||||
|
|
||||||
[k3s_agents]
|
|
||||||
agent1.zeruel.fuku
|
|
||||||
sandalphon.fuku
|
|
||||||
|
|
||||||
[k3s_masters:vars]
|
|
||||||
ansible_user=ci
|
|
||||||
|
|
||||||
[k3s_agents:vars]
|
|
||||||
ansible_user=ci
|
|
||||||
|
|
||||||
[giteas]
|
|
||||||
gitea.fuku
|
|
||||||
|
|
||||||
[giteas:vars]
|
|
||||||
ansible_user=root
|
|
||||||
|
|
@ -1,9 +0,0 @@
|
||||||
---
|
|
||||||
- name: Apply base configuration
|
|
||||||
hosts:
|
|
||||||
- k3s_agents
|
|
||||||
- k3s_masters
|
|
||||||
|
|
||||||
roles:
|
|
||||||
- role: ../roles/base
|
|
||||||
become: true
|
|
||||||
|
|
@ -1,13 +0,0 @@
|
||||||
---
|
|
||||||
- name: Install k3s nodes
|
|
||||||
hosts:
|
|
||||||
- k3s_masters
|
|
||||||
- k3s_agents
|
|
||||||
|
|
||||||
roles:
|
|
||||||
- role: ../roles/k3s
|
|
||||||
become: true
|
|
||||||
vars:
|
|
||||||
first_master_hostname: "{{ groups['k3s_masters'][0] }}"
|
|
||||||
is_first_master: "{{ inventory_hostname in groups['k3s_masters'][0] }}"
|
|
||||||
short_hostname: "{{ inventory_hostname.split('.')[0] }}"
|
|
||||||
|
|
@ -1,36 +0,0 @@
|
||||||
---
|
|
||||||
- name: Set same timezone on every Server
|
|
||||||
community.general.timezone:
|
|
||||||
name: '{{ system_timezone }}'
|
|
||||||
when: (system_timezone is defined) and (system_timezone != "Europe/Madrid")
|
|
||||||
|
|
||||||
- name: Enable IPv4 forwarding
|
|
||||||
ansible.posix.sysctl:
|
|
||||||
name: net.ipv4.ip_forward
|
|
||||||
value: '1'
|
|
||||||
state: present
|
|
||||||
reload: true
|
|
||||||
|
|
||||||
- name: Enable IPv6 forwarding
|
|
||||||
ansible.posix.sysctl:
|
|
||||||
name: net.ipv6.conf.all.forwarding
|
|
||||||
value: '1'
|
|
||||||
state: present
|
|
||||||
reload: true
|
|
||||||
|
|
||||||
- name: Enable IPv6 router advertisements
|
|
||||||
ansible.posix.sysctl:
|
|
||||||
name: net.ipv6.conf.all.accept_ra
|
|
||||||
value: '2'
|
|
||||||
state: present
|
|
||||||
reload: true
|
|
||||||
|
|
||||||
- import_tasks: packages.yml
|
|
||||||
name: Install base packages
|
|
||||||
tags:
|
|
||||||
- packages
|
|
||||||
|
|
||||||
- import_tasks: mounts.yml
|
|
||||||
name: Mount NFS shares
|
|
||||||
tags:
|
|
||||||
- nfs
|
|
||||||
|
|
@ -1,19 +0,0 @@
|
||||||
---
|
|
||||||
- name: Create mountpoint directory
|
|
||||||
file:
|
|
||||||
path: /nfs/nas1
|
|
||||||
state: directory
|
|
||||||
owner: 10000
|
|
||||||
group: 10000
|
|
||||||
|
|
||||||
- name: Mount nas1 share
|
|
||||||
mount:
|
|
||||||
fstype: nfs
|
|
||||||
src: zeruel.fuku:/mnt/pool1/nas1
|
|
||||||
path: /nfs/nas1
|
|
||||||
state: mounted
|
|
||||||
fstab: /etc/fstab
|
|
||||||
opts: _netdev,nofail,tcp,bg,retrans=2,timeo=150,rsize=32768,wsize=32768,noresvport
|
|
||||||
backup: true
|
|
||||||
become: true
|
|
||||||
become_user: root
|
|
||||||
|
|
@ -1,18 +0,0 @@
|
||||||
---
|
|
||||||
- name: Install base packages
|
|
||||||
apt:
|
|
||||||
name: '{{ item }}'
|
|
||||||
state: present
|
|
||||||
update_cache: true
|
|
||||||
loop:
|
|
||||||
- qemu-guest-agent
|
|
||||||
- git
|
|
||||||
- tmux
|
|
||||||
- vim
|
|
||||||
- curl
|
|
||||||
- nfs-common
|
|
||||||
|
|
||||||
- name: Update all packages
|
|
||||||
apt:
|
|
||||||
upgrade: dist
|
|
||||||
update_cache: true
|
|
||||||
|
|
@ -1,17 +0,0 @@
|
||||||
---
|
|
||||||
- name: Create rancher folder
|
|
||||||
file:
|
|
||||||
state: directory
|
|
||||||
path: /etc/rancher/k3s
|
|
||||||
owner: root
|
|
||||||
group: root
|
|
||||||
mode: 755
|
|
||||||
|
|
||||||
- name: Copy k3s config file
|
|
||||||
template:
|
|
||||||
src: agent.config.yaml.j2
|
|
||||||
dest: /etc/rancher/k3s/config.yaml
|
|
||||||
mode: 600
|
|
||||||
|
|
||||||
- name: Install k3s agent
|
|
||||||
shell: bash /tmp/k3s.install.sh agent
|
|
||||||
|
|
@ -1,19 +0,0 @@
|
||||||
---
|
|
||||||
- name: Create .kube directory
|
|
||||||
become: true
|
|
||||||
file:
|
|
||||||
path: /home/ci/.kube
|
|
||||||
state: directory
|
|
||||||
mode: '0755'
|
|
||||||
owner: ci
|
|
||||||
group: ci
|
|
||||||
|
|
||||||
- name: Copy kubeconfig
|
|
||||||
copy:
|
|
||||||
remote_src: true
|
|
||||||
src: /etc/rancher/k3s/k3s.yaml
|
|
||||||
dest: /home/ci/.kube/config
|
|
||||||
mode: 0644
|
|
||||||
owner: ci
|
|
||||||
group: ci
|
|
||||||
become: true
|
|
||||||
|
|
@ -1,5 +0,0 @@
|
||||||
---
|
|
||||||
- name: Download k3s script
|
|
||||||
get_url:
|
|
||||||
url: https://get.k3s.io
|
|
||||||
dest: /tmp/k3s.install.sh
|
|
||||||
|
|
@ -1,15 +0,0 @@
|
||||||
---
|
|
||||||
- import_tasks: download.yml
|
|
||||||
name: Download install script
|
|
||||||
|
|
||||||
- import_tasks: master.yml
|
|
||||||
name: Install master node
|
|
||||||
when: inventory_hostname in groups["k3s_masters"]
|
|
||||||
|
|
||||||
- import_tasks: agent.yml
|
|
||||||
name: Install agent node
|
|
||||||
when: inventory_hostname in groups["k3s_agents"]
|
|
||||||
|
|
||||||
- import_tasks: copy-kubeconfig.yml
|
|
||||||
name: Copy kubeconfig
|
|
||||||
when: inventory_hostname in groups["k3s_masters"] and is_first_master
|
|
||||||
|
|
@ -1,19 +0,0 @@
|
||||||
---
|
|
||||||
- name: Create rancher folder
|
|
||||||
file:
|
|
||||||
state: directory
|
|
||||||
path: /etc/rancher/k3s
|
|
||||||
owner: root
|
|
||||||
group: root
|
|
||||||
mode: 755
|
|
||||||
|
|
||||||
- name: Copy k3s config file
|
|
||||||
template:
|
|
||||||
src: master.config.yaml.j2
|
|
||||||
dest: /etc/rancher/k3s/config.yaml
|
|
||||||
mode: 600
|
|
||||||
vars:
|
|
||||||
etcd_snapshot_dir: /nfs/nas1/backups/{{ short_hostname }}
|
|
||||||
|
|
||||||
- name: Install k3s master
|
|
||||||
command: bash /tmp/k3s.install.sh
|
|
||||||
|
|
@ -1,2 +0,0 @@
|
||||||
token: {{ cluster_token }}
|
|
||||||
server: https://{{ tls_san }}:6443
|
|
||||||
|
|
@ -1,12 +0,0 @@
|
||||||
tls-san:
|
|
||||||
- {{ inventory_hostname }}
|
|
||||||
- {{ tls_san }}
|
|
||||||
node-label:
|
|
||||||
- name={{ inventory_hostname }}
|
|
||||||
token: "{{ cluster_token }}"
|
|
||||||
etcd-snapshot-dir: {{ etcd_snapshot_dir }}
|
|
||||||
{% if is_first_master %}
|
|
||||||
cluster-init: "{{ is_first_master }}"
|
|
||||||
{% else %}
|
|
||||||
server: https://{{ first_master_hostname }}:6443
|
|
||||||
{% endif %}
|
|
||||||
|
|
@ -1,4 +0,0 @@
|
||||||
---
|
|
||||||
k3s_version: v1.27.4+k3s1
|
|
||||||
tls_san: "{{ lookup('env', 'ANSIBLE_TLS_SAN') | mandatory }}"
|
|
||||||
cluster_token: "{{ lookup('env', 'ANSIBLE_CLUSTER_TOKEN') | mandatory }}"
|
|
||||||
|
|
@ -1,2 +0,0 @@
|
||||||
ANSIBLE_K3S_CLUSTER_TOKEN=
|
|
||||||
ANSIBLE_K3S_TLS_SAN=
|
|
||||||
|
|
@ -1,27 +0,0 @@
|
||||||
---
|
|
||||||
- name: Install promtail
|
|
||||||
hosts:
|
|
||||||
- nextclouds
|
|
||||||
|
|
||||||
roles:
|
|
||||||
- role: patrickjahns.promtail
|
|
||||||
vars:
|
|
||||||
promtail_version: 2.9.4
|
|
||||||
promtail_config_clients:
|
|
||||||
- url: https://loki.fuku/loki/api/v1/push
|
|
||||||
basic_auth:
|
|
||||||
username: cloud
|
|
||||||
password: "{{ lookup('env', 'NEXTCLOUD_PROMTAIL_PASSWORD') | mandatory }}"
|
|
||||||
tls_config:
|
|
||||||
insecure_skip_verify: true
|
|
||||||
promtail_config_scrape_configs:
|
|
||||||
- job_name: system
|
|
||||||
static_configs:
|
|
||||||
- targets:
|
|
||||||
- localhost
|
|
||||||
labels:
|
|
||||||
nextcloud: cloud.fukurokuju.dev
|
|
||||||
__path__: /mnt/share/data/cloud/data/{nextcloud,audit}.log
|
|
||||||
promtail_config_limits_config:
|
|
||||||
readline_rate_enabled: true
|
|
||||||
readline_rate_drop: true
|
|
||||||
|
|
@ -1 +0,0 @@
|
||||||
NEXTCLOUD_PROMTAIL_PASSWORD=superdupersecure
|
|
||||||
|
|
@ -1,3 +0,0 @@
|
||||||
---
|
|
||||||
- name: patrickjahns.promtail
|
|
||||||
version: 1.31.0
|
|
||||||
|
|
@ -12,7 +12,7 @@ spec:
|
||||||
sources:
|
sources:
|
||||||
- chart: authentik
|
- chart: authentik
|
||||||
repoURL: https://charts.goauthentik.io/
|
repoURL: https://charts.goauthentik.io/
|
||||||
targetRevision: 2025.4.*
|
targetRevision: 2025.6.*
|
||||||
helm:
|
helm:
|
||||||
valuesObject:
|
valuesObject:
|
||||||
authentik:
|
authentik:
|
||||||
|
|
|
||||||
|
|
@ -1,35 +0,0 @@
|
||||||
#!/usr/bin/env bash
|
|
||||||
|
|
||||||
SCRIPT_DIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" &> /dev/null && pwd )"
|
|
||||||
|
|
||||||
cd "$SCRIPT_DIR/../docker/nextcloud" || exit
|
|
||||||
|
|
||||||
docker_exec() {
|
|
||||||
docker compose exec nextcloud "$@"
|
|
||||||
}
|
|
||||||
|
|
||||||
occ_exec() {
|
|
||||||
docker_exec sudo -E -u www-data php occ "$@"
|
|
||||||
}
|
|
||||||
|
|
||||||
case "$1" in
|
|
||||||
upgrade)
|
|
||||||
occ_exec upgrade
|
|
||||||
;;
|
|
||||||
htaccess)
|
|
||||||
occ_exec maintenance:update:htaccess
|
|
||||||
;;
|
|
||||||
indices)
|
|
||||||
occ_exec db:add-missing-indices
|
|
||||||
;;
|
|
||||||
occ)
|
|
||||||
occ_exec "$@"
|
|
||||||
;;
|
|
||||||
exec)
|
|
||||||
docker_exec "$@"
|
|
||||||
;;
|
|
||||||
*)
|
|
||||||
echo "Usage: $0 {upgrade|htaccess|indices|occ <custom occ command>|exec <custom command>}"
|
|
||||||
exit 1
|
|
||||||
;;
|
|
||||||
esac
|
|
||||||
|
|
@ -35,64 +35,16 @@ resource "adguard_rewrite" "argo_3" {
|
||||||
domain = "argo.fuku"
|
domain = "argo.fuku"
|
||||||
answer = "192.168.1.33"
|
answer = "192.168.1.33"
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
resource "adguard_rewrite" "loki_1" {
|
|
||||||
domain = "loki.fuku"
|
|
||||||
answer = "192.168.1.31"
|
|
||||||
}
|
|
||||||
|
|
||||||
resource "adguard_rewrite" "loki_2" {
|
|
||||||
domain = "loki.fuku"
|
|
||||||
answer = "192.168.1.32"
|
|
||||||
}
|
|
||||||
|
|
||||||
resource "adguard_rewrite" "loki_3" {
|
|
||||||
domain = "loki.fuku"
|
|
||||||
answer = "192.168.1.33"
|
|
||||||
}
|
|
||||||
|
|
||||||
resource "adguard_rewrite" "grafana_1" {
|
|
||||||
domain = "grafana.fuku"
|
|
||||||
answer = "192.168.1.31"
|
|
||||||
}
|
|
||||||
|
|
||||||
resource "adguard_rewrite" "grafana_2" {
|
|
||||||
domain = "grafana.fuku"
|
|
||||||
answer = "192.168.1.32"
|
|
||||||
}
|
|
||||||
|
|
||||||
resource "adguard_rewrite" "grafana_3" {
|
|
||||||
domain = "grafana.fuku"
|
|
||||||
answer = "192.168.1.33"
|
|
||||||
}
|
|
||||||
|
|
||||||
resource "adguard_rewrite" "feeds" {
|
resource "adguard_rewrite" "feeds" {
|
||||||
domain = "feeds.roboces.dev"
|
domain = "feeds.roboces.dev"
|
||||||
answer = "192.168.1.12"
|
answer = "192.168.1.12"
|
||||||
}
|
}
|
||||||
|
|
||||||
resource "adguard_rewrite" "feeds_local_1" {
|
|
||||||
domain = "feeds.fuku"
|
|
||||||
answer = "192.168.1.31"
|
|
||||||
}
|
|
||||||
|
|
||||||
resource "adguard_rewrite" "feeds_local_2" {
|
|
||||||
domain = "feeds.fuku"
|
|
||||||
answer = "192.168.1.32"
|
|
||||||
}
|
|
||||||
|
|
||||||
resource "adguard_rewrite" "feeds_local_3" {
|
|
||||||
domain = "feeds.fuku"
|
|
||||||
answer = "192.168.1.33"
|
|
||||||
}
|
|
||||||
|
|
||||||
resource "adguard_rewrite" "authentik" {
|
resource "adguard_rewrite" "authentik" {
|
||||||
domain = "auth.fukurokuju.dev"
|
domain = "auth.fukurokuju.dev"
|
||||||
answer = "192.168.1.12"
|
answer = "192.168.1.12"
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
resource "adguard_rewrite" "dd02" {
|
resource "adguard_rewrite" "dd02" {
|
||||||
domain = "dd02.fuku"
|
domain = "dd02.fuku"
|
||||||
answer = "192.168.1.19"
|
answer = "192.168.1.19"
|
||||||
|
|
@ -117,3 +69,23 @@ resource "adguard_rewrite" "elastic_3" {
|
||||||
domain = "elastic.fuku"
|
domain = "elastic.fuku"
|
||||||
answer = "192.168.1.33"
|
answer = "192.168.1.33"
|
||||||
}
|
}
|
||||||
|
|
||||||
|
resource "adguard_rewrite" "agent1" {
|
||||||
|
domain = "agent1.fuku"
|
||||||
|
answer = "192.168.1.34"
|
||||||
|
}
|
||||||
|
|
||||||
|
resource "adguard_rewrite" "master1" {
|
||||||
|
domain = "master1.ramiel.fuku"
|
||||||
|
answer = "192.168.1.31"
|
||||||
|
}
|
||||||
|
|
||||||
|
resource "adguard_rewrite" "master2" {
|
||||||
|
domain = "master2.ramiel.fuku"
|
||||||
|
answer = "192.168.1.32"
|
||||||
|
}
|
||||||
|
|
||||||
|
resource "adguard_rewrite" "master3" {
|
||||||
|
domain = "master3.ramiel.fuku"
|
||||||
|
answer = "192.168.1.33"
|
||||||
|
}
|
||||||
|
|
|
||||||
23
tofu/modules/proxmox-vm/.terraform.lock.hcl
generated
23
tofu/modules/proxmox-vm/.terraform.lock.hcl
generated
|
|
@ -1,23 +0,0 @@
|
||||||
# This file is maintained automatically by "tofu init".
|
|
||||||
# Manual edits may be lost in future updates.
|
|
||||||
|
|
||||||
provider "registry.opentofu.org/thegameprofi/proxmox" {
|
|
||||||
version = "2.10.0"
|
|
||||||
constraints = "2.10.0"
|
|
||||||
hashes = [
|
|
||||||
"h1:0wJlZKYgmnyXJ4AGKfML5m0/bhRUXty3vn4mSEYhFDM=",
|
|
||||||
"h1:2fmj6FkuQ8CyWdxJ83s87GJhRHiQFnzlSYSlA2q7pXg=",
|
|
||||||
"h1:5FQbfjTDMO4FCMXAQ5S+2CdVHw1gJoOicQRLVTCfrDI=",
|
|
||||||
"h1:60DcgYwyBroh3oUsdsZU/Z03LfEIidq2lTYKbhZDibU=",
|
|
||||||
"h1:6hrWJT18l4jfWrzO8GweRA1bIdzY8dDUkzB7+GW0Zgs=",
|
|
||||||
"h1:8bQZLEyLJqYnLnZC0zopbPSJVYXwQIb+G3nQ9Tc9bMU=",
|
|
||||||
"h1:DArex0up+F+sN3arAKanOSVc70b9kyTeTkuy6Abkor8=",
|
|
||||||
"h1:WUh8KlPtUdI+O7FJz6/d3BC+77tSBYjlNCxKjIGX7Kg=",
|
|
||||||
"h1:atdwdBFAPjyz7ICxoeMBrUn7tllqjJyIIClO5Mv3dlw=",
|
|
||||||
"h1:mnIwCGDAZ9O/UiQEXGWKXUeA6H1Tt/AkII/MjHw1SMg=",
|
|
||||||
"h1:sW+6tq318P9g0P+BwqEM74AHsJ6CmN+s6NgAJJxZk/Y=",
|
|
||||||
"h1:tWdlbcPuby+thPgC07FCBbIzzQAOn3XSvRibuHcdK04=",
|
|
||||||
"h1:yHr19pmBvjRCSbyWmNd3Y4iI7mCjeDgzC622q5kSSKE=",
|
|
||||||
"h1:zbDNS+UBlhewZUPt0s5ntUeAGOSTLj2jjHweCDKw6qo=",
|
|
||||||
]
|
|
||||||
}
|
|
||||||
|
|
@ -1,46 +0,0 @@
|
||||||
terraform {
|
|
||||||
required_version = ">= 1.6"
|
|
||||||
required_providers {
|
|
||||||
proxmox = {
|
|
||||||
source = "thegameprofi/proxmox"
|
|
||||||
version = "2.10.0"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
resource "proxmox_vm_qemu" "vm" {
|
|
||||||
name = var.vm_name
|
|
||||||
target_node = var.node_name
|
|
||||||
vmid = var.vm_id
|
|
||||||
clone = var.cloud_init_template
|
|
||||||
os_type = "cloudinit"
|
|
||||||
qemu_os = "other"
|
|
||||||
ipconfig0 = var.ipconfig0
|
|
||||||
cores = var.core_count
|
|
||||||
sockets = 1
|
|
||||||
pool = "k3s"
|
|
||||||
memory = var.memory
|
|
||||||
ciuser = var.ci_username
|
|
||||||
agent = 0
|
|
||||||
serial {
|
|
||||||
id = 0
|
|
||||||
type = "socket"
|
|
||||||
}
|
|
||||||
disk {
|
|
||||||
size = "50G"
|
|
||||||
storage = var.disk_storage_name
|
|
||||||
type = "scsi"
|
|
||||||
}
|
|
||||||
network {
|
|
||||||
bridge = var.network_bridge_name
|
|
||||||
firewall = false
|
|
||||||
link_down = false
|
|
||||||
model = "virtio"
|
|
||||||
mtu = 0
|
|
||||||
queues = 0
|
|
||||||
rate = 0
|
|
||||||
tag = -1
|
|
||||||
}
|
|
||||||
sshkeys = var.ssh_keys
|
|
||||||
nameserver = "192.168.1.7 192.168.1.3"
|
|
||||||
}
|
|
||||||
|
|
@ -1,66 +0,0 @@
|
||||||
variable "vm_name" {
|
|
||||||
description = "Name of the VM"
|
|
||||||
type = string
|
|
||||||
}
|
|
||||||
|
|
||||||
variable "node_name" {
|
|
||||||
description = "Name of the Proxmox node"
|
|
||||||
type = string
|
|
||||||
}
|
|
||||||
|
|
||||||
variable "vm_id" {
|
|
||||||
description = "ID of the new VM"
|
|
||||||
type = number
|
|
||||||
}
|
|
||||||
|
|
||||||
variable "cloud_init_template" {
|
|
||||||
description = "Cloud-init enabled template to be cloned"
|
|
||||||
type = string
|
|
||||||
default = "ci-debian12"
|
|
||||||
}
|
|
||||||
|
|
||||||
variable "ipconfig0" {
|
|
||||||
description = "Default inet configuration"
|
|
||||||
type = string
|
|
||||||
}
|
|
||||||
|
|
||||||
variable "ci_username" {
|
|
||||||
description = "Cloud-init username"
|
|
||||||
type = string
|
|
||||||
default = "ci"
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
variable "memory" {
|
|
||||||
description = "Available RAM, in MB"
|
|
||||||
type = number
|
|
||||||
default = 512
|
|
||||||
}
|
|
||||||
|
|
||||||
variable "core_count" {
|
|
||||||
description = "Available cores per socket"
|
|
||||||
type = number
|
|
||||||
default = 2
|
|
||||||
|
|
||||||
}
|
|
||||||
|
|
||||||
variable "network_bridge_name" {
|
|
||||||
description = "Network bridge name"
|
|
||||||
type = string
|
|
||||||
default = "vmbr0"
|
|
||||||
}
|
|
||||||
|
|
||||||
variable "ssh_keys" {
|
|
||||||
description = "SSH public keys to be provisioned"
|
|
||||||
type = string
|
|
||||||
default = <<EOF
|
|
||||||
ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIL++DKyzE0+zM6A4lpjWTNvntQnmEdA6FoOh6jIdsHMT catalin@jupiter
|
|
||||||
ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIGreLVacZyxq2EtgievpXgC/L7czKyJa/kWpgqDoPgnA phireh@GenPhireh
|
|
||||||
EOF
|
|
||||||
}
|
|
||||||
|
|
||||||
variable "disk_storage_name" {
|
|
||||||
description = "Disk storage name"
|
|
||||||
type = string
|
|
||||||
default = "storage"
|
|
||||||
}
|
|
||||||
45
tofu/proxmox/.terraform.lock.hcl
generated
45
tofu/proxmox/.terraform.lock.hcl
generated
|
|
@ -1,45 +0,0 @@
|
||||||
# This file is maintained automatically by "tofu init".
|
|
||||||
# Manual edits may be lost in future updates.
|
|
||||||
|
|
||||||
provider "registry.opentofu.org/telmate/proxmox" {
|
|
||||||
version = "2.9.14"
|
|
||||||
constraints = "2.9.14"
|
|
||||||
hashes = [
|
|
||||||
"h1:H/f+LbVyPOLslHLAYnGuMMRqWFZ65K6E3V+MCYgfAyk=",
|
|
||||||
"zh:0d049d33f705e5b814d30028770c084151218439424e99684ce31d7e26a720b5",
|
|
||||||
"zh:20b1c64ed56d81de95f3f37b82b45b4654c0de26670c0e87a474c5cce13cd015",
|
|
||||||
"zh:2946058abd1d8e50e475b9ec39781eb02576b40dbd80f4653fade4493a4514c6",
|
|
||||||
"zh:29e50a25c456f040ce072f23ac57b5b82ebd3b916ca5ae6688332b5ec62adc4a",
|
|
||||||
"zh:3612932306ce5f08db94868f526cbb8c56d0d3c6ebe1c11a83f92bbf94354296",
|
|
||||||
"zh:42d1699b0abebaac82ea5a19f4393541d8bb2741bde204a8ac1028cdc29d1b14",
|
|
||||||
"zh:5ffd5dc567262eb8aafdf2f6eac63f7f21361da9c5d75a3c36b479638a0001b0",
|
|
||||||
"zh:6692ef323e3b89de99934ad731f6a1850525bf8142916ae28ea4e4048d73a787",
|
|
||||||
"zh:a5afc98e9a4038516bb58e788cb77dea67a60dce780dfcd206d7373c5a56b776",
|
|
||||||
"zh:bf902cded709d84fa27fbf91b589c241f2238a6c4924e4e479eebd74320b93a5",
|
|
||||||
"zh:cab0e1e72c9cebcf669fc6f35ec28cb8ab2dffb0237afc8860aa40d23bf8a49f",
|
|
||||||
"zh:e523b99a48beec83d9bc04b2d336266044f9f53514cefb652fe6768611847196",
|
|
||||||
"zh:f593915e8a24829d322d2eaeedcb153328cf9042f0d84f66040dde1be70ede04",
|
|
||||||
"zh:fba1aff541133e2129dfda0160369635ab48503d5c44b8407ce5922ecc15d0bd",
|
|
||||||
]
|
|
||||||
}
|
|
||||||
|
|
||||||
provider "registry.opentofu.org/thegameprofi/proxmox" {
|
|
||||||
version = "2.10.0"
|
|
||||||
constraints = "2.10.0"
|
|
||||||
hashes = [
|
|
||||||
"h1:0wJlZKYgmnyXJ4AGKfML5m0/bhRUXty3vn4mSEYhFDM=",
|
|
||||||
"h1:2fmj6FkuQ8CyWdxJ83s87GJhRHiQFnzlSYSlA2q7pXg=",
|
|
||||||
"h1:5FQbfjTDMO4FCMXAQ5S+2CdVHw1gJoOicQRLVTCfrDI=",
|
|
||||||
"h1:60DcgYwyBroh3oUsdsZU/Z03LfEIidq2lTYKbhZDibU=",
|
|
||||||
"h1:6hrWJT18l4jfWrzO8GweRA1bIdzY8dDUkzB7+GW0Zgs=",
|
|
||||||
"h1:8bQZLEyLJqYnLnZC0zopbPSJVYXwQIb+G3nQ9Tc9bMU=",
|
|
||||||
"h1:DArex0up+F+sN3arAKanOSVc70b9kyTeTkuy6Abkor8=",
|
|
||||||
"h1:WUh8KlPtUdI+O7FJz6/d3BC+77tSBYjlNCxKjIGX7Kg=",
|
|
||||||
"h1:atdwdBFAPjyz7ICxoeMBrUn7tllqjJyIIClO5Mv3dlw=",
|
|
||||||
"h1:mnIwCGDAZ9O/UiQEXGWKXUeA6H1Tt/AkII/MjHw1SMg=",
|
|
||||||
"h1:sW+6tq318P9g0P+BwqEM74AHsJ6CmN+s6NgAJJxZk/Y=",
|
|
||||||
"h1:tWdlbcPuby+thPgC07FCBbIzzQAOn3XSvRibuHcdK04=",
|
|
||||||
"h1:yHr19pmBvjRCSbyWmNd3Y4iI7mCjeDgzC622q5kSSKE=",
|
|
||||||
"h1:zbDNS+UBlhewZUPt0s5ntUeAGOSTLj2jjHweCDKw6qo=",
|
|
||||||
]
|
|
||||||
}
|
|
||||||
|
|
@ -1,50 +0,0 @@
|
||||||
terraform {
|
|
||||||
required_version = ">= 1.6"
|
|
||||||
backend "s3" {
|
|
||||||
bucket = "fuku-terraform"
|
|
||||||
key = "vm-foundation/terraform"
|
|
||||||
region = "us-east-1"
|
|
||||||
}
|
|
||||||
required_providers {
|
|
||||||
proxmox = {
|
|
||||||
source = "thegameprofi/proxmox"
|
|
||||||
version = "2.10.0"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
provider "proxmox" {
|
|
||||||
pm_debug = true
|
|
||||||
}
|
|
||||||
|
|
||||||
module "master1" {
|
|
||||||
source = "../modules/proxmox-vm"
|
|
||||||
vm_id = 3001
|
|
||||||
vm_name = "master1.ramiel.fuku"
|
|
||||||
node_name = "ramiel"
|
|
||||||
ipconfig0 = "ip=192.168.1.31/24,gw=192.168.1.1"
|
|
||||||
memory = 5120
|
|
||||||
disk_storage_name = "storage"
|
|
||||||
core_count = 2
|
|
||||||
}
|
|
||||||
|
|
||||||
module "master2" {
|
|
||||||
source = "../modules/proxmox-vm"
|
|
||||||
vm_id = 3002
|
|
||||||
vm_name = "master2.ireul.fuku"
|
|
||||||
node_name = "ireul"
|
|
||||||
ipconfig0 = "ip=192.168.1.32/24,gw=192.168.1.1"
|
|
||||||
memory = 4096
|
|
||||||
core_count = 2
|
|
||||||
}
|
|
||||||
|
|
||||||
module "master3" {
|
|
||||||
source = "../modules/proxmox-vm"
|
|
||||||
vm_id = 3003
|
|
||||||
vm_name = "master3.ireul.fuku"
|
|
||||||
node_name = "ireul"
|
|
||||||
ipconfig0 = "ip=192.168.1.33/24,gw=192.168.1.1"
|
|
||||||
disk_storage_name = "local-lvm"
|
|
||||||
memory = 4096
|
|
||||||
core_count = 2
|
|
||||||
}
|
|
||||||
|
|
@ -1,4 +0,0 @@
|
||||||
PM_USER=
|
|
||||||
PM_PASS=
|
|
||||||
PM_API_URL=
|
|
||||||
TF_VAR_ci_password=
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue