From 6e2abce331048d98f894d39b2525fe885fcfde1d Mon Sep 17 00:00:00 2001 From: root Date: Sat, 28 Feb 2026 18:08:58 +0000 Subject: [PATCH] . --- terraform/control-plane1.bu | 6 +- terraform/main.tf | 6 +- terraform/terraform.tfstate | 2 +- terraform/terraform.tfstate.backup | 1174 +++++++++++++++++++++++++++- terraform/worker3.bu | 2 +- 5 files changed, 1171 insertions(+), 19 deletions(-) diff --git a/terraform/control-plane1.bu b/terraform/control-plane1.bu index 81e3ddd..a6452b9 100644 --- a/terraform/control-plane1.bu +++ b/terraform/control-plane1.bu @@ -354,6 +354,7 @@ systemd: Type=oneshot StandardOutput=journal+console StandardError=journal+console + ExecStart=/opt/bin/kubeadm reset -f ExecStart=/bin/sh -c 'echo "kubeadm-init.service started..."' @@ -377,6 +378,7 @@ systemd: RestartSec=120s [Install] WantedBy=multi-user.target + - name: install-calico.service enabled: true contents: | @@ -444,8 +446,8 @@ systemd: enabled: true contents: | [Unit] - Wants=kubeadm-init.service - After=kubeadm-init.service + Wants=install-calico.service + After=install-calico.service ConditionPathExists=!/var/lib/undercloud-stamps/install-ceph.done [Service] diff --git a/terraform/main.tf b/terraform/main.tf index 42eeacb..39860c5 100644 --- a/terraform/main.tf +++ b/terraform/main.tf @@ -319,7 +319,7 @@ resource "proxmox_virtual_environment_vm" "control_plane1" { # Hardware cpu { type = "host" - cores = 2 + cores = 4 } memory { dedicated = 4096 @@ -364,7 +364,7 @@ resource "proxmox_virtual_environment_vm" "control_plane2" { # Hardware cpu { type = "host" - cores = 2 + cores = 4 } memory { dedicated = 3072 @@ -408,7 +408,7 @@ resource "proxmox_virtual_environment_vm" "control_plane3" { # Hardware cpu { type = "host" - cores = 2 + cores = 4 } memory { dedicated = 3072 diff --git a/terraform/terraform.tfstate b/terraform/terraform.tfstate index 3f06c7b..7a1605c 100644 --- a/terraform/terraform.tfstate +++ b/terraform/terraform.tfstate @@ -1,7 +1,7 @@ { "version": 4, "terraform_version": "1.12.2", - "serial": 1412, + "serial": 1506, "lineage": "d92c42be-29f9-bad9-ef9a-3dc952ff5fa5", "outputs": {}, "resources": [], diff --git a/terraform/terraform.tfstate.backup b/terraform/terraform.tfstate.backup index 2c9bed6..c7efc9e 100644 --- a/terraform/terraform.tfstate.backup +++ b/terraform/terraform.tfstate.backup @@ -1,7 +1,7 @@ { "version": 4, "terraform_version": "1.12.2", - "serial": 1393, + "serial": 1479, "lineage": "d92c42be-29f9-bad9-ef9a-3dc952ff5fa5", "outputs": {}, "resources": [ @@ -14,10 +14,10 @@ { "schema_version": 0, "attributes": { - "content": "variant: flatcar\nversion: 1.1.0\n\npasswd:\n users:\n - name: core\n ssh_authorized_keys:\n - \"ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIHHEAlPo3v4U67Y3411pTjIMkQxwlFWdXrBJkSzXenDH flatcar@undercloud\"\n\nstorage:\n directories:\n - path: /opt/bin\n overwrite: true\n mode: 0755\n - path: /opt/cni/bin\n overwrite: true\n mode: 755\n - path: /etc/kubernetes/manifests\n #overwrite: true\n mode: 0755\n - path: /etc/install-calico\n overwrite: true\n mode: 0755\n - path: /var/lib/undercloud-stamps\n mode: 0755\n - path: /var/lib/rsyslog\n overwrite: true\n mode: 0755\n\n files:\n - path: /etc/hostname\n mode: 0644\n contents:\n inline: |\n control-plane1\n\n - path: /etc/systemd/network/00-eth.network\n mode: 0644\n contents:\n inline: |\n [Match]\n Name=eth*\n \n [Network]\n Address=fd00:0:0:2::91/64\n Address=2001:470:7116:2::91/64\n Gateway=2001:470:7116:2::3\n DNS=fd00:0:0:1::1\n Address=10.0.2.91/24\n Gateway=10.0.2.3\n DNS=10.0.1.1\n Domains=undercloud.local\n IPv6AcceptRA=no\n IPv6PrivacyExtensions=no\n\n - path: /etc/hosts\n mode: 0644\n overwrite: true\n contents:\n inline: |\n 127.0.0.1 localhost\n ::1 localhost ip6-localhost ip6-loopback\n\n 2001:470:7116:2::91 control-plane1.undercloud.local control-plane1\n 2001:470:7116:2::92 control-plane2.undercloud.local control-plane2\n 2001:470:7116:2::93 control-plane3.undercloud.local control-plane3\n 2001:470:7116:2::101 worker1.undercloud.local worker1\n 2001:470:7116:2::102 worker2.undercloud.local worker2\n 2001:470:7116:2::103 worker3.undercloud.local worker3\n\n fd00:0:0:2::91 control-plane1.undercloud.local control-plane1\n fd00:0:0:2::92 control-plane2.undercloud.local control-plane2\n fd00:0:0:2::93 control-plane3.undercloud.local control-plane3\n fd00:0:0:2::101 worker1.undercloud.local worker1\n fd00:0:0:2::102 worker2.undercloud.local worker2\n fd00:0:0:2::103 worker3.undercloud.local worker3\n\n 10.0.2.91 control-plane1.undercloud.local control-plane1\n 10.0.2.92 control-plane2.undercloud.local control-plane2\n 10.0.2.93 control-plane3.undercloud.local control-plane3\n 10.0.2.101 worker1.undercloud.local worker1\n 10.0.2.102 worker2.undercloud.local worker2\n 10.0.2.103 worker3.undercloud.local worker3\n\n - path: /etc/motd\n mode: 0644\n overwrite: true\n contents:\n inline: |\n *******************************************************************\n * AUTHORIZED ACCESS ONLY *\n * *\n * This system is part of a secured infrastructure. *\n * All activities are monitored and logged. *\n * Unauthorized access or misuse is strictly prohibited and *\n * may result in disciplinary and legal action. *\n *******************************************************************\n\n --------------------------------------------------------------------------------\n kubernetes controle plane Node\n\n Manage via:\n kubectl (kubectl)\n calico (calicoctl)\n velero - backup (velero)\n argocd https://argocd-server.argocd.svc.k8aux.undercloud.cf/\n --------------------------------------------------------------------------------\n\n - path: /etc/sysctl.d/99-k8s.conf\n mode: 0644\n contents:\n inline: |\n net.ipv4.ip_forward = 1\n net.ipv6.ip_forward = 1\n net.ipv6.conf.all.forwarding = 1\n net.ipv4.conf.all.forwarding = 1\n net.bridge.bridge-nf-call-iptables = 1\n net.bridge.bridge-nf-call-ip6tables = 1\n net.netfilter.nf_conntrack_max = 1000000\n net.ipv4.conf.all.rp_filter = 0\n net.ipv6.conf.all.disable_ipv6 = 0\n vm.overcommit_memory = 1\n fs.inotify.max_user_watches = 524288\n fs.inotify.max_user_instances = 4096\n kernel.panic = 10\n kernel.panic_on_oops = 1\n\n - path: /etc/flatcar/update.conf\n overwrite: true\n mode: 0420\n contents:\n inline: |\n REBOOT_STRATEGY=off\n\n - path: /opt/bin/kubeadm\n mode: 0755\n contents:\n source: \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubeadm\"\n\n - path: /opt/bin/kubelet\n mode: 0755\n contents:\n source: \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubelet\"\n\n - path: /opt/bin/kubectl\n mode: 0755\n contents:\n source: \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubectl\"\n\n - path: /opt/bin/calicoctl\n mode: 0755\n contents:\n source: \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/calicoctl\"\n\n - path: /opt/bin/velero\n mode: 0755\n contents:\n source: \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/velero\"\n \n - path: /opt/bin/rsyslogd\n mode: 0755\n contents:\n source: \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/rsyslogd\"\n\n - path: /etc/systemd/journald.conf.d/10-forward-to-syslog.conf\n mode: 0644\n contents:\n inline: |\n [Journal]\n ForwardToSyslog=yes\n Syslog=yes\n \n - path: /etc/rsyslog.conf\n mode: 0644\n contents:\n inline: |\n # Minimal rsyslog: receive from journald syslog socket and forward to remote\n\n global(workDirectory=\"/var/lib/rsyslog\")\n\n # This is the important part: pick up what journald forwards when ForwardToSyslog=yes\n module(load=\"imuxsock\" SysSock.Name=\"/run/systemd/journal/syslog\")\n\n # Forward everything to your syslog server (TCP recommended)\n action(\n type=\"omfwd\"\n target=\"syslog.undercloud.local\"\n port=\"514\"\n protocol=\"tcp\"\n action.resumeRetryCount=\"-1\"\n queue.type=\"linkedList\"\n queue.size=\"50000\"\n )\n\n - path: /etc/kubernetes/kubeadm-init.yaml\n mode: 0644\n contents:\n inline: |\n apiVersion: kubeadm.k8s.io/v1beta3\n kind: InitConfiguration\n bootstrapTokens:\n - token: \"kvg1hc.t3rewovrps426rof\"\n description: \"default kubeadm bootstrap token\"\n ttl: \"0\"\n nodeRegistration:\n name: control-plane1\n criSocket: unix:///run/containerd/containerd.sock\n kubeletExtraArgs:\n node-ip: \"2001:470:7116:2::91\"\n cluster-dns: \"10.0.91.53,2001:470:7116:f:1::53\"\n volume-plugin-dir: \"/opt/libexec/kubernetes/kubelet-plugins/volume/exec/\"\n localAPIEndpoint:\n advertiseAddress: \"2001:470:7116:2::91\"\n bindPort: 6443\n certificateKey: \"fee7c3e5cfcac7e4774c6efca0464a42d897f30f7300340d6578b5cfb4a3d34b\"\n ---\n apiVersion: kubeadm.k8s.io/v1beta3\n kind: ClusterConfiguration\n controlPlaneEndpoint: \"[fd00:0:0:2::100]:6443\"\n networking:\n podSubnet: \"2001:470:7116:a::/64,10.0.10.0/24\"\n serviceSubnet: \"2001:470:7116:f:1::/108,10.0.91.0/24\"\n dnsDomain: \"k8s.undercloud.local\"\n controllerManager:\n extraArgs:\n flex-volume-plugin-dir: \"/opt/libexec/kubernetes/kubelet-plugins/volume/exec/\"\n bind-address: '::'\n apiServer:\n extraArgs:\n enable-aggregator-routing: \"true\"\n proxy-client-cert-file: /etc/kubernetes/pki/front-proxy-client.crt\n proxy-client-key-file: /etc/kubernetes/pki/front-proxy-client.key\n requestheader-client-ca-file: /etc/kubernetes/pki/front-proxy-ca.crt\n requestheader-allowed-names: front-proxy-client\n requestheader-extra-headers-prefix: X-Remote-Extra-\n requestheader-group-headers: X-Remote-Group\n requestheader-username-headers: X-Remote-User\n\n ---\n apiVersion: kubelet.config.k8s.io/v1beta1\n kind: KubeletConfiguration\n address: \"::\"\n healthzBindAddress: \"::\"\n clusterDomain: \"k8s.undercloud.local\"\n clusterDNS:\n - \"2001:470:7116:f:1::53\"\n - \"10.0.91.53\"\n volumePluginDir: /opt/libexec/kubernetes/kubelet-plugins/volume/exec\n cgroupDriver: \"systemd\"\n authentication:\n anonymous:\n enabled: true\n webhook:\n enabled: true\n authorization:\n mode: Webhook\n\n - path: /etc/kubernetes/addons/kube-dns-fixed-svc.yaml\n mode: 0644\n contents:\n inline: |\n apiVersion: v1\n kind: Service\n metadata:\n name: kube-dns\n namespace: kube-system\n labels:\n k8s-app: kube-dns\n spec:\n type: ClusterIP\n ipFamilyPolicy: RequireDualStack\n ipFamilies: [IPv6, IPv4]\n clusterIP: 2001:470:7116:f:1::53\n clusterIPs:\n - 2001:470:7116:f:1::53\n - 10.0.91.53\n ports:\n - name: dns\n port: 53\n protocol: UDP\n targetPort: 53\n - name: dns-tcp\n port: 53\n protocol: TCP\n targetPort: 53\n - name: metrics\n port: 9153\n protocol: TCP\n targetPort: 9153\n selector:\n k8s-app: kube-dns\n\nsystemd:\n units:\n - name: rsyslog.service\n enabled: true\n contents: |\n [Unit]\n Description=rsyslog (journald -\u003e remote syslog)\n Wants=network-online.target\n After=network-online.target systemd-journald.service\n Requires=systemd-journald.service\n\n [Service]\n Type=simple\n ExecStart=/opt/bin/rsyslogd -n -f /etc/rsyslog.conf\n Restart=always\n RestartSec=2\n\n [Install]\n WantedBy=multi-user.target\n # --- Boot entrypoint: only this target is enabled at boot ---\n - name: undercloud-bootstrap.target\n enabled: true\n contents: |\n [Unit]\n Description=Undercloud Bootstrap Chain\n Wants=network-online.target\n After=network-online.target\n # Start the chain entry\n Wants=containerd.service kubelet.service kubeadm-init.service\n After=containerd.service kubelet.service kubeadm-init.service\n\n [Install]\n WantedBy=multi-user.target\n\n - name: modules-load.service\n enabled: true\n contents: |\n [Unit]\n Description=Load necessary kernel modules\n Before=containerd.service kubeadm-init.service\n\n [Service]\n Type=oneshot\n ExecStart=/usr/bin/modprobe br_netfilter\n ExecStart=/usr/bin/modprobe overlay\n RemainAfterExit=yes\n\n [Install]\n WantedBy=multi-user.target\n\n - name: systemd-networkd-wait-online.service\n enabled: true\n\n - name: containerd.service\n enabled: true\n contents: |\n [Unit]\n Description=containerd container runtime\n After=network.target modules-load.service\n Wants=modules-load.service\n\n [Service]\n ExecStart=/usr/bin/containerd\n Restart=always\n RestartSec=5\n Delegate=yes\n KillMode=process\n OOMScoreAdjust=-999\n\n [Install]\n WantedBy=multi-user.target\n\n - name: set-timezone.service\n enabled: true\n contents: |\n [Unit]\n Description=Set Timezone\n After=network-online.target\n Wants=network-online.target\n\n [Service]\n Type=oneshot\n StandardOutput=journal+console\n StandardError=journal+console\n ExecStart=/bin/sh -c 'echo \"setting timezone to Europe/Berlin\"'\n ExecStart=/usr/bin/timedatectl set-timezone Europe/Berlin\n ExecStart=/usr/bin/timedatectl set-ntp true\n\n [Install]\n WantedBy=multi-user.target\n\n - name: kubelet.service\n enabled: true\n contents: |\n [Unit]\n Description=kubelet, the Kubernetes Node Agent\n Documentation=https://kubernetes.io/docs/home\n Wants=network-online.target\n After=network-online.target containerd.service\n Requires=containerd.service\n\n [Service]\n Environment=\"KUBELET_KUBECONFIG_ARGS=--bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf\"\n Environment=\"KUBELET_CONFIG_ARGS=--config=/var/lib/kubelet/config.yaml\"\n EnvironmentFile=-/var/lib/kubelet/kubeadm-flags.env\n ExecStart=/opt/bin/kubelet $KUBELET_KUBECONFIG_ARGS $KUBELET_CONFIG_ARGS $KUBELET_KUBEADM_ARGS $KUBELET_EXTRA_ARGS\n Restart=always\n StartLimitInterval=0\n RestartSec=10\n\n [Install]\n WantedBy=multi-user.target\n\n # --- Chain step 1 ---\n - name: kubeadm-init.service\n enabled: false\n contents: |\n [Unit]\n Description=Kubeadm Init Cluster\n Wants=network-online.target\n After=network-online.target containerd.service kubelet.service\n Requires=containerd.service kubelet.service\n ConditionPathExists=!/etc/kubernetes/kubelet.conf\n\n [Service]\n Type=oneshot\n StandardOutput=journal+console\n StandardError=journal+console\n\n Environment=KUBECONFIG=/etc/kubernetes/admin.conf\n Environment=DATASTORE_TYPE=kubernetes\n Environment=PATH=/usr/bin/:/usr/sbin:/opt/bin:/opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent-uds/\n\n ExecStart=/bin/sh -eu -c '\\\n echo \"[kubeadm-init] started...\" ; \\\n echo \"[kubeadm-init] waiting for containerd socket...\" ; \\\n for i in $(seq 1 60); do test -S /run/containerd/containerd.sock \u0026\u0026 break; sleep 1; done ; \\\n echo \"[kubeadm-init] running kubeadm init...\" ; \\\n /opt/bin/kubeadm init --upload-certs --config=/etc/kubernetes/kubeadm-init.yaml ; \\\n echo \"[kubeadm-init] copying kubeconfig to core...\" ; \\\n mkdir -p /home/core/.kube ; \\\n cp -f /etc/kubernetes/admin.conf /home/core/.kube/config ; \\\n chown core:core /home/core/.kube/config ; \\\n echo \"[kubeadm-init] done.\" \\\n '\n\n # strictly start next step (serialization)\n ExecStartPost=/usr/bin/systemctl start install-calico.service\n\n [Install]\n WantedBy=undercloud-bootstrap.target\n\n # --- Chain step 2 ---\n - name: install-calico.service\n enabled: false\n contents: |\n [Unit]\n Description=Install Calico\n Requires=kubeadm-init.service\n After=kubeadm-init.service\n ConditionPathExists=!/var/lib/undercloud-stamps/install-calico.done\n\n [Service]\n Type=oneshot\n StandardOutput=journal+console\n StandardError=journal+console\n Environment=KUBECONFIG=/etc/kubernetes/admin.conf\n Environment=DATASTORE_TYPE=kubernetes\n Environment=PATH=/usr/bin/:/usr/sbin:/opt/bin\n\n ExecStart=/bin/sh -eu -c '\\\n echo \"[calico] waiting for API /readyz...\" ; \\\n for i in $(seq 1 180); do kubectl get --raw=/readyz \u003e/dev/null 2\u003e\u00261 \u0026\u0026 break; sleep 2; done ; \\\n echo \"[calico] create namespace + operator...\" ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/namespace.yaml ; \\\n kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/operator-crds.yaml || true ; \\\n kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/tigera-operator.yaml || true ; \\\n echo \"[calico] wait for tigera-operator...\" ; \\\n kubectl wait deployment -n tigera-operator tigera-operator --for condition=Available=True --timeout=1200s ; \\\n echo \"[calico] apply custom resources...\" ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/custom-resources.yaml ; \\\n echo \"[calico] wait for calico-apiserver...\" ; \\\n kubectl wait deployment -n calico-apiserver calico-apiserver --for condition=Available=True --timeout=1200s ; \\\n echo \"[calico] apply peers + pools...\" ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/calico-peer.yaml ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/ippools.yaml ; \\\n echo \"[calico] done.\" \\\n '\n\n ExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-calico.done\n ExecStartPost=/usr/bin/systemctl start install-ceph.service\n\n [Install]\n WantedBy=undercloud-bootstrap.target\n\n # --- Chain step 3 ---\n - name: install-ceph.service\n enabled: false\n contents: |\n [Unit]\n Description=Install Ceph CSI\n Requires=install-calico.service\n After=install-calico.service\n ConditionPathExists=!/var/lib/undercloud-stamps/install-ceph.done\n\n [Service]\n Type=oneshot\n StandardOutput=journal+console\n StandardError=journal+console\n Environment=KUBECONFIG=/etc/kubernetes/admin.conf\n Environment=DATASTORE_TYPE=kubernetes\n Environment=PATH=/usr/bin/:/usr/sbin:/opt/bin\n\n ExecStart=/bin/sh -eu -c '\\\n echo \"[ceph] waiting for API /readyz...\" ; \\\n for i in $(seq 1 180); do kubectl get --raw=/readyz \u003e/dev/null 2\u003e\u00261 \u0026\u0026 break; sleep 2; done ; \\\n echo \"[ceph] apply manifests...\" ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/namespace.yaml ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-nodeplugin-rbac.yaml ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-provisioner-rbac.yaml ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/secrets.yaml ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/ceph-conf.yaml ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-cephfsplugin-provisioner.yaml ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-cephfsplugin.yaml ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-config-map.yaml ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-encryption-kms-config.yaml ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csidriver.yaml ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/storage-class.yaml ; \\\n echo \"[ceph] done.\" \\\n '\n\n ExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-ceph.done\n ExecStartPost=/usr/bin/systemctl start install-gitea.service\n\n [Install]\n WantedBy=undercloud-bootstrap.target\n\n # --- Chain step 4 ---\n - name: install-gitea.service\n enabled: false\n contents: |\n [Unit]\n Description=Install Gitea\n Requires=install-ceph.service\n After=install-ceph.service\n ConditionPathExists=!/var/lib/undercloud-stamps/install-gitea.done\n\n [Service]\n Type=oneshot\n StandardOutput=journal+console\n StandardError=journal+console\n Environment=KUBECONFIG=/etc/kubernetes/admin.conf\n Environment=DATASTORE_TYPE=kubernetes\n Environment=PATH=/usr/bin/:/usr/sbin:/opt/bin\n\n ExecStart=/bin/sh -eu -c '\\\n echo \"[gitea] wait for ceph provisioner...\" ; \\\n kubectl wait deployment -n ceph csi-cephfsplugin-provisioner --for condition=Available=True --timeout=1200s ; \\\n echo \"[gitea] apply manifests...\" ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/namespace.yaml ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/secrets.yaml ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/db.yaml ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/adminer.yaml ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/gitea.yaml ; \\\n echo \"[gitea] wait for gitea deployment...\" ; \\\n kubectl -n gitea wait deployment gitea --for=condition=Available=True --timeout=1200s ; \\\n echo \"[gitea] run startup...\" ; \\\n kubectl exec deploy/gitea -n gitea -- /bin/startup.sh ; \\\n echo \"[gitea] done.\" \\\n '\n\n ExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-gitea.done\n ExecStartPost=/usr/bin/systemctl start install-argocd.service\n\n [Install]\n WantedBy=undercloud-bootstrap.target\n\n # --- Chain step 5 ---\n - name: install-argocd.service\n enabled: false\n contents: |\n [Unit]\n Description=Install ArgoCD\n Requires=install-calico.service install-gitea.service\n After=install-calico.service install-gitea.service\n ConditionPathExists=!/var/lib/undercloud-stamps/install-argocd.done\n\n [Service]\n Type=oneshot\n StandardOutput=journal+console\n StandardError=journal+console\n Environment=KUBECONFIG=/etc/kubernetes/admin.conf\n Environment=DATASTORE_TYPE=kubernetes\n Environment=PATH=/usr/bin/:/usr/sbin:/opt/bin\n\n ExecStart=/bin/sh -eu -c '\\\n echo \"[argocd] wait for coredns...\" ; \\\n kubectl -n kube-system wait deploy coredns --for=condition=Available=True --timeout=1200s ; \\\n echo \"[argocd] install...\" ; \\\n kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/namespace.yaml ; \\\n kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/install.yaml ; \\\n kubectl -n argocd wait deploy argocd-server --for=condition=Available=True --timeout=1200s ; \\\n kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/repo.yaml ; \\\n kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/apps.yaml ; \\\n kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/ingress.yaml ; \\\n echo \"[argocd] done.\" \\\n '\n\n ExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-argocd.done\n ExecStartPost=/usr/bin/systemctl start pin-service-ips.service\n\n [Install]\n WantedBy=undercloud-bootstrap.target\n\n # --- Chain step 6 (final) ---\n - name: pin-service-ips.service\n enabled: false\n contents: |\n [Unit]\n Description=Pin fixed dual-stack ClusterIPs for kube-dns, argocd-server and whisker\n Requires=install-argocd.service install-calico.service kubeadm-init.service\n After=install-argocd.service install-calico.service kubeadm-init.service network-online.target\n Wants=network-online.target\n ConditionPathExists=!/var/lib/undercloud-stamps/pin-service-ips.done\n\n [Service]\n Type=oneshot\n StandardOutput=journal+console\n StandardError=journal+console\n Environment=KUBECONFIG=/etc/kubernetes/admin.conf\n Environment=PATH=/usr/bin:/usr/sbin:/opt/bin\n ExecStart=/bin/sh -eu -c '\\\n echo \"[pin-service-ips] waiting for API...\" ; \\\n for i in $(seq 1 120); do kubectl get --raw=/readyz \u003e/dev/null 2\u003e\u00261 \u0026\u0026 break; sleep 2; done ; \\\n echo \"[pin-service-ips] ensure namespaces exist...\" ; \\\n kubectl get ns kube-system \u003e/dev/null ; \\\n kubectl get ns argocd \u003e/dev/null 2\u003e\u00261 || kubectl create ns argocd ; \\\n kubectl get ns calico-system \u003e/dev/null ; \\\n echo \"[pin-service-ips] wait for coredns/argocd readiness (best effort)...\" ; \\\n kubectl -n kube-system wait deploy coredns --for=condition=Available=True --timeout=300s || true ; \\\n kubectl -n argocd wait deploy argocd-server --for=condition=Available=True --timeout=600s || true ; \\\n echo \"[pin-service-ips] replace Services with fixed ClusterIPs...\" ; \\\n kubectl -n kube-system delete svc kube-dns --ignore-not-found ; \\\n kubectl apply -f /etc/kubernetes/addons/kube-dns-fixed-svc.yaml ; \\\n kubectl -n argocd delete svc argocd-server --ignore-not-found ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/service.yaml ; \\\n kubectl -n calico-system delete svc whisker --ignore-not-found || true ; \\\n kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/whisker.yaml || true ; \\\n echo \"[pin-service-ips] done.\" \\\n '\n\n ExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/pin-service-ips.done\n\n [Install]\n WantedBy=undercloud-bootstrap.target\n", - "id": "1153550859", + "content": "variant: flatcar\nversion: 1.1.0\n\npasswd:\n users:\n - name: core\n ssh_authorized_keys:\n - \"ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIHHEAlPo3v4U67Y3411pTjIMkQxwlFWdXrBJkSzXenDH flatcar@undercloud\"\n\nstorage:\n directories:\n - path: /opt/bin\n overwrite: true\n mode: 0755\n - path: /opt/cni/bin\n overwrite: true\n mode: 755\n - path: /etc/kubernetes/manifests\n #overwrite: true\n mode: 0755\n - path: /etc/install-calico\n overwrite: true\n mode: 0755\n - path: /var/lib/undercloud-stamps\n mode: 0755\n\n files:\n - path: /etc/hostname\n mode: 0644\n contents:\n inline: |\n control-plane1\n\n - path: /etc/systemd/network/00-eth.network\n mode: 0644\n contents:\n inline: |\n [Match]\n Name=eth*\n \n [Network]\n Address=fd00:0:0:2::91/64\n Address=2001:470:7116:2::91/64\n Gateway=2001:470:7116:2::3\n DNS=fd00:0:0:1::1\n Address=10.0.2.91/24\n Gateway=10.0.2.3\n DNS=10.0.1.1\n Domains=undercloud.local\n IPv6AcceptRA=no\n IPv6PrivacyExtensions=no\n\n - path: /etc/hosts\n mode: 0644\n overwrite: true\n contents:\n inline: |\n 127.0.0.1 localhost\n ::1 localhost ip6-localhost ip6-loopback\n\n 2001:470:7116:2::91 control-plane1.undercloud.local control-plane1\n 2001:470:7116:2::92 control-plane2.undercloud.local control-plane2\n 2001:470:7116:2::93 control-plane3.undercloud.local control-plane3\n 2001:470:7116:2::101 worker1.undercloud.local worker1\n 2001:470:7116:2::102 worker2.undercloud.local worker2\n 2001:470:7116:2::103 worker3.undercloud.local worker3\n\n fd00:0:0:2::91 control-plane1.undercloud.local control-plane1\n fd00:0:0:2::92 control-plane2.undercloud.local control-plane2\n fd00:0:0:2::93 control-plane3.undercloud.local control-plane3\n fd00:0:0:2::101 worker1.undercloud.local worker1\n fd00:0:0:2::102 worker2.undercloud.local worker2\n fd00:0:0:2::103 worker3.undercloud.local worker3\n\n 10.0.2.91 control-plane1.undercloud.local control-plane1\n 10.0.2.92 control-plane2.undercloud.local control-plane2\n 10.0.2.93 control-plane3.undercloud.local control-plane3\n 10.0.2.101 worker1.undercloud.local worker1\n 10.0.2.102 worker2.undercloud.local worker2\n 10.0.2.103 worker3.undercloud.local worker3\n\n - path: /etc/motd\n mode: 0644\n overwrite: true\n contents:\n inline: |\n *******************************************************************\n * AUTHORIZED ACCESS ONLY *\n * *\n * This system is part of a secured infrastructure. *\n * All activities are monitored and logged. *\n * Unauthorized access or misuse is strictly prohibited and *\n * may result in disciplinary and legal action. *\n *******************************************************************\n\n --------------------------------------------------------------------------------\n kubernetes controle plane Node\n\n Manage via:\n kubectl (kubectl)\n calico (calicoctl)\n velero - backup (velero)\n argocd https://argocd-server.argocd.svc.k8aux.undercloud.cf/\n --------------------------------------------------------------------------------\n\n - path: /etc/sysctl.d/99-k8s.conf\n mode: 0644\n contents:\n inline: |\n net.ipv4.ip_forward = 1\n net.ipv6.ip_forward = 1\n net.ipv6.conf.all.forwarding = 1\n net.ipv4.conf.all.forwarding = 1\n net.bridge.bridge-nf-call-iptables = 1\n net.bridge.bridge-nf-call-ip6tables = 1\n net.netfilter.nf_conntrack_max = 1000000\n net.ipv4.conf.all.rp_filter = 0\n net.ipv6.conf.all.disable_ipv6 = 0\n vm.overcommit_memory = 1\n fs.inotify.max_user_watches = 524288\n fs.inotify.max_user_instances = 4096\n kernel.panic = 10\n kernel.panic_on_oops = 1\n\n - path: /etc/flatcar/update.conf\n overwrite: true\n mode: 0420\n contents:\n inline: |\n REBOOT_STRATEGY=off\n\n - path: /opt/bin/kubeadm\n mode: 0755\n contents:\n source: \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubeadm\"\n\n - path: /opt/bin/kubelet\n mode: 0755\n contents:\n source: \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubelet\"\n\n - path: /opt/bin/kubectl\n mode: 0755\n contents:\n source: \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubectl\"\n\n - path: /opt/bin/calicoctl\n mode: 0755\n contents:\n source: \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/calicoctl\"\n\n - path: /opt/bin/velero\n mode: 0755\n contents:\n source: \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/velero\"\n\n - path: /etc/kubernetes/kubeadm-init.yaml\n mode: 0644\n contents:\n inline: |\n apiVersion: kubeadm.k8s.io/v1beta3\n kind: InitConfiguration\n bootstrapTokens:\n - token: \"kvg1hc.t3rewovrps426rof\"\n description: \"default kubeadm bootstrap token\"\n ttl: \"0\"\n nodeRegistration:\n name: control-plane1\n criSocket: unix:///run/containerd/containerd.sock\n kubeletExtraArgs:\n node-ip: \"2001:470:7116:2::91\"\n cluster-dns: \"10.0.91.53,2001:470:7116:f:1::53\"\n volume-plugin-dir: \"/opt/libexec/kubernetes/kubelet-plugins/volume/exec/\"\n localAPIEndpoint:\n advertiseAddress: \"2001:470:7116:2::91\"\n bindPort: 6443\n certificateKey: \"fee7c3e5cfcac7e4774c6efca0464a42d897f30f7300340d6578b5cfb4a3d34b\"\n ---\n apiVersion: kubeadm.k8s.io/v1beta3\n kind: ClusterConfiguration\n controlPlaneEndpoint: \"[fd00:0:0:2::100]:6443\"\n networking:\n podSubnet: \"2001:470:7116:a::/64,10.0.10.0/24\"\n serviceSubnet: \"2001:470:7116:f:1::/108,10.0.91.0/24\"\n dnsDomain: \"k8s.undercloud.local\"\n controllerManager:\n extraArgs:\n flex-volume-plugin-dir: \"/opt/libexec/kubernetes/kubelet-plugins/volume/exec/\"\n bind-address: '::'\n apiServer:\n extraArgs:\n enable-aggregator-routing: \"true\"\n proxy-client-cert-file: /etc/kubernetes/pki/front-proxy-client.crt\n proxy-client-key-file: /etc/kubernetes/pki/front-proxy-client.key\n requestheader-client-ca-file: /etc/kubernetes/pki/front-proxy-ca.crt\n requestheader-allowed-names: front-proxy-client\n requestheader-extra-headers-prefix: X-Remote-Extra-\n requestheader-group-headers: X-Remote-Group\n requestheader-username-headers: X-Remote-User\n\n ---\n apiVersion: kubelet.config.k8s.io/v1beta1\n kind: KubeletConfiguration\n address: \"::\"\n healthzBindAddress: \"::\"\n clusterDomain: \"k8s.undercloud.local\"\n clusterDNS:\n - \"2001:470:7116:f:1::53\"\n - \"10.0.91.53\"\n volumePluginDir: /opt/libexec/kubernetes/kubelet-plugins/volume/exec\n cgroupDriver: \"systemd\"\n authentication:\n anonymous:\n enabled: true\n webhook:\n enabled: true\n authorization:\n mode: Webhook\n\n - path: /etc/kubernetes/addons/kube-dns-fixed-svc.yaml\n mode: 0644\n contents:\n inline: |\n apiVersion: v1\n kind: Service\n metadata:\n name: kube-dns\n namespace: kube-system\n labels:\n k8s-app: kube-dns\n spec:\n type: ClusterIP\n ipFamilyPolicy: RequireDualStack\n ipFamilies: [IPv6, IPv4]\n clusterIP: 2001:470:7116:f:1::53\n clusterIPs:\n - 2001:470:7116:f:1::53\n - 10.0.91.53\n ports:\n - name: dns\n port: 53\n protocol: UDP\n targetPort: 53\n - name: dns-tcp\n port: 53\n protocol: TCP\n targetPort: 53\n - name: metrics\n port: 9153\n protocol: TCP\n targetPort: 9153\n selector:\n k8s-app: kube-dns\n\n\nsystemd:\n units:\n - name: modules-load.service\n enabled: true\n contents: |\n [Unit]\n Description=Load necessary kernel modules\n Before=containerd.service kubeadm-init.service\n\n [Service]\n Type=oneshot\n\n ExecStart=/usr/bin/modprobe br_netfilter\n ExecStart=/usr/bin/modprobe overlay\n RemainAfterExit=yes\n\n [Install]\n WantedBy=multi-user.target\n\n - name: systemd-networkd-wait-online.service\n enabled: true\n\n - name: containerd.service\n enabled: true\n contents: |\n [Unit]\n Description=containerd container runtime\n After=network.target modules-load.service\n\n [Service]\n #StandardOutput=journal+console\n #StandardError=journal+console\n\n ExecStart=/usr/bin/containerd\n Restart=always\n RestartSec=5\n Delegate=yes\n KillMode=process\n OOMScoreAdjust=-999\n\n [Install]\n WantedBy=multi-user.target\n\n - name: set-timezone.service\n enabled: true\n contents: |\n [Unit]\n Description=Set Timezone\n After=network-online.target\n Wants=network-online.target\n [Service]\n StandardOutput=journal+console\n StandardError=journal+console\n\n ExecStart=/bin/sh -c 'echo \"setting timezone to Europe/Berlin\"'\n StandardOutput=journal+console\n StandardError=journal+console\n Type=oneshot\n Restart=on-failure\n ExecStart=/usr/bin/timedatectl set-timezone Europe/Berlin\n ExecStart=/usr/bin/timedatectl set-ntp true \n [Install]\n WantedBy=multi-user.target\n\n - name: kubelet.service\n enabled: true\n contents: |\n [Unit]\n Description=kubelet, the Kubernetes Node Agent\n Documentation=https://kubernets.io/docs/home\n Wants=network-online.target\n After=network-online.target\n [Service]\n #StandardOutput=journal+console\n #StandardError=journal+console\n\n #EnvironmentFile=/run/metadata/coreos\n Environment=\"KUBELET_KUBECONFIG_ARGS=--bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf\"\n Environment=\"KUBELET_CONFIG_ARGS=--config=/var/lib/kubelet/config.yaml\"\n # This is a file that \"kubeadm init\" and \"kubeadm join\" generates at runtime, populating the KUBELET_KUBEADM_ARGS variable dynamically\n EnvironmentFile=-/var/lib/kubelet/kubeadm-flags.env\n ExecStart=/opt/bin/kubelet $KUBELET_KUBECONFIG_ARGS $KUBELET_CONFIG_ARGS $KUBELET_KUBEADM_ARGS $KUBELET_EXTRA_ARGS\n Restart=always\n StartLimitInterval=0\n RestartSec=10\n [Install]\n WantedBy=multi-user.target\n\n - name: kubeadm-init.service\n enabled: true\n contents: |\n [Unit]\n Description=Kubeadm Init Cluster\n After=network-online.target containerd.service kubelet.service\n Wants=network-online.target\n ConditionPathExists=!/etc/kubernetes/kubelet.conf\n\n [Service]\n Type=oneshot\n StandardOutput=journal+console\n StandardError=journal+console\n ExecStart=/opt/bin/kubeadm reset -f\n\n ExecStart=/bin/sh -c 'echo \"kubeadm-init.service started...\"'\n\n # Environment\n Environment=KUBECONFIG=/etc/kubernetes/admin.conf\n Environment=DATASTORE_TYPE=kubernetes\n Environment=PATH=/usr/bin/:/usr/sbin:/opt/bin:/opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent-uds/\n \n ExecStartPre=/bin/sleep 30s\n ExecStart=/bin/sh -c 'echo \"running kubeadm init...\"'\n ExecStart=/opt/bin/kubeadm init --upload-certs --config=/etc/kubernetes/kubeadm-init.yaml\n \n # copy files for kubectl\n ExecStart=/bin/sh -c 'echo \"copying files (admin.conf) to core home folder.\"'\n ExecStartPost=/usr/bin/mkdir -p /home/core/.kube\n ExecStartPost=/usr/bin/cp -i /etc/kubernetes/admin.conf /home/core/.kube/config\n ExecStartPost=/usr/bin/chown core:core /home/core/.kube/config\n \n ExecStartPost=/usr/bin/systemctl disable kubeadm-init.service\n Restart=on-failure\n RestartSec=120s\n [Install]\n WantedBy=multi-user.target\n\n - name: install-calico.service\n enabled: true\n contents: |\n [Unit]\n Wants=kubeadm-init.service\n After=kubeadm-init.service\n ConditionPathExists=!/var/lib/undercloud-stamps/install-calico.done\n\n\n [Service]\n StandardOutput=journal+console\n StandardError=journal+console\n\n ExecStart=/bin/sh -c 'echo \"install.calico.service started...\"'\n Environment=KUBECONFIG=/etc/kubernetes/admin.conf\n Environment=DATASTORE_TYPE=kubernetes\n Environment=PATH=/usr/bin/:/usr/sbin:/opt/bin\n Type=oneshot\n StandardOutput=journal+console\n StandardError=journal+console\n ExecStart=/bin/sh -c 'echo \"witing 30s...\"'\n ExecStart=/bin/sleep 30s\n ExecStart=/bin/sh -c 'echo \"create calico namespace...\"'\n ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/namespace.yaml\n ExecStart=/bin/sh -c 'echo \"install tigera operator...\"'\n ExecStart=-/opt/bin/kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/operator-crds.yaml\n ExecStart=-/opt/bin/kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/tigera-operator.yaml\n ExecStart=/bin/sh -c 'echo \"witing 60s...\"'\n ExecStart=/bin/sleep 60s\n ExecStart=/bin/sh -c 'echo \"witing for tigera operator... (20mini max)\"'\n ExecStart=/opt/bin/kubectl wait deployment -n tigera-operator tigera-operator --for condition=Available=True --timeout=1200s\n ExecStart=/bin/sh -c 'echo \"create clico custom ressources...\"'\n ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/custom-resources.yaml\n \n ExecStart=/bin/sh -c 'echo \"witing 3m..\"'\n ExecStart=/bin/sleep 3m\n #ExecStart=/bin/sh -c 'echo \"apply calico (calico-apiserver)...\"'\n #ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/calico.yaml\n #ExecStart=/bin/sh -c 'echo \"witing 1m...\"'\n #ExecStart=/bin/sleep 2m\n ExecStart=/bin/sh -c 'echo \"witing calico-apiserver... (20mini max)\"'\n ExecStart=/opt/bin/kubectl wait deployment -n calico-apiserver calico-apiserver --for condition=Available=True --timeout=1200s\n ExecStart=/bin/sh -c 'echo \"witing 120s...\"'\n ExecStart=/bin/sleep 2m\n ExecStart=/bin/sh -c 'echo \"apply calico-peers...\"'\n ExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/calico-peer.yaml\n ExecStart=/bin/sh -c 'echo \"witing 60s...\"'\n ExecStart=/bin/sleep 1m\n ExecStart=/bin/sh -c 'echo \"apply calico-ippools...\"'\n ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/ippools.yaml\n \n #ExecStart=/bin/sh -c 'echo \"witing for whisker..\"'\n #ExecStart=/opt/bin/kubectl wait deployment -n calico-system whisker --for condition=Available=True --timeout=1200s\n #ExecStart=/bin/sh -c 'echo \"port-forward -n calico-system service/whisker 8081:8081\"'\n #ExecStart=/opt/bin/kubectl port-forward -n calico-system service/whisker 8081:8081\n \n ExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-calico.done\n ExecStart=/usr/bin/systemctl disable install-calico.service\n #RemainAfterExit=true\n Restart=on-failure\n RestartSec=120s\n [Install]\n WantedBy=multi-user.target\n - name: install-ceph.service\n enabled: true\n contents: |\n [Unit]\n Wants=kubeadm-init.service\n After=kubeadm-init.service\n ConditionPathExists=!/var/lib/undercloud-stamps/install-ceph.done\n\n [Service]\n StandardOutput=journal+console\n StandardError=journal+console\n\n ExecStart=/bin/sh -c 'echo \"install.ceph.service started...\"'\n Environment=KUBECONFIG=/etc/kubernetes/admin.conf\n Environment=DATASTORE_TYPE=kubernetes\n Environment=PATH=/usr/bin/:/usr/sbin:/opt/bin\n Type=oneshot\n \n StandardOutput=journal+console\n StandardError=journal+console\n ExecStart=/bin/sh -c 'echo \"witing 30s...\"'\n ExecStart=/bin/sleep 30s\n ExecStart=/bin/sh -c 'echo \"create ceph namespace...\"'\n ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/namespace.yaml\n \n ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-nodeplugin-rbac.yaml\n ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-provisioner-rbac.yaml\n\n ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/secrets.yaml\n\n ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/ceph-conf.yaml\n ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-cephfsplugin-provisioner.yaml\n ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-cephfsplugin.yaml\n ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-config-map.yaml\n ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-encryption-kms-config.yaml\n \n ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csidriver.yaml\n \n ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/storage-class.yaml\n \n ExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-ceph.done\n ExecStart=/usr/bin/systemctl disable install-ceph.service\n #RemainAfterExit=true\n Restart=on-failure\n RestartSec=120s\n [Install]\n WantedBy=multi-user.target\n - name: install-gitea.service\n enabled: true\n contents: |\n [Unit]\n Wants=install-ceph.service\n After=install-ceph.service\n ConditionPathExists=!/var/lib/undercloud-stamps/install-gitea.done\n \n [Service]\n StandardOutput=journal+console\n StandardError=journal+console\n Environment=KUBECONFIG=/etc/kubernetes/admin.conf\n Environment=DATASTORE_TYPE=kubernetes\n Environment=PATH=/usr/bin/:/usr/sbin:/opt/bin\n Type=oneshot\n \n ExecStart=/opt/bin/kubectl wait deployment -n ceph csi-cephfsplugin-provisioner --for condition=Available=True --timeout=1200s\n ExecStart=/bin/sleep 4m\n ExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/namespace.yaml\n ExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/secrets.yaml\n ExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/db.yaml\n ExecStart=/bin/sleep 60s\n ExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/adminer.yaml\n ExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/gitea.yaml\n ExecStart=/bin/sleep 3m\n ExecStart=/opt/bin/kubectl exec deploy/gitea -n gitea -- /bin/startup.sh\n\n ExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-gitea.done\n ExecStart=/usr/bin/systemctl disable install-gitea.service\n Restart=on-failure\n RestartSec=120s\n [Install]\n WantedBy=multi-user.target\n - name: install-argocd.service\n enabled: true\n contents: |\n [Unit]\n Wants=install-calico.service\n After=install-calico.service\n ConditionPathExists=!/var/lib/undercloud-stamps/install-argocd.done\n [Service]\n StandardOutput=journal+console\n StandardError=journal+console\n\n Environment=KUBECONFIG=/etc/kubernetes/admin.conf\n Environment=DATASTORE_TYPE=kubernetes\n Environment=PATH=/usr/bin/:/usr/sbin:/opt/bin\n Type=oneshot\n\n ExecStart=/opt/bin/kubectl wait deployment -n kube-system coredns --for condition=Available=True --timeout=600s\n \n ExecStart=/bin/sleep 1m\n ExecStart=/opt/bin/kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/namespace.yaml\n ExecStart=/opt/bin/kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/install.yaml\n ExecStart=/opt/bin/kubectl wait deployment -n argocd argocd-server --for condition=Available=True --timeout=600s\n\n ExecStart=/opt/bin/kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/repo.yaml\n ExecStart=/bin/sleep 10s\n ExecStart=/opt/bin/kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/apps.yaml\n\n ExecStart=/bin/sleep 10s\n ExecStart=/opt/bin/kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/ingress.yaml\n\n\n #ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/repos/k8aux-bootstrap.yaml\n #ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/repos/k8aux-apps.yaml\n #ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/apps/argocd.yaml\n #ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/apps/calico.yaml\n #ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/apps/rook-ceph.yaml\n #ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/apps/gitea.yaml\n #ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/argocd/argocd-secret.yaml\n ##ExecStart=/bin/sleep 10m\n #ExecStart=/opt/bin/kubectl wait deployment -n gitea gitea --for condition=Available=True --timeout=4800s\n #ExecStart=/bin/sleep 10m\n #ExecStart=/opt/bin/kubectl apply -n argocd -f http://gitea.gitea.svc.k8aux.undercloud.cf:3000/undercloud/k8aux-apps/raw/branch/main/app-of-apps/app-of-apps.yaml\n \n ExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-argocd.done\n ExecStart=/usr/bin/systemctl disable install-argocd.service\n Restart=on-failure\n RestartSec=120s\n [Install]\n WantedBy=multi-user.target\n - name: pin-service-ips.service\n enabled: true\n contents: |\n [Unit]\n Description=Pin fixed dual-stack ClusterIPs for kube-dns, argocd-server and whisker\n After=install-argocd.service install-calico.service kubeadm-init.service network-online.target\n Wants=install-argocd.service install-calico.service kubeadm-init.service network-online.target\n ConditionPathExists=!/var/lib/undercloud-stamps/pin-service-ips.done\n\n\n [Service]\n Type=oneshot\n StandardOutput=journal+console\n StandardError=journal+console\n Environment=KUBECONFIG=/etc/kubernetes/admin.conf\n Environment=PATH=/usr/bin:/usr/sbin:/opt/bin\n ExecStart=/bin/sh -eu -c '\\\n echo \"[pin-service-ips] waiting for API...\" ; \\\n for i in $(seq 1 120); do kubectl get --raw=/readyz \u003e/dev/null 2\u003e\u00261 \u0026\u0026 break; sleep 2; done ; \\\n echo \"[pin-service-ips] ensure namespaces exist...\" ; \\\n kubectl get ns kube-system \u003e/dev/null ; \\\n kubectl get ns argocd \u003e/dev/null 2\u003e\u00261 || kubectl create ns argocd ; \\\n kubectl get ns calico-system \u003e/dev/null ; \\\n echo \"[pin-service-ips] wait for coredns/argocd readiness (best effort)...\" ; \\\n kubectl -n kube-system wait deploy coredns --for=condition=Available=True --timeout=300s || true ; \\\n kubectl -n argocd wait deploy argocd-server --for=condition=Available=True --timeout=600s || true ; \\\n echo \"[pin-service-ips] replace Services with fixed ClusterIPs...\" ; \\\n kubectl -n kube-system delete svc kube-dns --ignore-not-found ; \\\n kubectl apply -f /etc/kubernetes/addons/kube-dns-fixed-svc.yaml ; \\\n kubectl -n argocd delete svc argocd-server --ignore-not-found ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/service.yaml ; \\\n kubectl -n calico-system delete svc whisker --ignore-not-found || true ; \\\n kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/whisker.yaml || true ; \\\n echo \"[pin-service-ips] done.\" \\\n '\n\n ExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/pin-service-ips.done\n Restart=on-failure\n RestartSec=120s\n [Install]\n WantedBy=multi-user.target\n\n", + "id": "1481435391", "pretty_print": true, - "rendered": "{\n \"ignition\": {\n \"config\": {\n \"replace\": {\n \"verification\": {}\n }\n },\n \"proxy\": {},\n \"security\": {\n \"tls\": {}\n },\n \"timeouts\": {},\n \"version\": \"3.4.0\"\n },\n \"kernelArguments\": {},\n \"passwd\": {\n \"users\": [\n {\n \"name\": \"core\",\n \"sshAuthorizedKeys\": [\n \"ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIHHEAlPo3v4U67Y3411pTjIMkQxwlFWdXrBJkSzXenDH flatcar@undercloud\"\n ]\n }\n ]\n },\n \"storage\": {\n \"directories\": [\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/opt/bin\",\n \"user\": {},\n \"mode\": 493\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/opt/cni/bin\",\n \"user\": {},\n \"mode\": 755\n },\n {\n \"group\": {},\n \"path\": \"/etc/kubernetes/manifests\",\n \"user\": {},\n \"mode\": 493\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/install-calico\",\n \"user\": {},\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/var/lib/undercloud-stamps\",\n \"user\": {},\n \"mode\": 493\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/var/lib/rsyslog\",\n \"user\": {},\n \"mode\": 493\n }\n ],\n \"files\": [\n {\n \"group\": {},\n \"path\": \"/etc/hostname\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"\",\n \"source\": \"data:,control-plane1%0A\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"path\": \"/etc/systemd/network/00-eth.network\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/2SMMWuFMBRG9/yUDjE3FcVABsFSOlSkHcUhJLco1aQkUeu/f+h76PC407mH77SfKuq+I7WaUGLsXwhpa4yr878dKY3xGIL8MYyJ/bgQBSRZehrOGIg0ZyIHyC79riKuanvWr6Sqv68eCAFnCxhllNMCEn4VHs/77gCgQCo3qcEGOVuDXo9uNnR0Wo3ko1myUmv8i1+ltO7gxg+L0tvbf0QbBmfDLm4BAAD//4CQfvX4AAAA\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/hosts\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/5ST7W6DIBRA//MUvEDJ/Vhqdt/GqcuWEjGszV5/wVEjlBa5/sLLOYknEakzYMCg1tq6obdf7ueqRML5f7a3+ns5n/KTWz764aIUAaC8dSAd4llI5B21Htx89c6eFtvPE5rbPE5+sO42mlWT7QsOyhxUcVDBwZmDKw5+dCCg/nX+MvnCR8RFiaK4pGcUlSiOS35GsVKfI4CEJ5YO01g7UVBBUY+dKLigqLfeK0LmMNXUKUQ76EXpFOId9Co0hp+DzNq4MfAdpfawd5Tbg0Z0bVkNuV2mAwG3y3wg3F8AAAD//+yy88FYBAAA\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/motd\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/6yRwYobMQyG7/MUOmYX4lzL3sJ2oYV2A01yaG+KrMyIeOxBkodOn74kk0MPaSg0/8WWkD4+4+f/T/MMf8l6v/u0+fb5x9tHWL++vm23sHn/8v3G4B3GP+fC2HViYJM59yAGA6pDOQKCMVXlCJKPiuZayatyuMVYpwRILqO4sAEqQ1+yeDnvY46QSttyDPc89hmrd0Xl13mHiM2gKPRi1fhsZq5CniYYtHRyEL+y/2D0OIGy1eQgGaIYyZAko06zBbc4i5Ycbns84G+b5YPTnOqBNbOzAZXsWhLDkDAzvJfITfMVM7YMo+DLZZY8weJ6eWoIk1CBxXxeWiMn1gJLOCCd6gCLufHUoLaFInTug72sVnO5NNaRNcxVsJHC6QPWn6HmyEqp1BjouHr8u38HAAD//04R3qpqAwAA\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"path\": \"/etc/sysctl.d/99-k8s.conf\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/4zQX27DIAwG8PeeIhcIyqos6h56FuQS01kFGxmWtrefsmXZHzE13wsS+iH7g7EYSlNvKFkvegUdm785Nk+7xQ2bnRP2BkIwiyY+V1y/xZ2UxjMuR8u+dRBCS6nAKWB+7IYFfjnG4ikUVMPeOmEuCu5iI9zWud1HKltqsp9vf7TuKq1HyvNQO99+uykamVCdxEjFRoyi99ov+myIpZC/mwg3+5ZR7RWKe53rru553+8PhyomzgXYrfzY9N3LsLugMgaTgMk1/2Ru/wtaYSuScm3R9wAAAP//X0LFgj8CAAA=\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/flatcar/update.conf\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"\",\n \"source\": \"data:,REBOOT_STRATEGY%3Doff%0A\",\n \"verification\": {}\n },\n \"mode\": 272\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/kubeadm\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubeadm\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/kubelet\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubelet\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/kubectl\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubectl\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/calicoctl\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/calicoctl\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/velero\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/velero\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/rsyslogd\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/rsyslogd\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/etc/systemd/journald.conf.d/10-forward-to-syslog.conf\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"\",\n \"source\": \"data:,%5BJournal%5D%0AForwardToSyslog%3Dyes%0ASyslog%3Dyes%0A\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"path\": \"/etc/rsyslog.conf\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/0yQwW4UMRBE7/6KlnPZPTCzkcglkk9BnAAhdn/AsXt3zNjuod3ewXw9cmZEOJbLelVdD/A15JBsBC6tRLo9A6PDcEe4MiX4SZWzjR42Fwq5GQVs9nAlXi17EALGRIJK3SK92nhYiedPgdEJcTN6vFseY3gd9wh9VOoBLlMoEArIhBDSQiw2CyyW5RmW4GaoC6yTlfcKe2CBdcIMnzd1ofMb1DQsKpGvEQ+RrDc6pPq719VwbuVMbh6+2YRGj1zzWFoRTH7c4eP/zXYy4B25yRTyrd/YqPK/FZDvyHC4vHzvc1FKmD36o7JOAuWDApC2oNGUrqvXXVq+oRi9AYaaPbKLVP0QydnYv/QJjH56/PgmmIQcRaPFLf1hIw+MpSb8gcLthWoWoz88dvtXxYrDFhpDntF/CUXenRL+oNFPp9PppNVR/Q0AAP//7zzZyfcBAAA=\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"path\": \"/etc/kubernetes/kubeadm-init.yaml\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/6xV3W7jNhO951MQvNmbUD+WYie8y7dZfFgsWgSb/gFFLyhyJBOiSXVIOfY+fUHJNqKugwZFbcCQoTlHhzPnjORgfgEMxjtB+7EBqXdZfxcy4/N92UCUFemN04J+diZ+9K413YgyGu9I430MEeXwk+/BBUEo5TSma0FZv+/KrcpihfDi9ziEerVG3zJCKaUagkIzxOmpTEMrRxvPj6cX3plshsRoBWUFI85r+AqdSRUTAaHUyR0IqryL6C0frHRQEkoVmmeveoiCjs4cRJ7nOLo81UnjAPWryyx41RM6ibAQPx0iygfsplNRmh7KzSAoWxVFKepNITZluRYrIe7LWaCyY4iAXLsgKCuLrMjuy+y2ulkiWlEKcVvNmL234w74YMfOOK4NCspyP8TcmgYOoPKkBh1ECPlJ2Kk45DM2n8oYsV5J+/D0+ZPTgzcuJtlS7wGjCfCgNUIIb6tvjNNPHqOg67quiEqw1igZ4QscBWUtwEZVcKtaJdUG6s2mVmtolSzqdS3rlb6737RV0W6qoqjqQq9vN3fNrWqbWla6qhtGOOfk3Vb7OLdy6bbTdJ/ScC+npOz3VheFSN+VEGVR/CHSERhxEF889sZ1qRWD189j45IT/tYDKUS+rm+meaWffFWnjgTAvVHwBmgaYl4WdzfnOZ9x2oVHv5NmisBdyEanAZX1o86mEbHzOSzgD9LJDjDpg6XdWgsH/t+ZI1GmEXN5NsIHIT6keTwD7q8qACcbC1x2HUIno0eOfoypm5RFHGEmHdAfjlxZAy7y5BreGguC5hAX8obe5C16F/lrRKYwfk/Tw/HE8l6aHo4TDcKfI4S4BakBL6rkezXJi54lkbTWv4DmackEQb8XcAUzdZPPfwIfEFpzEPQ3/hV2PgKftgu/guvQj8MZJ+j8ucD+n+5eQY0BMKlbAC+onwMguRpACzFTU8iWOSxPOfwy1yxzePEQE4KRLUgbt9/+Z5x+WNw4rcN/CsOp6sfn+e3B3lyW/PVOZWQ2+NNk98cUjX+RDKKmdj+iSRmgLBxDhJ1mRI5xCy6mBXh6wUjn3XHnx0U6tKApCoTSF2i23vfXbiYuj+bbhWrnNQj664wgfwUAAP//F9AABn8HAAA=\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"path\": \"/etc/kubernetes/addons/kube-dns-fixed-svc.yaml\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/4yRwU7rMBBF9/6K+YCXKn6Uls62FVJ3FgU2iMXUGZAVJzb2JFL+HoWmEBYV9fLOuUejMUX3zCm70CL0WtWurRAOnHpnWTUsVJEQKoCWGkaouyMXVZunIEey5zQPWbhRAJ6O7PPYAajvckExzoo5sh1nMkRG2PouC6e9UQAu3lPj/GCCd3ZAeOCPziXedeQPQrb+QRxnhJe96Vf/YG/65asCsGcTwv+y1Lhcl7jWeoVvqBFvb+bItFxxkRxnulyUi41efAUxJPlunU5xusL4xiHCVASIKUiwwSM87cyUCaV3FjMHZ55CbPzL9bi9wtWwJGd/77XRV9smNLNnKyFd+sLPAAAA//8MLaPANAIAAA==\",\n \"verification\": {}\n },\n \"mode\": 420\n }\n ]\n },\n \"systemd\": {\n \"units\": [\n {\n \"contents\": \"[Unit]\\nDescription=rsyslog (journald -\\u003e remote syslog)\\nWants=network-online.target\\nAfter=network-online.target systemd-journald.service\\nRequires=systemd-journald.service\\n\\n[Service]\\nType=simple\\nExecStart=/opt/bin/rsyslogd -n -f /etc/rsyslog.conf\\nRestart=always\\nRestartSec=2\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"rsyslog.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Undercloud Bootstrap Chain\\nWants=network-online.target\\nAfter=network-online.target\\n# Start the chain entry\\nWants=containerd.service kubelet.service kubeadm-init.service\\nAfter=containerd.service kubelet.service kubeadm-init.service\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"undercloud-bootstrap.target\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Load necessary kernel modules\\nBefore=containerd.service kubeadm-init.service\\n\\n[Service]\\nType=oneshot\\nExecStart=/usr/bin/modprobe br_netfilter\\nExecStart=/usr/bin/modprobe overlay\\nRemainAfterExit=yes\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"modules-load.service\"\n },\n {\n \"enabled\": true,\n \"name\": \"systemd-networkd-wait-online.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=containerd container runtime\\nAfter=network.target modules-load.service\\nWants=modules-load.service\\n\\n[Service]\\nExecStart=/usr/bin/containerd\\nRestart=always\\nRestartSec=5\\nDelegate=yes\\nKillMode=process\\nOOMScoreAdjust=-999\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"containerd.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Set Timezone\\nAfter=network-online.target\\nWants=network-online.target\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nExecStart=/bin/sh -c 'echo \\\"setting timezone to Europe/Berlin\\\"'\\nExecStart=/usr/bin/timedatectl set-timezone Europe/Berlin\\nExecStart=/usr/bin/timedatectl set-ntp true\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"set-timezone.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=kubelet, the Kubernetes Node Agent\\nDocumentation=https://kubernetes.io/docs/home\\nWants=network-online.target\\nAfter=network-online.target containerd.service\\nRequires=containerd.service\\n\\n[Service]\\nEnvironment=\\\"KUBELET_KUBECONFIG_ARGS=--bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf\\\"\\nEnvironment=\\\"KUBELET_CONFIG_ARGS=--config=/var/lib/kubelet/config.yaml\\\"\\nEnvironmentFile=-/var/lib/kubelet/kubeadm-flags.env\\nExecStart=/opt/bin/kubelet $KUBELET_KUBECONFIG_ARGS $KUBELET_CONFIG_ARGS $KUBELET_KUBEADM_ARGS $KUBELET_EXTRA_ARGS\\nRestart=always\\nStartLimitInterval=0\\nRestartSec=10\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"kubelet.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Kubeadm Init Cluster\\nWants=network-online.target\\nAfter=network-online.target containerd.service kubelet.service\\nRequires=containerd.service kubelet.service\\nConditionPathExists=!/etc/kubernetes/kubelet.conf\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\n\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin:/opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent-uds/\\n\\nExecStart=/bin/sh -eu -c '\\\\\\n echo \\\"[kubeadm-init] started...\\\" ; \\\\\\n echo \\\"[kubeadm-init] waiting for containerd socket...\\\" ; \\\\\\n for i in $(seq 1 60); do test -S /run/containerd/containerd.sock \\u0026\\u0026 break; sleep 1; done ; \\\\\\n echo \\\"[kubeadm-init] running kubeadm init...\\\" ; \\\\\\n /opt/bin/kubeadm init --upload-certs --config=/etc/kubernetes/kubeadm-init.yaml ; \\\\\\n echo \\\"[kubeadm-init] copying kubeconfig to core...\\\" ; \\\\\\n mkdir -p /home/core/.kube ; \\\\\\n cp -f /etc/kubernetes/admin.conf /home/core/.kube/config ; \\\\\\n chown core:core /home/core/.kube/config ; \\\\\\n echo \\\"[kubeadm-init] done.\\\" \\\\\\n'\\n\\n# strictly start next step (serialization)\\nExecStartPost=/usr/bin/systemctl start install-calico.service\\n\\n[Install]\\nWantedBy=undercloud-bootstrap.target\\n\",\n \"enabled\": false,\n \"name\": \"kubeadm-init.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Install Calico\\nRequires=kubeadm-init.service\\nAfter=kubeadm-init.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-calico.done\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\n\\nExecStart=/bin/sh -eu -c '\\\\\\n echo \\\"[calico] waiting for API /readyz...\\\" ; \\\\\\n for i in $(seq 1 180); do kubectl get --raw=/readyz \\u003e/dev/null 2\\u003e\\u00261 \\u0026\\u0026 break; sleep 2; done ; \\\\\\n echo \\\"[calico] create namespace + operator...\\\" ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/namespace.yaml ; \\\\\\n kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/operator-crds.yaml || true ; \\\\\\n kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/tigera-operator.yaml || true ; \\\\\\n echo \\\"[calico] wait for tigera-operator...\\\" ; \\\\\\n kubectl wait deployment -n tigera-operator tigera-operator --for condition=Available=True --timeout=1200s ; \\\\\\n echo \\\"[calico] apply custom resources...\\\" ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/custom-resources.yaml ; \\\\\\n echo \\\"[calico] wait for calico-apiserver...\\\" ; \\\\\\n kubectl wait deployment -n calico-apiserver calico-apiserver --for condition=Available=True --timeout=1200s ; \\\\\\n echo \\\"[calico] apply peers + pools...\\\" ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/calico-peer.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/ippools.yaml ; \\\\\\n echo \\\"[calico] done.\\\" \\\\\\n'\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-calico.done\\nExecStartPost=/usr/bin/systemctl start install-ceph.service\\n\\n[Install]\\nWantedBy=undercloud-bootstrap.target\\n\",\n \"enabled\": false,\n \"name\": \"install-calico.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Install Ceph CSI\\nRequires=install-calico.service\\nAfter=install-calico.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-ceph.done\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\n\\nExecStart=/bin/sh -eu -c '\\\\\\n echo \\\"[ceph] waiting for API /readyz...\\\" ; \\\\\\n for i in $(seq 1 180); do kubectl get --raw=/readyz \\u003e/dev/null 2\\u003e\\u00261 \\u0026\\u0026 break; sleep 2; done ; \\\\\\n echo \\\"[ceph] apply manifests...\\\" ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/namespace.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-nodeplugin-rbac.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-provisioner-rbac.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/secrets.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/ceph-conf.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-cephfsplugin-provisioner.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-cephfsplugin.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-config-map.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-encryption-kms-config.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csidriver.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/storage-class.yaml ; \\\\\\n echo \\\"[ceph] done.\\\" \\\\\\n'\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-ceph.done\\nExecStartPost=/usr/bin/systemctl start install-gitea.service\\n\\n[Install]\\nWantedBy=undercloud-bootstrap.target\\n\",\n \"enabled\": false,\n \"name\": \"install-ceph.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Install Gitea\\nRequires=install-ceph.service\\nAfter=install-ceph.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-gitea.done\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\n\\nExecStart=/bin/sh -eu -c '\\\\\\n echo \\\"[gitea] wait for ceph provisioner...\\\" ; \\\\\\n kubectl wait deployment -n ceph csi-cephfsplugin-provisioner --for condition=Available=True --timeout=1200s ; \\\\\\n echo \\\"[gitea] apply manifests...\\\" ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/namespace.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/secrets.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/db.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/adminer.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/gitea.yaml ; \\\\\\n echo \\\"[gitea] wait for gitea deployment...\\\" ; \\\\\\n kubectl -n gitea wait deployment gitea --for=condition=Available=True --timeout=1200s ; \\\\\\n echo \\\"[gitea] run startup...\\\" ; \\\\\\n kubectl exec deploy/gitea -n gitea -- /bin/startup.sh ; \\\\\\n echo \\\"[gitea] done.\\\" \\\\\\n'\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-gitea.done\\nExecStartPost=/usr/bin/systemctl start install-argocd.service\\n\\n[Install]\\nWantedBy=undercloud-bootstrap.target\\n\",\n \"enabled\": false,\n \"name\": \"install-gitea.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Install ArgoCD\\nRequires=install-calico.service install-gitea.service\\nAfter=install-calico.service install-gitea.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-argocd.done\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\n\\nExecStart=/bin/sh -eu -c '\\\\\\n echo \\\"[argocd] wait for coredns...\\\" ; \\\\\\n kubectl -n kube-system wait deploy coredns --for=condition=Available=True --timeout=1200s ; \\\\\\n echo \\\"[argocd] install...\\\" ; \\\\\\n kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/namespace.yaml ; \\\\\\n kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/install.yaml ; \\\\\\n kubectl -n argocd wait deploy argocd-server --for=condition=Available=True --timeout=1200s ; \\\\\\n kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/repo.yaml ; \\\\\\n kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/apps.yaml ; \\\\\\n kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/ingress.yaml ; \\\\\\n echo \\\"[argocd] done.\\\" \\\\\\n'\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-argocd.done\\nExecStartPost=/usr/bin/systemctl start pin-service-ips.service\\n\\n[Install]\\nWantedBy=undercloud-bootstrap.target\\n\",\n \"enabled\": false,\n \"name\": \"install-argocd.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Pin fixed dual-stack ClusterIPs for kube-dns, argocd-server and whisker\\nRequires=install-argocd.service install-calico.service kubeadm-init.service\\nAfter=install-argocd.service install-calico.service kubeadm-init.service network-online.target\\nWants=network-online.target\\nConditionPathExists=!/var/lib/undercloud-stamps/pin-service-ips.done\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=PATH=/usr/bin:/usr/sbin:/opt/bin\\nExecStart=/bin/sh -eu -c '\\\\\\n echo \\\"[pin-service-ips] waiting for API...\\\" ; \\\\\\n for i in $(seq 1 120); do kubectl get --raw=/readyz \\u003e/dev/null 2\\u003e\\u00261 \\u0026\\u0026 break; sleep 2; done ; \\\\\\n echo \\\"[pin-service-ips] ensure namespaces exist...\\\" ; \\\\\\n kubectl get ns kube-system \\u003e/dev/null ; \\\\\\n kubectl get ns argocd \\u003e/dev/null 2\\u003e\\u00261 || kubectl create ns argocd ; \\\\\\n kubectl get ns calico-system \\u003e/dev/null ; \\\\\\n echo \\\"[pin-service-ips] wait for coredns/argocd readiness (best effort)...\\\" ; \\\\\\n kubectl -n kube-system wait deploy coredns --for=condition=Available=True --timeout=300s || true ; \\\\\\n kubectl -n argocd wait deploy argocd-server --for=condition=Available=True --timeout=600s || true ; \\\\\\n echo \\\"[pin-service-ips] replace Services with fixed ClusterIPs...\\\" ; \\\\\\n kubectl -n kube-system delete svc kube-dns --ignore-not-found ; \\\\\\n kubectl apply -f /etc/kubernetes/addons/kube-dns-fixed-svc.yaml ; \\\\\\n kubectl -n argocd delete svc argocd-server --ignore-not-found ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/service.yaml ; \\\\\\n kubectl -n calico-system delete svc whisker --ignore-not-found || true ; \\\\\\n kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/whisker.yaml || true ; \\\\\\n echo \\\"[pin-service-ips] done.\\\" \\\\\\n'\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/pin-service-ips.done\\n\\n[Install]\\nWantedBy=undercloud-bootstrap.target\\n\",\n \"enabled\": false,\n \"name\": \"pin-service-ips.service\"\n }\n ]\n }\n}", + "rendered": "{\n \"ignition\": {\n \"config\": {\n \"replace\": {\n \"verification\": {}\n }\n },\n \"proxy\": {},\n \"security\": {\n \"tls\": {}\n },\n \"timeouts\": {},\n \"version\": \"3.4.0\"\n },\n \"kernelArguments\": {},\n \"passwd\": {\n \"users\": [\n {\n \"name\": \"core\",\n \"sshAuthorizedKeys\": [\n \"ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIHHEAlPo3v4U67Y3411pTjIMkQxwlFWdXrBJkSzXenDH flatcar@undercloud\"\n ]\n }\n ]\n },\n \"storage\": {\n \"directories\": [\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/opt/bin\",\n \"user\": {},\n \"mode\": 493\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/opt/cni/bin\",\n \"user\": {},\n \"mode\": 755\n },\n {\n \"group\": {},\n \"path\": \"/etc/kubernetes/manifests\",\n \"user\": {},\n \"mode\": 493\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/install-calico\",\n \"user\": {},\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/var/lib/undercloud-stamps\",\n \"user\": {},\n \"mode\": 493\n }\n ],\n \"files\": [\n {\n \"group\": {},\n \"path\": \"/etc/hostname\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"\",\n \"source\": \"data:,control-plane1%0A\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"path\": \"/etc/systemd/network/00-eth.network\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/2SMMWuFMBRG9/yUDjE3FcVABsFSOlSkHcUhJLco1aQkUeu/f+h76PC407mH77SfKuq+I7WaUGLsXwhpa4yr878dKY3xGIL8MYyJ/bgQBSRZehrOGIg0ZyIHyC79riKuanvWr6Sqv68eCAFnCxhllNMCEn4VHs/77gCgQCo3qcEGOVuDXo9uNnR0Wo3ko1myUmv8i1+ltO7gxg+L0tvbf0QbBmfDLm4BAAD//4CQfvX4AAAA\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/hosts\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/5ST7W6DIBRA//MUvEDJ/Vhqdt/GqcuWEjGszV5/wVEjlBa5/sLLOYknEakzYMCg1tq6obdf7ueqRML5f7a3+ns5n/KTWz764aIUAaC8dSAd4llI5B21Htx89c6eFtvPE5rbPE5+sO42mlWT7QsOyhxUcVDBwZmDKw5+dCCg/nX+MvnCR8RFiaK4pGcUlSiOS35GsVKfI4CEJ5YO01g7UVBBUY+dKLigqLfeK0LmMNXUKUQ76EXpFOId9Co0hp+DzNq4MfAdpfawd5Tbg0Z0bVkNuV2mAwG3y3wg3F8AAAD//+yy88FYBAAA\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/motd\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/6yRwYobMQyG7/MUOmYX4lzL3sJ2oYV2A01yaG+KrMyIeOxBkodOn74kk0MPaSg0/8WWkD4+4+f/T/MMf8l6v/u0+fb5x9tHWL++vm23sHn/8v3G4B3GP+fC2HViYJM59yAGA6pDOQKCMVXlCJKPiuZayatyuMVYpwRILqO4sAEqQ1+yeDnvY46QSttyDPc89hmrd0Xl13mHiM2gKPRi1fhsZq5CniYYtHRyEL+y/2D0OIGy1eQgGaIYyZAko06zBbc4i5Ycbns84G+b5YPTnOqBNbOzAZXsWhLDkDAzvJfITfMVM7YMo+DLZZY8weJ6eWoIk1CBxXxeWiMn1gJLOCCd6gCLufHUoLaFInTug72sVnO5NNaRNcxVsJHC6QPWn6HmyEqp1BjouHr8u38HAAD//04R3qpqAwAA\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"path\": \"/etc/sysctl.d/99-k8s.conf\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/4zQX27DIAwG8PeeIhcIyqos6h56FuQS01kFGxmWtrefsmXZHzE13wsS+iH7g7EYSlNvKFkvegUdm785Nk+7xQ2bnRP2BkIwiyY+V1y/xZ2UxjMuR8u+dRBCS6nAKWB+7IYFfjnG4ikUVMPeOmEuCu5iI9zWud1HKltqsp9vf7TuKq1HyvNQO99+uykamVCdxEjFRoyi99ov+myIpZC/mwg3+5ZR7RWKe53rru553+8PhyomzgXYrfzY9N3LsLugMgaTgMk1/2Ru/wtaYSuScm3R9wAAAP//X0LFgj8CAAA=\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/flatcar/update.conf\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"\",\n \"source\": \"data:,REBOOT_STRATEGY%3Doff%0A\",\n \"verification\": {}\n },\n \"mode\": 272\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/kubeadm\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubeadm\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/kubelet\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubelet\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/kubectl\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubectl\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/calicoctl\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/calicoctl\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/velero\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/velero\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/etc/kubernetes/kubeadm-init.yaml\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/6xV3W7jNhO951MQvNmbUD+WYie8y7dZfFgsWgSb/gFFLyhyJBOiSXVIOfY+fUHJNqKugwZFbcCQoTlHhzPnjORgfgEMxjtB+7EBqXdZfxcy4/N92UCUFemN04J+diZ+9K413YgyGu9I430MEeXwk+/BBUEo5TSma0FZv+/KrcpihfDi9ziEerVG3zJCKaUagkIzxOmpTEMrRxvPj6cX3plshsRoBWUFI85r+AqdSRUTAaHUyR0IqryL6C0frHRQEkoVmmeveoiCjs4cRJ7nOLo81UnjAPWryyx41RM6ibAQPx0iygfsplNRmh7KzSAoWxVFKepNITZluRYrIe7LWaCyY4iAXLsgKCuLrMjuy+y2ulkiWlEKcVvNmL234w74YMfOOK4NCspyP8TcmgYOoPKkBh1ECPlJ2Kk45DM2n8oYsV5J+/D0+ZPTgzcuJtlS7wGjCfCgNUIIb6tvjNNPHqOg67quiEqw1igZ4QscBWUtwEZVcKtaJdUG6s2mVmtolSzqdS3rlb6737RV0W6qoqjqQq9vN3fNrWqbWla6qhtGOOfk3Vb7OLdy6bbTdJ/ScC+npOz3VheFSN+VEGVR/CHSERhxEF889sZ1qRWD189j45IT/tYDKUS+rm+meaWffFWnjgTAvVHwBmgaYl4WdzfnOZ9x2oVHv5NmisBdyEanAZX1o86mEbHzOSzgD9LJDjDpg6XdWgsH/t+ZI1GmEXN5NsIHIT6keTwD7q8qACcbC1x2HUIno0eOfoypm5RFHGEmHdAfjlxZAy7y5BreGguC5hAX8obe5C16F/lrRKYwfk/Tw/HE8l6aHo4TDcKfI4S4BakBL6rkezXJi54lkbTWv4DmackEQb8XcAUzdZPPfwIfEFpzEPQ3/hV2PgKftgu/guvQj8MZJ+j8ucD+n+5eQY0BMKlbAC+onwMguRpACzFTU8iWOSxPOfwy1yxzePEQE4KRLUgbt9/+Z5x+WNw4rcN/CsOp6sfn+e3B3lyW/PVOZWQ2+NNk98cUjX+RDKKmdj+iSRmgLBxDhJ1mRI5xCy6mBXh6wUjn3XHnx0U6tKApCoTSF2i23vfXbiYuj+bbhWrnNQj664wgfwUAAP//F9AABn8HAAA=\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"path\": \"/etc/kubernetes/addons/kube-dns-fixed-svc.yaml\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/4yRwU7rMBBF9/6K+YCXKn6Uls62FVJ3FgU2iMXUGZAVJzb2JFL+HoWmEBYV9fLOuUejMUX3zCm70CL0WtWurRAOnHpnWTUsVJEQKoCWGkaouyMXVZunIEey5zQPWbhRAJ6O7PPYAajvckExzoo5sh1nMkRG2PouC6e9UQAu3lPj/GCCd3ZAeOCPziXedeQPQrb+QRxnhJe96Vf/YG/65asCsGcTwv+y1Lhcl7jWeoVvqBFvb+bItFxxkRxnulyUi41efAUxJPlunU5xusL4xiHCVASIKUiwwSM87cyUCaV3FjMHZ55CbPzL9bi9wtWwJGd/77XRV9smNLNnKyFd+sLPAAAA//8MLaPANAIAAA==\",\n \"verification\": {}\n },\n \"mode\": 420\n }\n ]\n },\n \"systemd\": {\n \"units\": [\n {\n \"contents\": \"[Unit]\\nDescription=Load necessary kernel modules\\nBefore=containerd.service kubeadm-init.service\\n\\n[Service]\\nType=oneshot\\n\\nExecStart=/usr/bin/modprobe br_netfilter\\nExecStart=/usr/bin/modprobe overlay\\nRemainAfterExit=yes\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"modules-load.service\"\n },\n {\n \"enabled\": true,\n \"name\": \"systemd-networkd-wait-online.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=containerd container runtime\\nAfter=network.target modules-load.service\\n\\n[Service]\\n#StandardOutput=journal+console\\n#StandardError=journal+console\\n\\nExecStart=/usr/bin/containerd\\nRestart=always\\nRestartSec=5\\nDelegate=yes\\nKillMode=process\\nOOMScoreAdjust=-999\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"containerd.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Set Timezone\\nAfter=network-online.target\\nWants=network-online.target\\n[Service]\\nStandardOutput=journal+console\\nStandardError=journal+console\\n\\nExecStart=/bin/sh -c 'echo \\\"setting timezone to Europe/Berlin\\\"'\\nStandardOutput=journal+console\\nStandardError=journal+console\\nType=oneshot\\nRestart=on-failure\\nExecStart=/usr/bin/timedatectl set-timezone Europe/Berlin\\nExecStart=/usr/bin/timedatectl set-ntp true \\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"set-timezone.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=kubelet, the Kubernetes Node Agent\\nDocumentation=https://kubernets.io/docs/home\\nWants=network-online.target\\nAfter=network-online.target\\n[Service]\\n#StandardOutput=journal+console\\n#StandardError=journal+console\\n\\n#EnvironmentFile=/run/metadata/coreos\\nEnvironment=\\\"KUBELET_KUBECONFIG_ARGS=--bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf\\\"\\nEnvironment=\\\"KUBELET_CONFIG_ARGS=--config=/var/lib/kubelet/config.yaml\\\"\\n# This is a file that \\\"kubeadm init\\\" and \\\"kubeadm join\\\" generates at runtime, populating the KUBELET_KUBEADM_ARGS variable dynamically\\nEnvironmentFile=-/var/lib/kubelet/kubeadm-flags.env\\nExecStart=/opt/bin/kubelet $KUBELET_KUBECONFIG_ARGS $KUBELET_CONFIG_ARGS $KUBELET_KUBEADM_ARGS $KUBELET_EXTRA_ARGS\\nRestart=always\\nStartLimitInterval=0\\nRestartSec=10\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"kubelet.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Kubeadm Init Cluster\\nAfter=network-online.target containerd.service kubelet.service\\nWants=network-online.target\\nConditionPathExists=!/etc/kubernetes/kubelet.conf\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nExecStart=/opt/bin/kubeadm reset -f\\n\\nExecStart=/bin/sh -c 'echo \\\"kubeadm-init.service started...\\\"'\\n\\n# Environment\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin:/opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent-uds/\\n\\nExecStartPre=/bin/sleep 30s\\nExecStart=/bin/sh -c 'echo \\\"running kubeadm init...\\\"'\\nExecStart=/opt/bin/kubeadm init --upload-certs --config=/etc/kubernetes/kubeadm-init.yaml\\n\\n# copy files for kubectl\\nExecStart=/bin/sh -c 'echo \\\"copying files (admin.conf) to core home folder.\\\"'\\nExecStartPost=/usr/bin/mkdir -p /home/core/.kube\\nExecStartPost=/usr/bin/cp -i /etc/kubernetes/admin.conf /home/core/.kube/config\\nExecStartPost=/usr/bin/chown core:core /home/core/.kube/config\\n\\nExecStartPost=/usr/bin/systemctl disable kubeadm-init.service\\nRestart=on-failure\\nRestartSec=120s\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"kubeadm-init.service\"\n },\n {\n \"contents\": \"[Unit]\\nWants=kubeadm-init.service\\nAfter=kubeadm-init.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-calico.done\\n\\n\\n[Service]\\nStandardOutput=journal+console\\nStandardError=journal+console\\n\\nExecStart=/bin/sh -c 'echo \\\"install.calico.service started...\\\"'\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nExecStart=/bin/sh -c 'echo \\\"witing 30s...\\\"'\\nExecStart=/bin/sleep 30s\\nExecStart=/bin/sh -c 'echo \\\"create calico namespace...\\\"'\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/namespace.yaml\\nExecStart=/bin/sh -c 'echo \\\"install tigera operator...\\\"'\\nExecStart=-/opt/bin/kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/operator-crds.yaml\\nExecStart=-/opt/bin/kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/tigera-operator.yaml\\nExecStart=/bin/sh -c 'echo \\\"witing 60s...\\\"'\\nExecStart=/bin/sleep 60s\\nExecStart=/bin/sh -c 'echo \\\"witing for tigera operator... (20mini max)\\\"'\\nExecStart=/opt/bin/kubectl wait deployment -n tigera-operator tigera-operator --for condition=Available=True --timeout=1200s\\nExecStart=/bin/sh -c 'echo \\\"create clico custom ressources...\\\"'\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/custom-resources.yaml\\n\\nExecStart=/bin/sh -c 'echo \\\"witing 3m..\\\"'\\nExecStart=/bin/sleep 3m\\n#ExecStart=/bin/sh -c 'echo \\\"apply calico (calico-apiserver)...\\\"'\\n#ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/calico.yaml\\n#ExecStart=/bin/sh -c 'echo \\\"witing 1m...\\\"'\\n#ExecStart=/bin/sleep 2m\\nExecStart=/bin/sh -c 'echo \\\"witing calico-apiserver... (20mini max)\\\"'\\nExecStart=/opt/bin/kubectl wait deployment -n calico-apiserver calico-apiserver --for condition=Available=True --timeout=1200s\\nExecStart=/bin/sh -c 'echo \\\"witing 120s...\\\"'\\nExecStart=/bin/sleep 2m\\nExecStart=/bin/sh -c 'echo \\\"apply calico-peers...\\\"'\\nExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/calico-peer.yaml\\nExecStart=/bin/sh -c 'echo \\\"witing 60s...\\\"'\\nExecStart=/bin/sleep 1m\\nExecStart=/bin/sh -c 'echo \\\"apply calico-ippools...\\\"'\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/ippools.yaml\\n\\n#ExecStart=/bin/sh -c 'echo \\\"witing for whisker..\\\"'\\n#ExecStart=/opt/bin/kubectl wait deployment -n calico-system whisker --for condition=Available=True --timeout=1200s\\n#ExecStart=/bin/sh -c 'echo \\\"port-forward -n calico-system service/whisker 8081:8081\\\"'\\n#ExecStart=/opt/bin/kubectl port-forward -n calico-system service/whisker 8081:8081\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-calico.done\\nExecStart=/usr/bin/systemctl disable install-calico.service\\n#RemainAfterExit=true\\nRestart=on-failure\\nRestartSec=120s\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"install-calico.service\"\n },\n {\n \"contents\": \"[Unit]\\nWants=kubeadm-init.service\\nAfter=kubeadm-init.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-ceph.done\\n\\n[Service]\\nStandardOutput=journal+console\\nStandardError=journal+console\\n\\nExecStart=/bin/sh -c 'echo \\\"install.ceph.service started...\\\"'\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\nType=oneshot\\n\\nStandardOutput=journal+console\\nStandardError=journal+console\\nExecStart=/bin/sh -c 'echo \\\"witing 30s...\\\"'\\nExecStart=/bin/sleep 30s\\nExecStart=/bin/sh -c 'echo \\\"create ceph namespace...\\\"'\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/namespace.yaml\\n\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-nodeplugin-rbac.yaml\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-provisioner-rbac.yaml\\n\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/secrets.yaml\\n\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/ceph-conf.yaml\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-cephfsplugin-provisioner.yaml\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-cephfsplugin.yaml\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-config-map.yaml\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-encryption-kms-config.yaml\\n\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csidriver.yaml\\n\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/storage-class.yaml\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-ceph.done\\nExecStart=/usr/bin/systemctl disable install-ceph.service\\n#RemainAfterExit=true\\nRestart=on-failure\\nRestartSec=120s\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"install-ceph.service\"\n },\n {\n \"contents\": \"[Unit]\\nWants=install-ceph.service\\nAfter=install-ceph.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-gitea.done\\n\\n[Service]\\nStandardOutput=journal+console\\nStandardError=journal+console\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\nType=oneshot\\n\\nExecStart=/opt/bin/kubectl wait deployment -n ceph csi-cephfsplugin-provisioner --for condition=Available=True --timeout=1200s\\nExecStart=/bin/sleep 4m\\nExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/namespace.yaml\\nExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/secrets.yaml\\nExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/db.yaml\\nExecStart=/bin/sleep 60s\\nExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/adminer.yaml\\nExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/gitea.yaml\\nExecStart=/bin/sleep 3m\\nExecStart=/opt/bin/kubectl exec deploy/gitea -n gitea -- /bin/startup.sh\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-gitea.done\\nExecStart=/usr/bin/systemctl disable install-gitea.service\\nRestart=on-failure\\nRestartSec=120s\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"install-gitea.service\"\n },\n {\n \"contents\": \"[Unit]\\nWants=install-calico.service\\nAfter=install-calico.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-argocd.done\\n[Service]\\nStandardOutput=journal+console\\nStandardError=journal+console\\n\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\nType=oneshot\\n\\nExecStart=/opt/bin/kubectl wait deployment -n kube-system coredns --for condition=Available=True --timeout=600s\\n\\nExecStart=/bin/sleep 1m\\nExecStart=/opt/bin/kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/namespace.yaml\\nExecStart=/opt/bin/kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/install.yaml\\nExecStart=/opt/bin/kubectl wait deployment -n argocd argocd-server --for condition=Available=True --timeout=600s\\n\\nExecStart=/opt/bin/kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/repo.yaml\\nExecStart=/bin/sleep 10s\\nExecStart=/opt/bin/kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/apps.yaml\\n\\nExecStart=/bin/sleep 10s\\nExecStart=/opt/bin/kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/ingress.yaml\\n\\n\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/repos/k8aux-bootstrap.yaml\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/repos/k8aux-apps.yaml\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/apps/argocd.yaml\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/apps/calico.yaml\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/apps/rook-ceph.yaml\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/apps/gitea.yaml\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/argocd/argocd-secret.yaml\\n##ExecStart=/bin/sleep 10m\\n#ExecStart=/opt/bin/kubectl wait deployment -n gitea gitea --for condition=Available=True --timeout=4800s\\n#ExecStart=/bin/sleep 10m\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://gitea.gitea.svc.k8aux.undercloud.cf:3000/undercloud/k8aux-apps/raw/branch/main/app-of-apps/app-of-apps.yaml\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-argocd.done\\nExecStart=/usr/bin/systemctl disable install-argocd.service\\nRestart=on-failure\\nRestartSec=120s\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"install-argocd.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Pin fixed dual-stack ClusterIPs for kube-dns, argocd-server and whisker\\nAfter=install-argocd.service install-calico.service kubeadm-init.service network-online.target\\nWants=install-argocd.service install-calico.service kubeadm-init.service network-online.target\\nConditionPathExists=!/var/lib/undercloud-stamps/pin-service-ips.done\\n\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=PATH=/usr/bin:/usr/sbin:/opt/bin\\nExecStart=/bin/sh -eu -c '\\\\\\n echo \\\"[pin-service-ips] waiting for API...\\\" ; \\\\\\n for i in $(seq 1 120); do kubectl get --raw=/readyz \\u003e/dev/null 2\\u003e\\u00261 \\u0026\\u0026 break; sleep 2; done ; \\\\\\n echo \\\"[pin-service-ips] ensure namespaces exist...\\\" ; \\\\\\n kubectl get ns kube-system \\u003e/dev/null ; \\\\\\n kubectl get ns argocd \\u003e/dev/null 2\\u003e\\u00261 || kubectl create ns argocd ; \\\\\\n kubectl get ns calico-system \\u003e/dev/null ; \\\\\\n echo \\\"[pin-service-ips] wait for coredns/argocd readiness (best effort)...\\\" ; \\\\\\n kubectl -n kube-system wait deploy coredns --for=condition=Available=True --timeout=300s || true ; \\\\\\n kubectl -n argocd wait deploy argocd-server --for=condition=Available=True --timeout=600s || true ; \\\\\\n echo \\\"[pin-service-ips] replace Services with fixed ClusterIPs...\\\" ; \\\\\\n kubectl -n kube-system delete svc kube-dns --ignore-not-found ; \\\\\\n kubectl apply -f /etc/kubernetes/addons/kube-dns-fixed-svc.yaml ; \\\\\\n kubectl -n argocd delete svc argocd-server --ignore-not-found ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/service.yaml ; \\\\\\n kubectl -n calico-system delete svc whisker --ignore-not-found || true ; \\\\\\n kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/whisker.yaml || true ; \\\\\\n echo \\\"[pin-service-ips] done.\\\" \\\\\\n'\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/pin-service-ips.done\\nRestart=on-failure\\nRestartSec=120s\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"pin-service-ips.service\"\n }\n ]\n }\n}", "snippets": null, "strict": false }, @@ -152,6 +152,38 @@ } ] }, + { + "mode": "managed", + "type": "null_resource", + "name": "wait_for_cluster_ready", + "provider": "provider[\"registry.terraform.io/hashicorp/null\"]", + "instances": [ + { + "schema_version": 0, + "attributes": { + "id": "5616813553365462819", + "triggers": null + }, + "sensitive_attributes": [], + "identity_schema_version": 0, + "dependencies": [ + "data.ct_config.control_plane1_ignition", + "data.ct_config.control_plane2_ignition", + "data.ct_config.control_plane3_ignition", + "null_resource.wait_for_cp1_api", + "null_resource.wait_for_cp2_cp3_api", + "proxmox_virtual_environment_download_file.flatcar_image", + "proxmox_virtual_environment_file.control_plane1_ignition", + "proxmox_virtual_environment_file.control_plane2_ignition", + "proxmox_virtual_environment_file.control_plane3_ignition", + "proxmox_virtual_environment_vm.control_plane1", + "proxmox_virtual_environment_vm.control_plane2", + "proxmox_virtual_environment_vm.control_plane3", + "proxmox_virtual_environment_vm.flatcar_template" + ] + } + ] + }, { "mode": "managed", "type": "null_resource", @@ -159,10 +191,9 @@ "provider": "provider[\"registry.terraform.io/hashicorp/null\"]", "instances": [ { - "status": "tainted", "schema_version": 0, "attributes": { - "id": "4053921073969850663", + "id": "4626356098463939642", "triggers": null }, "sensitive_attributes": [], @@ -177,6 +208,37 @@ } ] }, + { + "mode": "managed", + "type": "null_resource", + "name": "wait_for_cp2_cp3_api", + "provider": "provider[\"registry.terraform.io/hashicorp/null\"]", + "instances": [ + { + "schema_version": 0, + "attributes": { + "id": "163078617735657231", + "triggers": null + }, + "sensitive_attributes": [], + "identity_schema_version": 0, + "dependencies": [ + "data.ct_config.control_plane1_ignition", + "data.ct_config.control_plane2_ignition", + "data.ct_config.control_plane3_ignition", + "null_resource.wait_for_cp1_api", + "proxmox_virtual_environment_download_file.flatcar_image", + "proxmox_virtual_environment_file.control_plane1_ignition", + "proxmox_virtual_environment_file.control_plane2_ignition", + "proxmox_virtual_environment_file.control_plane3_ignition", + "proxmox_virtual_environment_vm.control_plane1", + "proxmox_virtual_environment_vm.control_plane2", + "proxmox_virtual_environment_vm.control_plane3", + "proxmox_virtual_environment_vm.flatcar_template" + ] + } + ] + }, { "mode": "managed", "type": "proxmox_virtual_environment_download_file", @@ -228,7 +290,7 @@ "source_file": [], "source_raw": [ { - "data": "{\n \"ignition\": {\n \"config\": {\n \"replace\": {\n \"verification\": {}\n }\n },\n \"proxy\": {},\n \"security\": {\n \"tls\": {}\n },\n \"timeouts\": {},\n \"version\": \"3.4.0\"\n },\n \"kernelArguments\": {},\n \"passwd\": {\n \"users\": [\n {\n \"name\": \"core\",\n \"sshAuthorizedKeys\": [\n \"ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIHHEAlPo3v4U67Y3411pTjIMkQxwlFWdXrBJkSzXenDH flatcar@undercloud\"\n ]\n }\n ]\n },\n \"storage\": {\n \"directories\": [\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/opt/bin\",\n \"user\": {},\n \"mode\": 493\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/opt/cni/bin\",\n \"user\": {},\n \"mode\": 755\n },\n {\n \"group\": {},\n \"path\": \"/etc/kubernetes/manifests\",\n \"user\": {},\n \"mode\": 493\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/install-calico\",\n \"user\": {},\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/var/lib/undercloud-stamps\",\n \"user\": {},\n \"mode\": 493\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/var/lib/rsyslog\",\n \"user\": {},\n \"mode\": 493\n }\n ],\n \"files\": [\n {\n \"group\": {},\n \"path\": \"/etc/hostname\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"\",\n \"source\": \"data:,control-plane1%0A\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"path\": \"/etc/systemd/network/00-eth.network\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/2SMMWuFMBRG9/yUDjE3FcVABsFSOlSkHcUhJLco1aQkUeu/f+h76PC407mH77SfKuq+I7WaUGLsXwhpa4yr878dKY3xGIL8MYyJ/bgQBSRZehrOGIg0ZyIHyC79riKuanvWr6Sqv68eCAFnCxhllNMCEn4VHs/77gCgQCo3qcEGOVuDXo9uNnR0Wo3ko1myUmv8i1+ltO7gxg+L0tvbf0QbBmfDLm4BAAD//4CQfvX4AAAA\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/hosts\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/5ST7W6DIBRA//MUvEDJ/Vhqdt/GqcuWEjGszV5/wVEjlBa5/sLLOYknEakzYMCg1tq6obdf7ueqRML5f7a3+ns5n/KTWz764aIUAaC8dSAd4llI5B21Htx89c6eFtvPE5rbPE5+sO42mlWT7QsOyhxUcVDBwZmDKw5+dCCg/nX+MvnCR8RFiaK4pGcUlSiOS35GsVKfI4CEJ5YO01g7UVBBUY+dKLigqLfeK0LmMNXUKUQ76EXpFOId9Co0hp+DzNq4MfAdpfawd5Tbg0Z0bVkNuV2mAwG3y3wg3F8AAAD//+yy88FYBAAA\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/motd\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/6yRwYobMQyG7/MUOmYX4lzL3sJ2oYV2A01yaG+KrMyIeOxBkodOn74kk0MPaSg0/8WWkD4+4+f/T/MMf8l6v/u0+fb5x9tHWL++vm23sHn/8v3G4B3GP+fC2HViYJM59yAGA6pDOQKCMVXlCJKPiuZayatyuMVYpwRILqO4sAEqQ1+yeDnvY46QSttyDPc89hmrd0Xl13mHiM2gKPRi1fhsZq5CniYYtHRyEL+y/2D0OIGy1eQgGaIYyZAko06zBbc4i5Ycbns84G+b5YPTnOqBNbOzAZXsWhLDkDAzvJfITfMVM7YMo+DLZZY8weJ6eWoIk1CBxXxeWiMn1gJLOCCd6gCLufHUoLaFInTug72sVnO5NNaRNcxVsJHC6QPWn6HmyEqp1BjouHr8u38HAAD//04R3qpqAwAA\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"path\": \"/etc/sysctl.d/99-k8s.conf\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/4zQX27DIAwG8PeeIhcIyqos6h56FuQS01kFGxmWtrefsmXZHzE13wsS+iH7g7EYSlNvKFkvegUdm785Nk+7xQ2bnRP2BkIwiyY+V1y/xZ2UxjMuR8u+dRBCS6nAKWB+7IYFfjnG4ikUVMPeOmEuCu5iI9zWud1HKltqsp9vf7TuKq1HyvNQO99+uykamVCdxEjFRoyi99ov+myIpZC/mwg3+5ZR7RWKe53rru553+8PhyomzgXYrfzY9N3LsLugMgaTgMk1/2Ru/wtaYSuScm3R9wAAAP//X0LFgj8CAAA=\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/flatcar/update.conf\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"\",\n \"source\": \"data:,REBOOT_STRATEGY%3Doff%0A\",\n \"verification\": {}\n },\n \"mode\": 272\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/kubeadm\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubeadm\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/kubelet\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubelet\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/kubectl\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubectl\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/calicoctl\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/calicoctl\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/velero\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/velero\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/rsyslogd\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/rsyslogd\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/etc/systemd/journald.conf.d/10-forward-to-syslog.conf\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"\",\n \"source\": \"data:,%5BJournal%5D%0AForwardToSyslog%3Dyes%0ASyslog%3Dyes%0A\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"path\": \"/etc/rsyslog.conf\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/0yQwW4UMRBE7/6KlnPZPTCzkcglkk9BnAAhdn/AsXt3zNjuod3ewXw9cmZEOJbLelVdD/A15JBsBC6tRLo9A6PDcEe4MiX4SZWzjR42Fwq5GQVs9nAlXi17EALGRIJK3SK92nhYiedPgdEJcTN6vFseY3gd9wh9VOoBLlMoEArIhBDSQiw2CyyW5RmW4GaoC6yTlfcKe2CBdcIMnzd1ofMb1DQsKpGvEQ+RrDc6pPq719VwbuVMbh6+2YRGj1zzWFoRTH7c4eP/zXYy4B25yRTyrd/YqPK/FZDvyHC4vHzvc1FKmD36o7JOAuWDApC2oNGUrqvXXVq+oRi9AYaaPbKLVP0QydnYv/QJjH56/PgmmIQcRaPFLf1hIw+MpSb8gcLthWoWoz88dvtXxYrDFhpDntF/CUXenRL+oNFPp9PppNVR/Q0AAP//7zzZyfcBAAA=\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"path\": \"/etc/kubernetes/kubeadm-init.yaml\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/6xV3W7jNhO951MQvNmbUD+WYie8y7dZfFgsWgSb/gFFLyhyJBOiSXVIOfY+fUHJNqKugwZFbcCQoTlHhzPnjORgfgEMxjtB+7EBqXdZfxcy4/N92UCUFemN04J+diZ+9K413YgyGu9I430MEeXwk+/BBUEo5TSma0FZv+/KrcpihfDi9ziEerVG3zJCKaUagkIzxOmpTEMrRxvPj6cX3plshsRoBWUFI85r+AqdSRUTAaHUyR0IqryL6C0frHRQEkoVmmeveoiCjs4cRJ7nOLo81UnjAPWryyx41RM6ibAQPx0iygfsplNRmh7KzSAoWxVFKepNITZluRYrIe7LWaCyY4iAXLsgKCuLrMjuy+y2ulkiWlEKcVvNmL234w74YMfOOK4NCspyP8TcmgYOoPKkBh1ECPlJ2Kk45DM2n8oYsV5J+/D0+ZPTgzcuJtlS7wGjCfCgNUIIb6tvjNNPHqOg67quiEqw1igZ4QscBWUtwEZVcKtaJdUG6s2mVmtolSzqdS3rlb6737RV0W6qoqjqQq9vN3fNrWqbWla6qhtGOOfk3Vb7OLdy6bbTdJ/ScC+npOz3VheFSN+VEGVR/CHSERhxEF889sZ1qRWD189j45IT/tYDKUS+rm+meaWffFWnjgTAvVHwBmgaYl4WdzfnOZ9x2oVHv5NmisBdyEanAZX1o86mEbHzOSzgD9LJDjDpg6XdWgsH/t+ZI1GmEXN5NsIHIT6keTwD7q8qACcbC1x2HUIno0eOfoypm5RFHGEmHdAfjlxZAy7y5BreGguC5hAX8obe5C16F/lrRKYwfk/Tw/HE8l6aHo4TDcKfI4S4BakBL6rkezXJi54lkbTWv4DmackEQb8XcAUzdZPPfwIfEFpzEPQ3/hV2PgKftgu/guvQj8MZJ+j8ucD+n+5eQY0BMKlbAC+onwMguRpACzFTU8iWOSxPOfwy1yxzePEQE4KRLUgbt9/+Z5x+WNw4rcN/CsOp6sfn+e3B3lyW/PVOZWQ2+NNk98cUjX+RDKKmdj+iSRmgLBxDhJ1mRI5xCy6mBXh6wUjn3XHnx0U6tKApCoTSF2i23vfXbiYuj+bbhWrnNQj664wgfwUAAP//F9AABn8HAAA=\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"path\": \"/etc/kubernetes/addons/kube-dns-fixed-svc.yaml\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/4yRwU7rMBBF9/6K+YCXKn6Uls62FVJ3FgU2iMXUGZAVJzb2JFL+HoWmEBYV9fLOuUejMUX3zCm70CL0WtWurRAOnHpnWTUsVJEQKoCWGkaouyMXVZunIEey5zQPWbhRAJ6O7PPYAajvckExzoo5sh1nMkRG2PouC6e9UQAu3lPj/GCCd3ZAeOCPziXedeQPQrb+QRxnhJe96Vf/YG/65asCsGcTwv+y1Lhcl7jWeoVvqBFvb+bItFxxkRxnulyUi41efAUxJPlunU5xusL4xiHCVASIKUiwwSM87cyUCaV3FjMHZ55CbPzL9bi9wtWwJGd/77XRV9smNLNnKyFd+sLPAAAA//8MLaPANAIAAA==\",\n \"verification\": {}\n },\n \"mode\": 420\n }\n ]\n },\n \"systemd\": {\n \"units\": [\n {\n \"contents\": \"[Unit]\\nDescription=rsyslog (journald -\\u003e remote syslog)\\nWants=network-online.target\\nAfter=network-online.target systemd-journald.service\\nRequires=systemd-journald.service\\n\\n[Service]\\nType=simple\\nExecStart=/opt/bin/rsyslogd -n -f /etc/rsyslog.conf\\nRestart=always\\nRestartSec=2\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"rsyslog.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Undercloud Bootstrap Chain\\nWants=network-online.target\\nAfter=network-online.target\\n# Start the chain entry\\nWants=containerd.service kubelet.service kubeadm-init.service\\nAfter=containerd.service kubelet.service kubeadm-init.service\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"undercloud-bootstrap.target\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Load necessary kernel modules\\nBefore=containerd.service kubeadm-init.service\\n\\n[Service]\\nType=oneshot\\nExecStart=/usr/bin/modprobe br_netfilter\\nExecStart=/usr/bin/modprobe overlay\\nRemainAfterExit=yes\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"modules-load.service\"\n },\n {\n \"enabled\": true,\n \"name\": \"systemd-networkd-wait-online.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=containerd container runtime\\nAfter=network.target modules-load.service\\nWants=modules-load.service\\n\\n[Service]\\nExecStart=/usr/bin/containerd\\nRestart=always\\nRestartSec=5\\nDelegate=yes\\nKillMode=process\\nOOMScoreAdjust=-999\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"containerd.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Set Timezone\\nAfter=network-online.target\\nWants=network-online.target\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nExecStart=/bin/sh -c 'echo \\\"setting timezone to Europe/Berlin\\\"'\\nExecStart=/usr/bin/timedatectl set-timezone Europe/Berlin\\nExecStart=/usr/bin/timedatectl set-ntp true\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"set-timezone.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=kubelet, the Kubernetes Node Agent\\nDocumentation=https://kubernetes.io/docs/home\\nWants=network-online.target\\nAfter=network-online.target containerd.service\\nRequires=containerd.service\\n\\n[Service]\\nEnvironment=\\\"KUBELET_KUBECONFIG_ARGS=--bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf\\\"\\nEnvironment=\\\"KUBELET_CONFIG_ARGS=--config=/var/lib/kubelet/config.yaml\\\"\\nEnvironmentFile=-/var/lib/kubelet/kubeadm-flags.env\\nExecStart=/opt/bin/kubelet $KUBELET_KUBECONFIG_ARGS $KUBELET_CONFIG_ARGS $KUBELET_KUBEADM_ARGS $KUBELET_EXTRA_ARGS\\nRestart=always\\nStartLimitInterval=0\\nRestartSec=10\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"kubelet.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Kubeadm Init Cluster\\nWants=network-online.target\\nAfter=network-online.target containerd.service kubelet.service\\nRequires=containerd.service kubelet.service\\nConditionPathExists=!/etc/kubernetes/kubelet.conf\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\n\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin:/opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent-uds/\\n\\nExecStart=/bin/sh -eu -c '\\\\\\n echo \\\"[kubeadm-init] started...\\\" ; \\\\\\n echo \\\"[kubeadm-init] waiting for containerd socket...\\\" ; \\\\\\n for i in $(seq 1 60); do test -S /run/containerd/containerd.sock \\u0026\\u0026 break; sleep 1; done ; \\\\\\n echo \\\"[kubeadm-init] running kubeadm init...\\\" ; \\\\\\n /opt/bin/kubeadm init --upload-certs --config=/etc/kubernetes/kubeadm-init.yaml ; \\\\\\n echo \\\"[kubeadm-init] copying kubeconfig to core...\\\" ; \\\\\\n mkdir -p /home/core/.kube ; \\\\\\n cp -f /etc/kubernetes/admin.conf /home/core/.kube/config ; \\\\\\n chown core:core /home/core/.kube/config ; \\\\\\n echo \\\"[kubeadm-init] done.\\\" \\\\\\n'\\n\\n# strictly start next step (serialization)\\nExecStartPost=/usr/bin/systemctl start install-calico.service\\n\\n[Install]\\nWantedBy=undercloud-bootstrap.target\\n\",\n \"enabled\": false,\n \"name\": \"kubeadm-init.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Install Calico\\nRequires=kubeadm-init.service\\nAfter=kubeadm-init.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-calico.done\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\n\\nExecStart=/bin/sh -eu -c '\\\\\\n echo \\\"[calico] waiting for API /readyz...\\\" ; \\\\\\n for i in $(seq 1 180); do kubectl get --raw=/readyz \\u003e/dev/null 2\\u003e\\u00261 \\u0026\\u0026 break; sleep 2; done ; \\\\\\n echo \\\"[calico] create namespace + operator...\\\" ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/namespace.yaml ; \\\\\\n kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/operator-crds.yaml || true ; \\\\\\n kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/tigera-operator.yaml || true ; \\\\\\n echo \\\"[calico] wait for tigera-operator...\\\" ; \\\\\\n kubectl wait deployment -n tigera-operator tigera-operator --for condition=Available=True --timeout=1200s ; \\\\\\n echo \\\"[calico] apply custom resources...\\\" ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/custom-resources.yaml ; \\\\\\n echo \\\"[calico] wait for calico-apiserver...\\\" ; \\\\\\n kubectl wait deployment -n calico-apiserver calico-apiserver --for condition=Available=True --timeout=1200s ; \\\\\\n echo \\\"[calico] apply peers + pools...\\\" ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/calico-peer.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/ippools.yaml ; \\\\\\n echo \\\"[calico] done.\\\" \\\\\\n'\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-calico.done\\nExecStartPost=/usr/bin/systemctl start install-ceph.service\\n\\n[Install]\\nWantedBy=undercloud-bootstrap.target\\n\",\n \"enabled\": false,\n \"name\": \"install-calico.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Install Ceph CSI\\nRequires=install-calico.service\\nAfter=install-calico.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-ceph.done\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\n\\nExecStart=/bin/sh -eu -c '\\\\\\n echo \\\"[ceph] waiting for API /readyz...\\\" ; \\\\\\n for i in $(seq 1 180); do kubectl get --raw=/readyz \\u003e/dev/null 2\\u003e\\u00261 \\u0026\\u0026 break; sleep 2; done ; \\\\\\n echo \\\"[ceph] apply manifests...\\\" ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/namespace.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-nodeplugin-rbac.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-provisioner-rbac.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/secrets.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/ceph-conf.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-cephfsplugin-provisioner.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-cephfsplugin.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-config-map.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-encryption-kms-config.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csidriver.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/storage-class.yaml ; \\\\\\n echo \\\"[ceph] done.\\\" \\\\\\n'\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-ceph.done\\nExecStartPost=/usr/bin/systemctl start install-gitea.service\\n\\n[Install]\\nWantedBy=undercloud-bootstrap.target\\n\",\n \"enabled\": false,\n \"name\": \"install-ceph.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Install Gitea\\nRequires=install-ceph.service\\nAfter=install-ceph.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-gitea.done\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\n\\nExecStart=/bin/sh -eu -c '\\\\\\n echo \\\"[gitea] wait for ceph provisioner...\\\" ; \\\\\\n kubectl wait deployment -n ceph csi-cephfsplugin-provisioner --for condition=Available=True --timeout=1200s ; \\\\\\n echo \\\"[gitea] apply manifests...\\\" ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/namespace.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/secrets.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/db.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/adminer.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/gitea.yaml ; \\\\\\n echo \\\"[gitea] wait for gitea deployment...\\\" ; \\\\\\n kubectl -n gitea wait deployment gitea --for=condition=Available=True --timeout=1200s ; \\\\\\n echo \\\"[gitea] run startup...\\\" ; \\\\\\n kubectl exec deploy/gitea -n gitea -- /bin/startup.sh ; \\\\\\n echo \\\"[gitea] done.\\\" \\\\\\n'\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-gitea.done\\nExecStartPost=/usr/bin/systemctl start install-argocd.service\\n\\n[Install]\\nWantedBy=undercloud-bootstrap.target\\n\",\n \"enabled\": false,\n \"name\": \"install-gitea.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Install ArgoCD\\nRequires=install-calico.service install-gitea.service\\nAfter=install-calico.service install-gitea.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-argocd.done\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\n\\nExecStart=/bin/sh -eu -c '\\\\\\n echo \\\"[argocd] wait for coredns...\\\" ; \\\\\\n kubectl -n kube-system wait deploy coredns --for=condition=Available=True --timeout=1200s ; \\\\\\n echo \\\"[argocd] install...\\\" ; \\\\\\n kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/namespace.yaml ; \\\\\\n kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/install.yaml ; \\\\\\n kubectl -n argocd wait deploy argocd-server --for=condition=Available=True --timeout=1200s ; \\\\\\n kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/repo.yaml ; \\\\\\n kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/apps.yaml ; \\\\\\n kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/ingress.yaml ; \\\\\\n echo \\\"[argocd] done.\\\" \\\\\\n'\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-argocd.done\\nExecStartPost=/usr/bin/systemctl start pin-service-ips.service\\n\\n[Install]\\nWantedBy=undercloud-bootstrap.target\\n\",\n \"enabled\": false,\n \"name\": \"install-argocd.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Pin fixed dual-stack ClusterIPs for kube-dns, argocd-server and whisker\\nRequires=install-argocd.service install-calico.service kubeadm-init.service\\nAfter=install-argocd.service install-calico.service kubeadm-init.service network-online.target\\nWants=network-online.target\\nConditionPathExists=!/var/lib/undercloud-stamps/pin-service-ips.done\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=PATH=/usr/bin:/usr/sbin:/opt/bin\\nExecStart=/bin/sh -eu -c '\\\\\\n echo \\\"[pin-service-ips] waiting for API...\\\" ; \\\\\\n for i in $(seq 1 120); do kubectl get --raw=/readyz \\u003e/dev/null 2\\u003e\\u00261 \\u0026\\u0026 break; sleep 2; done ; \\\\\\n echo \\\"[pin-service-ips] ensure namespaces exist...\\\" ; \\\\\\n kubectl get ns kube-system \\u003e/dev/null ; \\\\\\n kubectl get ns argocd \\u003e/dev/null 2\\u003e\\u00261 || kubectl create ns argocd ; \\\\\\n kubectl get ns calico-system \\u003e/dev/null ; \\\\\\n echo \\\"[pin-service-ips] wait for coredns/argocd readiness (best effort)...\\\" ; \\\\\\n kubectl -n kube-system wait deploy coredns --for=condition=Available=True --timeout=300s || true ; \\\\\\n kubectl -n argocd wait deploy argocd-server --for=condition=Available=True --timeout=600s || true ; \\\\\\n echo \\\"[pin-service-ips] replace Services with fixed ClusterIPs...\\\" ; \\\\\\n kubectl -n kube-system delete svc kube-dns --ignore-not-found ; \\\\\\n kubectl apply -f /etc/kubernetes/addons/kube-dns-fixed-svc.yaml ; \\\\\\n kubectl -n argocd delete svc argocd-server --ignore-not-found ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/service.yaml ; \\\\\\n kubectl -n calico-system delete svc whisker --ignore-not-found || true ; \\\\\\n kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/whisker.yaml || true ; \\\\\\n echo \\\"[pin-service-ips] done.\\\" \\\\\\n'\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/pin-service-ips.done\\n\\n[Install]\\nWantedBy=undercloud-bootstrap.target\\n\",\n \"enabled\": false,\n \"name\": \"pin-service-ips.service\"\n }\n ]\n }\n}", + "data": "{\n \"ignition\": {\n \"config\": {\n \"replace\": {\n \"verification\": {}\n }\n },\n \"proxy\": {},\n \"security\": {\n \"tls\": {}\n },\n \"timeouts\": {},\n \"version\": \"3.4.0\"\n },\n \"kernelArguments\": {},\n \"passwd\": {\n \"users\": [\n {\n \"name\": \"core\",\n \"sshAuthorizedKeys\": [\n \"ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIHHEAlPo3v4U67Y3411pTjIMkQxwlFWdXrBJkSzXenDH flatcar@undercloud\"\n ]\n }\n ]\n },\n \"storage\": {\n \"directories\": [\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/opt/bin\",\n \"user\": {},\n \"mode\": 493\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/opt/cni/bin\",\n \"user\": {},\n \"mode\": 755\n },\n {\n \"group\": {},\n \"path\": \"/etc/kubernetes/manifests\",\n \"user\": {},\n \"mode\": 493\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/install-calico\",\n \"user\": {},\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/var/lib/undercloud-stamps\",\n \"user\": {},\n \"mode\": 493\n }\n ],\n \"files\": [\n {\n \"group\": {},\n \"path\": \"/etc/hostname\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"\",\n \"source\": \"data:,control-plane1%0A\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"path\": \"/etc/systemd/network/00-eth.network\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/2SMMWuFMBRG9/yUDjE3FcVABsFSOlSkHcUhJLco1aQkUeu/f+h76PC407mH77SfKuq+I7WaUGLsXwhpa4yr878dKY3xGIL8MYyJ/bgQBSRZehrOGIg0ZyIHyC79riKuanvWr6Sqv68eCAFnCxhllNMCEn4VHs/77gCgQCo3qcEGOVuDXo9uNnR0Wo3ko1myUmv8i1+ltO7gxg+L0tvbf0QbBmfDLm4BAAD//4CQfvX4AAAA\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/hosts\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/5ST7W6DIBRA//MUvEDJ/Vhqdt/GqcuWEjGszV5/wVEjlBa5/sLLOYknEakzYMCg1tq6obdf7ueqRML5f7a3+ns5n/KTWz764aIUAaC8dSAd4llI5B21Htx89c6eFtvPE5rbPE5+sO42mlWT7QsOyhxUcVDBwZmDKw5+dCCg/nX+MvnCR8RFiaK4pGcUlSiOS35GsVKfI4CEJ5YO01g7UVBBUY+dKLigqLfeK0LmMNXUKUQ76EXpFOId9Co0hp+DzNq4MfAdpfawd5Tbg0Z0bVkNuV2mAwG3y3wg3F8AAAD//+yy88FYBAAA\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/motd\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/6yRwYobMQyG7/MUOmYX4lzL3sJ2oYV2A01yaG+KrMyIeOxBkodOn74kk0MPaSg0/8WWkD4+4+f/T/MMf8l6v/u0+fb5x9tHWL++vm23sHn/8v3G4B3GP+fC2HViYJM59yAGA6pDOQKCMVXlCJKPiuZayatyuMVYpwRILqO4sAEqQ1+yeDnvY46QSttyDPc89hmrd0Xl13mHiM2gKPRi1fhsZq5CniYYtHRyEL+y/2D0OIGy1eQgGaIYyZAko06zBbc4i5Ycbns84G+b5YPTnOqBNbOzAZXsWhLDkDAzvJfITfMVM7YMo+DLZZY8weJ6eWoIk1CBxXxeWiMn1gJLOCCd6gCLufHUoLaFInTug72sVnO5NNaRNcxVsJHC6QPWn6HmyEqp1BjouHr8u38HAAD//04R3qpqAwAA\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"path\": \"/etc/sysctl.d/99-k8s.conf\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/4zQX27DIAwG8PeeIhcIyqos6h56FuQS01kFGxmWtrefsmXZHzE13wsS+iH7g7EYSlNvKFkvegUdm785Nk+7xQ2bnRP2BkIwiyY+V1y/xZ2UxjMuR8u+dRBCS6nAKWB+7IYFfjnG4ikUVMPeOmEuCu5iI9zWud1HKltqsp9vf7TuKq1HyvNQO99+uykamVCdxEjFRoyi99ov+myIpZC/mwg3+5ZR7RWKe53rru553+8PhyomzgXYrfzY9N3LsLugMgaTgMk1/2Ru/wtaYSuScm3R9wAAAP//X0LFgj8CAAA=\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/flatcar/update.conf\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"\",\n \"source\": \"data:,REBOOT_STRATEGY%3Doff%0A\",\n \"verification\": {}\n },\n \"mode\": 272\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/kubeadm\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubeadm\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/kubelet\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubelet\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/kubectl\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubectl\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/calicoctl\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/calicoctl\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/velero\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/velero\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/etc/kubernetes/kubeadm-init.yaml\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/6xV3W7jNhO951MQvNmbUD+WYie8y7dZfFgsWgSb/gFFLyhyJBOiSXVIOfY+fUHJNqKugwZFbcCQoTlHhzPnjORgfgEMxjtB+7EBqXdZfxcy4/N92UCUFemN04J+diZ+9K413YgyGu9I430MEeXwk+/BBUEo5TSma0FZv+/KrcpihfDi9ziEerVG3zJCKaUagkIzxOmpTEMrRxvPj6cX3plshsRoBWUFI85r+AqdSRUTAaHUyR0IqryL6C0frHRQEkoVmmeveoiCjs4cRJ7nOLo81UnjAPWryyx41RM6ibAQPx0iygfsplNRmh7KzSAoWxVFKepNITZluRYrIe7LWaCyY4iAXLsgKCuLrMjuy+y2ulkiWlEKcVvNmL234w74YMfOOK4NCspyP8TcmgYOoPKkBh1ECPlJ2Kk45DM2n8oYsV5J+/D0+ZPTgzcuJtlS7wGjCfCgNUIIb6tvjNNPHqOg67quiEqw1igZ4QscBWUtwEZVcKtaJdUG6s2mVmtolSzqdS3rlb6737RV0W6qoqjqQq9vN3fNrWqbWla6qhtGOOfk3Vb7OLdy6bbTdJ/ScC+npOz3VheFSN+VEGVR/CHSERhxEF889sZ1qRWD189j45IT/tYDKUS+rm+meaWffFWnjgTAvVHwBmgaYl4WdzfnOZ9x2oVHv5NmisBdyEanAZX1o86mEbHzOSzgD9LJDjDpg6XdWgsH/t+ZI1GmEXN5NsIHIT6keTwD7q8qACcbC1x2HUIno0eOfoypm5RFHGEmHdAfjlxZAy7y5BreGguC5hAX8obe5C16F/lrRKYwfk/Tw/HE8l6aHo4TDcKfI4S4BakBL6rkezXJi54lkbTWv4DmackEQb8XcAUzdZPPfwIfEFpzEPQ3/hV2PgKftgu/guvQj8MZJ+j8ucD+n+5eQY0BMKlbAC+onwMguRpACzFTU8iWOSxPOfwy1yxzePEQE4KRLUgbt9/+Z5x+WNw4rcN/CsOp6sfn+e3B3lyW/PVOZWQ2+NNk98cUjX+RDKKmdj+iSRmgLBxDhJ1mRI5xCy6mBXh6wUjn3XHnx0U6tKApCoTSF2i23vfXbiYuj+bbhWrnNQj664wgfwUAAP//F9AABn8HAAA=\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"path\": \"/etc/kubernetes/addons/kube-dns-fixed-svc.yaml\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/4yRwU7rMBBF9/6K+YCXKn6Uls62FVJ3FgU2iMXUGZAVJzb2JFL+HoWmEBYV9fLOuUejMUX3zCm70CL0WtWurRAOnHpnWTUsVJEQKoCWGkaouyMXVZunIEey5zQPWbhRAJ6O7PPYAajvckExzoo5sh1nMkRG2PouC6e9UQAu3lPj/GCCd3ZAeOCPziXedeQPQrb+QRxnhJe96Vf/YG/65asCsGcTwv+y1Lhcl7jWeoVvqBFvb+bItFxxkRxnulyUi41efAUxJPlunU5xusL4xiHCVASIKUiwwSM87cyUCaV3FjMHZ55CbPzL9bi9wtWwJGd/77XRV9smNLNnKyFd+sLPAAAA//8MLaPANAIAAA==\",\n \"verification\": {}\n },\n \"mode\": 420\n }\n ]\n },\n \"systemd\": {\n \"units\": [\n {\n \"contents\": \"[Unit]\\nDescription=Load necessary kernel modules\\nBefore=containerd.service kubeadm-init.service\\n\\n[Service]\\nType=oneshot\\n\\nExecStart=/usr/bin/modprobe br_netfilter\\nExecStart=/usr/bin/modprobe overlay\\nRemainAfterExit=yes\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"modules-load.service\"\n },\n {\n \"enabled\": true,\n \"name\": \"systemd-networkd-wait-online.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=containerd container runtime\\nAfter=network.target modules-load.service\\n\\n[Service]\\n#StandardOutput=journal+console\\n#StandardError=journal+console\\n\\nExecStart=/usr/bin/containerd\\nRestart=always\\nRestartSec=5\\nDelegate=yes\\nKillMode=process\\nOOMScoreAdjust=-999\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"containerd.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Set Timezone\\nAfter=network-online.target\\nWants=network-online.target\\n[Service]\\nStandardOutput=journal+console\\nStandardError=journal+console\\n\\nExecStart=/bin/sh -c 'echo \\\"setting timezone to Europe/Berlin\\\"'\\nStandardOutput=journal+console\\nStandardError=journal+console\\nType=oneshot\\nRestart=on-failure\\nExecStart=/usr/bin/timedatectl set-timezone Europe/Berlin\\nExecStart=/usr/bin/timedatectl set-ntp true \\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"set-timezone.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=kubelet, the Kubernetes Node Agent\\nDocumentation=https://kubernets.io/docs/home\\nWants=network-online.target\\nAfter=network-online.target\\n[Service]\\n#StandardOutput=journal+console\\n#StandardError=journal+console\\n\\n#EnvironmentFile=/run/metadata/coreos\\nEnvironment=\\\"KUBELET_KUBECONFIG_ARGS=--bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf\\\"\\nEnvironment=\\\"KUBELET_CONFIG_ARGS=--config=/var/lib/kubelet/config.yaml\\\"\\n# This is a file that \\\"kubeadm init\\\" and \\\"kubeadm join\\\" generates at runtime, populating the KUBELET_KUBEADM_ARGS variable dynamically\\nEnvironmentFile=-/var/lib/kubelet/kubeadm-flags.env\\nExecStart=/opt/bin/kubelet $KUBELET_KUBECONFIG_ARGS $KUBELET_CONFIG_ARGS $KUBELET_KUBEADM_ARGS $KUBELET_EXTRA_ARGS\\nRestart=always\\nStartLimitInterval=0\\nRestartSec=10\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"kubelet.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Kubeadm Init Cluster\\nAfter=network-online.target containerd.service kubelet.service\\nWants=network-online.target\\nConditionPathExists=!/etc/kubernetes/kubelet.conf\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nExecStart=/opt/bin/kubeadm reset -f\\n\\nExecStart=/bin/sh -c 'echo \\\"kubeadm-init.service started...\\\"'\\n\\n# Environment\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin:/opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent-uds/\\n\\nExecStartPre=/bin/sleep 30s\\nExecStart=/bin/sh -c 'echo \\\"running kubeadm init...\\\"'\\nExecStart=/opt/bin/kubeadm init --upload-certs --config=/etc/kubernetes/kubeadm-init.yaml\\n\\n# copy files for kubectl\\nExecStart=/bin/sh -c 'echo \\\"copying files (admin.conf) to core home folder.\\\"'\\nExecStartPost=/usr/bin/mkdir -p /home/core/.kube\\nExecStartPost=/usr/bin/cp -i /etc/kubernetes/admin.conf /home/core/.kube/config\\nExecStartPost=/usr/bin/chown core:core /home/core/.kube/config\\n\\nExecStartPost=/usr/bin/systemctl disable kubeadm-init.service\\nRestart=on-failure\\nRestartSec=120s\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"kubeadm-init.service\"\n },\n {\n \"contents\": \"[Unit]\\nWants=kubeadm-init.service\\nAfter=kubeadm-init.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-calico.done\\n\\n\\n[Service]\\nStandardOutput=journal+console\\nStandardError=journal+console\\n\\nExecStart=/bin/sh -c 'echo \\\"install.calico.service started...\\\"'\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nExecStart=/bin/sh -c 'echo \\\"witing 30s...\\\"'\\nExecStart=/bin/sleep 30s\\nExecStart=/bin/sh -c 'echo \\\"create calico namespace...\\\"'\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/namespace.yaml\\nExecStart=/bin/sh -c 'echo \\\"install tigera operator...\\\"'\\nExecStart=-/opt/bin/kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/operator-crds.yaml\\nExecStart=-/opt/bin/kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/tigera-operator.yaml\\nExecStart=/bin/sh -c 'echo \\\"witing 60s...\\\"'\\nExecStart=/bin/sleep 60s\\nExecStart=/bin/sh -c 'echo \\\"witing for tigera operator... (20mini max)\\\"'\\nExecStart=/opt/bin/kubectl wait deployment -n tigera-operator tigera-operator --for condition=Available=True --timeout=1200s\\nExecStart=/bin/sh -c 'echo \\\"create clico custom ressources...\\\"'\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/custom-resources.yaml\\n\\nExecStart=/bin/sh -c 'echo \\\"witing 3m..\\\"'\\nExecStart=/bin/sleep 3m\\n#ExecStart=/bin/sh -c 'echo \\\"apply calico (calico-apiserver)...\\\"'\\n#ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/calico.yaml\\n#ExecStart=/bin/sh -c 'echo \\\"witing 1m...\\\"'\\n#ExecStart=/bin/sleep 2m\\nExecStart=/bin/sh -c 'echo \\\"witing calico-apiserver... (20mini max)\\\"'\\nExecStart=/opt/bin/kubectl wait deployment -n calico-apiserver calico-apiserver --for condition=Available=True --timeout=1200s\\nExecStart=/bin/sh -c 'echo \\\"witing 120s...\\\"'\\nExecStart=/bin/sleep 2m\\nExecStart=/bin/sh -c 'echo \\\"apply calico-peers...\\\"'\\nExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/calico-peer.yaml\\nExecStart=/bin/sh -c 'echo \\\"witing 60s...\\\"'\\nExecStart=/bin/sleep 1m\\nExecStart=/bin/sh -c 'echo \\\"apply calico-ippools...\\\"'\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/ippools.yaml\\n\\n#ExecStart=/bin/sh -c 'echo \\\"witing for whisker..\\\"'\\n#ExecStart=/opt/bin/kubectl wait deployment -n calico-system whisker --for condition=Available=True --timeout=1200s\\n#ExecStart=/bin/sh -c 'echo \\\"port-forward -n calico-system service/whisker 8081:8081\\\"'\\n#ExecStart=/opt/bin/kubectl port-forward -n calico-system service/whisker 8081:8081\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-calico.done\\nExecStart=/usr/bin/systemctl disable install-calico.service\\n#RemainAfterExit=true\\nRestart=on-failure\\nRestartSec=120s\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"install-calico.service\"\n },\n {\n \"contents\": \"[Unit]\\nWants=kubeadm-init.service\\nAfter=kubeadm-init.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-ceph.done\\n\\n[Service]\\nStandardOutput=journal+console\\nStandardError=journal+console\\n\\nExecStart=/bin/sh -c 'echo \\\"install.ceph.service started...\\\"'\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\nType=oneshot\\n\\nStandardOutput=journal+console\\nStandardError=journal+console\\nExecStart=/bin/sh -c 'echo \\\"witing 30s...\\\"'\\nExecStart=/bin/sleep 30s\\nExecStart=/bin/sh -c 'echo \\\"create ceph namespace...\\\"'\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/namespace.yaml\\n\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-nodeplugin-rbac.yaml\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-provisioner-rbac.yaml\\n\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/secrets.yaml\\n\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/ceph-conf.yaml\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-cephfsplugin-provisioner.yaml\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-cephfsplugin.yaml\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-config-map.yaml\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-encryption-kms-config.yaml\\n\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csidriver.yaml\\n\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/storage-class.yaml\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-ceph.done\\nExecStart=/usr/bin/systemctl disable install-ceph.service\\n#RemainAfterExit=true\\nRestart=on-failure\\nRestartSec=120s\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"install-ceph.service\"\n },\n {\n \"contents\": \"[Unit]\\nWants=install-ceph.service\\nAfter=install-ceph.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-gitea.done\\n\\n[Service]\\nStandardOutput=journal+console\\nStandardError=journal+console\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\nType=oneshot\\n\\nExecStart=/opt/bin/kubectl wait deployment -n ceph csi-cephfsplugin-provisioner --for condition=Available=True --timeout=1200s\\nExecStart=/bin/sleep 4m\\nExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/namespace.yaml\\nExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/secrets.yaml\\nExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/db.yaml\\nExecStart=/bin/sleep 60s\\nExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/adminer.yaml\\nExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/gitea.yaml\\nExecStart=/bin/sleep 3m\\nExecStart=/opt/bin/kubectl exec deploy/gitea -n gitea -- /bin/startup.sh\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-gitea.done\\nExecStart=/usr/bin/systemctl disable install-gitea.service\\nRestart=on-failure\\nRestartSec=120s\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"install-gitea.service\"\n },\n {\n \"contents\": \"[Unit]\\nWants=install-calico.service\\nAfter=install-calico.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-argocd.done\\n[Service]\\nStandardOutput=journal+console\\nStandardError=journal+console\\n\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\nType=oneshot\\n\\nExecStart=/opt/bin/kubectl wait deployment -n kube-system coredns --for condition=Available=True --timeout=600s\\n\\nExecStart=/bin/sleep 1m\\nExecStart=/opt/bin/kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/namespace.yaml\\nExecStart=/opt/bin/kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/install.yaml\\nExecStart=/opt/bin/kubectl wait deployment -n argocd argocd-server --for condition=Available=True --timeout=600s\\n\\nExecStart=/opt/bin/kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/repo.yaml\\nExecStart=/bin/sleep 10s\\nExecStart=/opt/bin/kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/apps.yaml\\n\\nExecStart=/bin/sleep 10s\\nExecStart=/opt/bin/kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/ingress.yaml\\n\\n\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/repos/k8aux-bootstrap.yaml\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/repos/k8aux-apps.yaml\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/apps/argocd.yaml\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/apps/calico.yaml\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/apps/rook-ceph.yaml\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/apps/gitea.yaml\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/argocd/argocd-secret.yaml\\n##ExecStart=/bin/sleep 10m\\n#ExecStart=/opt/bin/kubectl wait deployment -n gitea gitea --for condition=Available=True --timeout=4800s\\n#ExecStart=/bin/sleep 10m\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://gitea.gitea.svc.k8aux.undercloud.cf:3000/undercloud/k8aux-apps/raw/branch/main/app-of-apps/app-of-apps.yaml\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-argocd.done\\nExecStart=/usr/bin/systemctl disable install-argocd.service\\nRestart=on-failure\\nRestartSec=120s\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"install-argocd.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Pin fixed dual-stack ClusterIPs for kube-dns, argocd-server and whisker\\nAfter=install-argocd.service install-calico.service kubeadm-init.service network-online.target\\nWants=install-argocd.service install-calico.service kubeadm-init.service network-online.target\\nConditionPathExists=!/var/lib/undercloud-stamps/pin-service-ips.done\\n\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=PATH=/usr/bin:/usr/sbin:/opt/bin\\nExecStart=/bin/sh -eu -c '\\\\\\n echo \\\"[pin-service-ips] waiting for API...\\\" ; \\\\\\n for i in $(seq 1 120); do kubectl get --raw=/readyz \\u003e/dev/null 2\\u003e\\u00261 \\u0026\\u0026 break; sleep 2; done ; \\\\\\n echo \\\"[pin-service-ips] ensure namespaces exist...\\\" ; \\\\\\n kubectl get ns kube-system \\u003e/dev/null ; \\\\\\n kubectl get ns argocd \\u003e/dev/null 2\\u003e\\u00261 || kubectl create ns argocd ; \\\\\\n kubectl get ns calico-system \\u003e/dev/null ; \\\\\\n echo \\\"[pin-service-ips] wait for coredns/argocd readiness (best effort)...\\\" ; \\\\\\n kubectl -n kube-system wait deploy coredns --for=condition=Available=True --timeout=300s || true ; \\\\\\n kubectl -n argocd wait deploy argocd-server --for=condition=Available=True --timeout=600s || true ; \\\\\\n echo \\\"[pin-service-ips] replace Services with fixed ClusterIPs...\\\" ; \\\\\\n kubectl -n kube-system delete svc kube-dns --ignore-not-found ; \\\\\\n kubectl apply -f /etc/kubernetes/addons/kube-dns-fixed-svc.yaml ; \\\\\\n kubectl -n argocd delete svc argocd-server --ignore-not-found ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/service.yaml ; \\\\\\n kubectl -n calico-system delete svc whisker --ignore-not-found || true ; \\\\\\n kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/whisker.yaml || true ; \\\\\\n echo \\\"[pin-service-ips] done.\\\" \\\\\\n'\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/pin-service-ips.done\\nRestart=on-failure\\nRestartSec=120s\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"pin-service-ips.service\"\n }\n ]\n }\n}", "file_name": "control-plane1-ignition-user-data", "resize": 0 } @@ -548,14 +610,14 @@ [ "2001:470:7116:2::91", "fd00:0:0:2::91", - "fe80::be24:11ff:fe8d:2749" + "fe80::be24:11ff:fe02:3135" ] ], "keyboard_layout": "en-us", "kvm_arguments": null, "mac_addresses": [ "00:00:00:00:00:00", - "BC:24:11:8D:27:49" + "BC:24:11:02:31:35" ], "machine": null, "memory": [ @@ -575,7 +637,7 @@ "disconnected": false, "enabled": true, "firewall": false, - "mac_address": "BC:24:11:8D:27:49", + "mac_address": "BC:24:11:02:31:35", "model": "virtio", "mtu": 0, "queues": 0, @@ -644,6 +706,358 @@ } ] }, + { + "mode": "managed", + "type": "proxmox_virtual_environment_vm", + "name": "control_plane2", + "provider": "provider[\"registry.terraform.io/bpg/proxmox\"]", + "instances": [ + { + "schema_version": 0, + "attributes": { + "acpi": true, + "agent": [], + "amd_sev": [], + "audio_device": [], + "bios": "seabios", + "boot_order": [ + "virtio0" + ], + "cdrom": [], + "clone": [ + { + "datastore_id": "", + "full": true, + "node_name": "hyper1", + "retries": 1, + "vm_id": 999 + } + ], + "cpu": [ + { + "affinity": "", + "architecture": "", + "cores": 2, + "flags": null, + "hotplugged": 0, + "limit": 0, + "numa": false, + "sockets": 1, + "type": "host", + "units": 1024 + } + ], + "description": "kubernetes control-plane2", + "disk": [], + "efi_disk": [], + "hook_script_file_id": null, + "hostpci": [], + "id": "1002", + "initialization": [ + { + "datastore_id": "Pool1", + "dns": [], + "interface": "ide2", + "ip_config": [], + "meta_data_file_id": "", + "network_data_file_id": "", + "type": "", + "user_account": [], + "user_data_file_id": "cephfs:snippets/control-plane2-ignition-user-data", + "vendor_data_file_id": "" + } + ], + "ipv4_addresses": [ + [ + "127.0.0.1" + ], + [ + "10.0.2.92" + ] + ], + "ipv6_addresses": [ + [ + "::1" + ], + [ + "2001:470:7116:2::92", + "fd00:0:0:2::92", + "fe80::be24:11ff:fe62:a48f" + ] + ], + "keyboard_layout": "en-us", + "kvm_arguments": null, + "mac_addresses": [ + "00:00:00:00:00:00", + "BC:24:11:62:A4:8F" + ], + "machine": null, + "memory": [ + { + "dedicated": 3072, + "floating": 3072, + "hugepages": "", + "keep_hugepages": false, + "shared": 0 + } + ], + "migrate": false, + "name": "control-plane2", + "network_device": [ + { + "bridge": "vmbr0", + "disconnected": false, + "enabled": true, + "firewall": false, + "mac_address": "BC:24:11:62:A4:8F", + "model": "virtio", + "mtu": 0, + "queues": 0, + "rate_limit": 0, + "trunks": "", + "vlan_id": 0 + } + ], + "network_interface_names": [ + "lo", + "eth0" + ], + "node_name": "hyper2", + "numa": [], + "on_boot": true, + "operating_system": [], + "pool_id": null, + "protection": false, + "reboot": false, + "reboot_after_update": true, + "rng": [], + "scsi_hardware": "virtio-scsi-pci", + "serial_device": [], + "smbios": [], + "started": true, + "startup": [ + { + "down_delay": -1, + "order": 1, + "up_delay": -1 + } + ], + "stop_on_destroy": false, + "tablet_device": true, + "tags": [ + "control-plane", + "flatcar", + "kubernetes", + "terraform" + ], + "template": false, + "timeout_clone": 1800, + "timeout_create": 1800, + "timeout_migrate": 1800, + "timeout_move_disk": 1800, + "timeout_reboot": 1800, + "timeout_shutdown_vm": 1800, + "timeout_start_vm": 1800, + "timeout_stop_vm": 300, + "tpm_state": [], + "usb": [], + "vga": [], + "virtiofs": [], + "vm_id": 1002, + "watchdog": [] + }, + "sensitive_attributes": [], + "identity_schema_version": 0, + "private": "bnVsbA==", + "dependencies": [ + "data.ct_config.control_plane1_ignition", + "data.ct_config.control_plane2_ignition", + "null_resource.wait_for_cp1_api", + "proxmox_virtual_environment_download_file.flatcar_image", + "proxmox_virtual_environment_file.control_plane1_ignition", + "proxmox_virtual_environment_file.control_plane2_ignition", + "proxmox_virtual_environment_vm.control_plane1", + "proxmox_virtual_environment_vm.flatcar_template" + ] + } + ] + }, + { + "mode": "managed", + "type": "proxmox_virtual_environment_vm", + "name": "control_plane3", + "provider": "provider[\"registry.terraform.io/bpg/proxmox\"]", + "instances": [ + { + "schema_version": 0, + "attributes": { + "acpi": true, + "agent": [], + "amd_sev": [], + "audio_device": [], + "bios": "seabios", + "boot_order": [ + "virtio0" + ], + "cdrom": [], + "clone": [ + { + "datastore_id": "", + "full": true, + "node_name": "hyper1", + "retries": 1, + "vm_id": 999 + } + ], + "cpu": [ + { + "affinity": "", + "architecture": "", + "cores": 2, + "flags": null, + "hotplugged": 0, + "limit": 0, + "numa": false, + "sockets": 1, + "type": "host", + "units": 1024 + } + ], + "description": "kubernetes control-plane3", + "disk": [], + "efi_disk": [], + "hook_script_file_id": null, + "hostpci": [], + "id": "1003", + "initialization": [ + { + "datastore_id": "Pool1", + "dns": [], + "interface": "ide2", + "ip_config": [], + "meta_data_file_id": "", + "network_data_file_id": "", + "type": "", + "user_account": [], + "user_data_file_id": "cephfs:snippets/control-plane3-ignition-user-data", + "vendor_data_file_id": "" + } + ], + "ipv4_addresses": [ + [ + "127.0.0.1" + ], + [ + "10.0.2.93" + ] + ], + "ipv6_addresses": [ + [ + "::1" + ], + [ + "2001:470:7116:2::93", + "fd00:0:0:2::93", + "fe80::be24:11ff:fe42:d661" + ] + ], + "keyboard_layout": "en-us", + "kvm_arguments": null, + "mac_addresses": [ + "00:00:00:00:00:00", + "BC:24:11:42:D6:61" + ], + "machine": null, + "memory": [ + { + "dedicated": 3072, + "floating": 3072, + "hugepages": "", + "keep_hugepages": false, + "shared": 0 + } + ], + "migrate": false, + "name": "control-plane3", + "network_device": [ + { + "bridge": "vmbr0", + "disconnected": false, + "enabled": true, + "firewall": false, + "mac_address": "BC:24:11:42:D6:61", + "model": "virtio", + "mtu": 0, + "queues": 0, + "rate_limit": 0, + "trunks": "", + "vlan_id": 0 + } + ], + "network_interface_names": [ + "lo", + "eth0" + ], + "node_name": "hyper3", + "numa": [], + "on_boot": true, + "operating_system": [], + "pool_id": null, + "protection": false, + "reboot": false, + "reboot_after_update": true, + "rng": [], + "scsi_hardware": "virtio-scsi-pci", + "serial_device": [], + "smbios": [], + "started": true, + "startup": [ + { + "down_delay": -1, + "order": 1, + "up_delay": -1 + } + ], + "stop_on_destroy": false, + "tablet_device": true, + "tags": [ + "control-plane", + "flatcar", + "kubernetes", + "terraform" + ], + "template": false, + "timeout_clone": 1800, + "timeout_create": 1800, + "timeout_migrate": 1800, + "timeout_move_disk": 1800, + "timeout_reboot": 1800, + "timeout_shutdown_vm": 1800, + "timeout_start_vm": 1800, + "timeout_stop_vm": 300, + "tpm_state": [], + "usb": [], + "vga": [], + "virtiofs": [], + "vm_id": 1003, + "watchdog": [] + }, + "sensitive_attributes": [], + "identity_schema_version": 0, + "private": "bnVsbA==", + "dependencies": [ + "data.ct_config.control_plane1_ignition", + "data.ct_config.control_plane3_ignition", + "null_resource.wait_for_cp1_api", + "proxmox_virtual_environment_download_file.flatcar_image", + "proxmox_virtual_environment_file.control_plane1_ignition", + "proxmox_virtual_environment_file.control_plane3_ignition", + "proxmox_virtual_environment_vm.control_plane1", + "proxmox_virtual_environment_vm.flatcar_template" + ] + } + ] + }, { "mode": "managed", "type": "proxmox_virtual_environment_vm", @@ -729,7 +1143,7 @@ "keyboard_layout": "en-us", "kvm_arguments": "", "mac_addresses": [ - "BC:24:11:6E:7C:BE" + "BC:24:11:2B:DB:03" ], "machine": "", "memory": [ @@ -749,7 +1163,7 @@ "disconnected": false, "enabled": true, "firewall": false, - "mac_address": "BC:24:11:6E:7C:BE", + "mac_address": "BC:24:11:2B:DB:03", "model": "virtio", "mtu": 0, "queues": 0, @@ -804,6 +1218,742 @@ ] } ] + }, + { + "mode": "managed", + "type": "proxmox_virtual_environment_vm", + "name": "worker1", + "provider": "provider[\"registry.terraform.io/bpg/proxmox\"]", + "instances": [ + { + "schema_version": 0, + "attributes": { + "acpi": true, + "agent": [], + "amd_sev": [], + "audio_device": [], + "bios": "seabios", + "boot_order": [ + "virtio0" + ], + "cdrom": [], + "clone": [ + { + "datastore_id": "", + "full": true, + "node_name": "hyper1", + "retries": 1, + "vm_id": 999 + } + ], + "cpu": [ + { + "affinity": "", + "architecture": "", + "cores": 2, + "flags": null, + "hotplugged": 0, + "limit": 0, + "numa": false, + "sockets": 1, + "type": "host", + "units": 1024 + } + ], + "description": "kubernetes worker1", + "disk": [], + "efi_disk": [], + "hook_script_file_id": null, + "hostpci": [], + "id": "1011", + "initialization": [ + { + "datastore_id": "Pool1", + "dns": [], + "interface": "ide2", + "ip_config": [], + "meta_data_file_id": "", + "network_data_file_id": "", + "type": "", + "user_account": [], + "user_data_file_id": "cephfs:snippets/worker1-ignition-user-data", + "vendor_data_file_id": "" + } + ], + "ipv4_addresses": [ + [ + "127.0.0.1" + ], + [ + "10.0.2.101" + ] + ], + "ipv6_addresses": [ + [ + "::1" + ], + [ + "2001:470:7116:2::101", + "fd00:0:0:2::101", + "fe80::be24:11ff:fe7a:bb6a" + ] + ], + "keyboard_layout": "en-us", + "kvm_arguments": null, + "mac_addresses": [ + "00:00:00:00:00:00", + "BC:24:11:7A:BB:6A" + ], + "machine": null, + "memory": [ + { + "dedicated": 10240, + "floating": 10240, + "hugepages": "", + "keep_hugepages": false, + "shared": 0 + } + ], + "migrate": false, + "name": "worker1", + "network_device": [ + { + "bridge": "vmbr0", + "disconnected": false, + "enabled": true, + "firewall": false, + "mac_address": "BC:24:11:7A:BB:6A", + "model": "virtio", + "mtu": 0, + "queues": 0, + "rate_limit": 0, + "trunks": "", + "vlan_id": 0 + } + ], + "network_interface_names": [ + "lo", + "eth0" + ], + "node_name": "hyper1", + "numa": [], + "on_boot": true, + "operating_system": [], + "pool_id": null, + "protection": false, + "reboot": false, + "reboot_after_update": true, + "rng": [], + "scsi_hardware": "virtio-scsi-pci", + "serial_device": [], + "smbios": [], + "started": true, + "startup": [ + { + "down_delay": -1, + "order": 1, + "up_delay": -1 + } + ], + "stop_on_destroy": false, + "tablet_device": true, + "tags": [ + "flatcar", + "kubernetes", + "terraform", + "worker" + ], + "template": false, + "timeout_clone": 1800, + "timeout_create": 1800, + "timeout_migrate": 1800, + "timeout_move_disk": 1800, + "timeout_reboot": 1800, + "timeout_shutdown_vm": 1800, + "timeout_start_vm": 1800, + "timeout_stop_vm": 300, + "tpm_state": [], + "usb": [], + "vga": [], + "virtiofs": [], + "vm_id": 1011, + "watchdog": [] + }, + "sensitive_attributes": [], + "identity_schema_version": 0, + "private": "bnVsbA==", + "dependencies": [ + "data.ct_config.control_plane1_ignition", + "data.ct_config.control_plane2_ignition", + "data.ct_config.control_plane3_ignition", + "data.ct_config.worker1_ignition", + "null_resource.wait_for_cluster_ready", + "null_resource.wait_for_cp1_api", + "null_resource.wait_for_cp2_cp3_api", + "proxmox_virtual_environment_download_file.flatcar_image", + "proxmox_virtual_environment_file.control_plane1_ignition", + "proxmox_virtual_environment_file.control_plane2_ignition", + "proxmox_virtual_environment_file.control_plane3_ignition", + "proxmox_virtual_environment_file.worker1_ignition", + "proxmox_virtual_environment_vm.control_plane1", + "proxmox_virtual_environment_vm.control_plane2", + "proxmox_virtual_environment_vm.control_plane3", + "proxmox_virtual_environment_vm.flatcar_template" + ] + } + ] + }, + { + "mode": "managed", + "type": "proxmox_virtual_environment_vm", + "name": "worker2", + "provider": "provider[\"registry.terraform.io/bpg/proxmox\"]", + "instances": [ + { + "schema_version": 0, + "attributes": { + "acpi": true, + "agent": [], + "amd_sev": [], + "audio_device": [], + "bios": "seabios", + "boot_order": [ + "virtio0" + ], + "cdrom": [], + "clone": [ + { + "datastore_id": "", + "full": true, + "node_name": "hyper1", + "retries": 1, + "vm_id": 999 + } + ], + "cpu": [ + { + "affinity": "", + "architecture": "", + "cores": 2, + "flags": null, + "hotplugged": 0, + "limit": 0, + "numa": false, + "sockets": 1, + "type": "host", + "units": 1024 + } + ], + "description": "kubernetes worker2", + "disk": [], + "efi_disk": [], + "hook_script_file_id": null, + "hostpci": [], + "id": "1012", + "initialization": [ + { + "datastore_id": "Pool1", + "dns": [], + "interface": "ide2", + "ip_config": [], + "meta_data_file_id": "", + "network_data_file_id": "", + "type": "", + "user_account": [], + "user_data_file_id": "cephfs:snippets/worker2-ignition-user-data", + "vendor_data_file_id": "" + } + ], + "ipv4_addresses": [ + [ + "127.0.0.1" + ], + [ + "10.0.2.102" + ] + ], + "ipv6_addresses": [ + [ + "::1" + ], + [ + "2001:470:7116:2::102", + "fd00:0:0:2::102", + "fe80::be24:11ff:fecb:3db3" + ] + ], + "keyboard_layout": "en-us", + "kvm_arguments": null, + "mac_addresses": [ + "00:00:00:00:00:00", + "BC:24:11:CB:3D:B3" + ], + "machine": null, + "memory": [ + { + "dedicated": 10240, + "floating": 10240, + "hugepages": "", + "keep_hugepages": false, + "shared": 0 + } + ], + "migrate": false, + "name": "worker2", + "network_device": [ + { + "bridge": "vmbr0", + "disconnected": false, + "enabled": true, + "firewall": false, + "mac_address": "BC:24:11:CB:3D:B3", + "model": "virtio", + "mtu": 0, + "queues": 0, + "rate_limit": 0, + "trunks": "", + "vlan_id": 0 + } + ], + "network_interface_names": [ + "lo", + "eth0" + ], + "node_name": "hyper2", + "numa": [], + "on_boot": true, + "operating_system": [], + "pool_id": null, + "protection": false, + "reboot": false, + "reboot_after_update": true, + "rng": [], + "scsi_hardware": "virtio-scsi-pci", + "serial_device": [], + "smbios": [], + "started": true, + "startup": [ + { + "down_delay": -1, + "order": 1, + "up_delay": -1 + } + ], + "stop_on_destroy": false, + "tablet_device": true, + "tags": [ + "flatcar", + "kubernetes", + "terraform", + "worker" + ], + "template": false, + "timeout_clone": 1800, + "timeout_create": 1800, + "timeout_migrate": 1800, + "timeout_move_disk": 1800, + "timeout_reboot": 1800, + "timeout_shutdown_vm": 1800, + "timeout_start_vm": 1800, + "timeout_stop_vm": 300, + "tpm_state": [], + "usb": [], + "vga": [], + "virtiofs": [], + "vm_id": 1012, + "watchdog": [] + }, + "sensitive_attributes": [], + "identity_schema_version": 0, + "private": "bnVsbA==", + "dependencies": [ + "data.ct_config.control_plane1_ignition", + "data.ct_config.control_plane2_ignition", + "data.ct_config.control_plane3_ignition", + "data.ct_config.worker2_ignition", + "null_resource.wait_for_cluster_ready", + "null_resource.wait_for_cp1_api", + "null_resource.wait_for_cp2_cp3_api", + "proxmox_virtual_environment_download_file.flatcar_image", + "proxmox_virtual_environment_file.control_plane1_ignition", + "proxmox_virtual_environment_file.control_plane2_ignition", + "proxmox_virtual_environment_file.control_plane3_ignition", + "proxmox_virtual_environment_file.worker2_ignition", + "proxmox_virtual_environment_vm.control_plane1", + "proxmox_virtual_environment_vm.control_plane2", + "proxmox_virtual_environment_vm.control_plane3", + "proxmox_virtual_environment_vm.flatcar_template" + ] + } + ] + }, + { + "mode": "managed", + "type": "proxmox_virtual_environment_vm", + "name": "worker3", + "provider": "provider[\"registry.terraform.io/bpg/proxmox\"]", + "instances": [ + { + "schema_version": 0, + "attributes": { + "acpi": true, + "agent": [], + "amd_sev": [], + "audio_device": [], + "bios": "seabios", + "boot_order": [ + "virtio0" + ], + "cdrom": [], + "clone": [ + { + "datastore_id": "", + "full": true, + "node_name": "hyper1", + "retries": 1, + "vm_id": 999 + } + ], + "cpu": [ + { + "affinity": "", + "architecture": "", + "cores": 2, + "flags": null, + "hotplugged": 0, + "limit": 0, + "numa": false, + "sockets": 1, + "type": "host", + "units": 1024 + } + ], + "description": "kubernetes worker3", + "disk": [], + "efi_disk": [], + "hook_script_file_id": null, + "hostpci": [], + "id": "1013", + "initialization": [ + { + "datastore_id": "Pool1", + "dns": [], + "interface": "ide2", + "ip_config": [], + "meta_data_file_id": "", + "network_data_file_id": "", + "type": "", + "user_account": [], + "user_data_file_id": "cephfs:snippets/worker3-ignition-user-data", + "vendor_data_file_id": "" + } + ], + "ipv4_addresses": [ + [ + "127.0.0.1" + ], + [ + "10.0.2.103" + ] + ], + "ipv6_addresses": [ + [ + "::1" + ], + [ + "2001:470:7116:2::103", + "fd00:0:0:2::103", + "fe80::be24:11ff:fe3a:10c5" + ] + ], + "keyboard_layout": "en-us", + "kvm_arguments": null, + "mac_addresses": [ + "00:00:00:00:00:00", + "BC:24:11:3A:10:C5" + ], + "machine": null, + "memory": [ + { + "dedicated": 10240, + "floating": 10240, + "hugepages": "", + "keep_hugepages": false, + "shared": 0 + } + ], + "migrate": false, + "name": "worker3", + "network_device": [ + { + "bridge": "vmbr0", + "disconnected": false, + "enabled": true, + "firewall": false, + "mac_address": "BC:24:11:3A:10:C5", + "model": "virtio", + "mtu": 0, + "queues": 0, + "rate_limit": 0, + "trunks": "", + "vlan_id": 0 + } + ], + "network_interface_names": [ + "lo", + "eth0" + ], + "node_name": "hyper3", + "numa": [], + "on_boot": true, + "operating_system": [], + "pool_id": null, + "protection": false, + "reboot": false, + "reboot_after_update": true, + "rng": [], + "scsi_hardware": "virtio-scsi-pci", + "serial_device": [], + "smbios": [], + "started": true, + "startup": [ + { + "down_delay": -1, + "order": 1, + "up_delay": -1 + } + ], + "stop_on_destroy": false, + "tablet_device": true, + "tags": [ + "flatcar", + "kubernetes", + "terraform", + "worker" + ], + "template": false, + "timeout_clone": 1800, + "timeout_create": 1800, + "timeout_migrate": 1800, + "timeout_move_disk": 1800, + "timeout_reboot": 1800, + "timeout_shutdown_vm": 1800, + "timeout_start_vm": 1800, + "timeout_stop_vm": 300, + "tpm_state": [], + "usb": [], + "vga": [], + "virtiofs": [], + "vm_id": 1013, + "watchdog": [] + }, + "sensitive_attributes": [], + "identity_schema_version": 0, + "private": "bnVsbA==", + "dependencies": [ + "data.ct_config.control_plane1_ignition", + "data.ct_config.control_plane2_ignition", + "data.ct_config.control_plane3_ignition", + "data.ct_config.worker3_ignition", + "null_resource.wait_for_cluster_ready", + "null_resource.wait_for_cp1_api", + "null_resource.wait_for_cp2_cp3_api", + "proxmox_virtual_environment_download_file.flatcar_image", + "proxmox_virtual_environment_file.control_plane1_ignition", + "proxmox_virtual_environment_file.control_plane2_ignition", + "proxmox_virtual_environment_file.control_plane3_ignition", + "proxmox_virtual_environment_file.worker3_ignition", + "proxmox_virtual_environment_vm.control_plane1", + "proxmox_virtual_environment_vm.control_plane2", + "proxmox_virtual_environment_vm.control_plane3", + "proxmox_virtual_environment_vm.flatcar_template" + ] + } + ] + }, + { + "mode": "managed", + "type": "proxmox_virtual_environment_vm", + "name": "worker4", + "provider": "provider[\"registry.terraform.io/bpg/proxmox\"]", + "instances": [ + { + "schema_version": 0, + "attributes": { + "acpi": true, + "agent": [], + "amd_sev": [], + "audio_device": [], + "bios": "seabios", + "boot_order": [ + "virtio0" + ], + "cdrom": [], + "clone": [ + { + "datastore_id": "", + "full": true, + "node_name": "hyper1", + "retries": 1, + "vm_id": 999 + } + ], + "cpu": [ + { + "affinity": "", + "architecture": "", + "cores": 2, + "flags": null, + "hotplugged": 0, + "limit": 0, + "numa": false, + "sockets": 1, + "type": "host", + "units": 1024 + } + ], + "description": "kubernetes worker4", + "disk": [], + "efi_disk": [], + "hook_script_file_id": null, + "hostpci": [], + "id": "1014", + "initialization": [ + { + "datastore_id": "Pool1", + "dns": [], + "interface": "ide2", + "ip_config": [], + "meta_data_file_id": "", + "network_data_file_id": "", + "type": "", + "user_account": [], + "user_data_file_id": "cephfs:snippets/worker4-ignition-user-data", + "vendor_data_file_id": "" + } + ], + "ipv4_addresses": [ + [ + "127.0.0.1" + ], + [ + "10.0.2.104" + ] + ], + "ipv6_addresses": [ + [ + "::1" + ], + [ + "2001:470:7116:2::104", + "fd00:0:0:2::104", + "fe80::be24:11ff:fe35:4ae3" + ] + ], + "keyboard_layout": "en-us", + "kvm_arguments": null, + "mac_addresses": [ + "00:00:00:00:00:00", + "BC:24:11:35:4A:E3" + ], + "machine": null, + "memory": [ + { + "dedicated": 12000, + "floating": 12000, + "hugepages": "", + "keep_hugepages": false, + "shared": 0 + } + ], + "migrate": false, + "name": "worker4", + "network_device": [ + { + "bridge": "vmbr0", + "disconnected": false, + "enabled": true, + "firewall": false, + "mac_address": "BC:24:11:35:4A:E3", + "model": "virtio", + "mtu": 0, + "queues": 0, + "rate_limit": 0, + "trunks": "", + "vlan_id": 0 + } + ], + "network_interface_names": [ + "lo", + "eth0" + ], + "node_name": "pbs", + "numa": [], + "on_boot": true, + "operating_system": [], + "pool_id": null, + "protection": false, + "reboot": false, + "reboot_after_update": true, + "rng": [], + "scsi_hardware": "virtio-scsi-pci", + "serial_device": [], + "smbios": [], + "started": true, + "startup": [ + { + "down_delay": -1, + "order": 1, + "up_delay": -1 + } + ], + "stop_on_destroy": false, + "tablet_device": true, + "tags": [ + "flatcar", + "kubernetes", + "terraform", + "worker" + ], + "template": false, + "timeout_clone": 1800, + "timeout_create": 1800, + "timeout_migrate": 1800, + "timeout_move_disk": 1800, + "timeout_reboot": 1800, + "timeout_shutdown_vm": 1800, + "timeout_start_vm": 1800, + "timeout_stop_vm": 300, + "tpm_state": [], + "usb": [], + "vga": [], + "virtiofs": [], + "vm_id": 1014, + "watchdog": [] + }, + "sensitive_attributes": [], + "identity_schema_version": 0, + "private": "bnVsbA==", + "dependencies": [ + "data.ct_config.control_plane1_ignition", + "data.ct_config.control_plane2_ignition", + "data.ct_config.control_plane3_ignition", + "data.ct_config.worker4_ignition", + "null_resource.wait_for_cluster_ready", + "null_resource.wait_for_cp1_api", + "null_resource.wait_for_cp2_cp3_api", + "proxmox_virtual_environment_download_file.flatcar_image", + "proxmox_virtual_environment_file.control_plane1_ignition", + "proxmox_virtual_environment_file.control_plane2_ignition", + "proxmox_virtual_environment_file.control_plane3_ignition", + "proxmox_virtual_environment_file.worker4_ignition", + "proxmox_virtual_environment_vm.control_plane1", + "proxmox_virtual_environment_vm.control_plane2", + "proxmox_virtual_environment_vm.control_plane3", + "proxmox_virtual_environment_vm.flatcar_template" + ] + } + ] } ], "check_results": null diff --git a/terraform/worker3.bu b/terraform/worker3.bu index 65c0585..875209e 100644 --- a/terraform/worker3.bu +++ b/terraform/worker3.bu @@ -203,7 +203,7 @@ systemd: OOMScoreAdjust=-999 [Install] - WantedBy=multi-user. + WantedBy=multi-user.target - name: set-timezone.service enabled: true