diff --git a/terraform/control-plane1.bu b/terraform/control-plane1.bu index 85304b6..f3f3cfc 100644 --- a/terraform/control-plane1.bu +++ b/terraform/control-plane1.bu @@ -154,15 +154,10 @@ storage: contents: source: "http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/velero" - - path: /opt/bin/remote-syslog + - path: /opt/bin/rsyslogd mode: 0755 contents: - source: "http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/remote-syslog" - - - path: /opt/bin/busybox - mode: 0755 - contents: - source: "http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/busybox" + source: "http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/rsyslogd" - path: /etc/systemd/journald.conf.d/10-forward-to-syslog.conf mode: 0644 @@ -271,23 +266,6 @@ storage: systemd: units: - - name: syslog-forward.service - enabled: true - contents: | - [Unit] - Description=Forward journald to central syslog server - Wants=network-online.target - After=network-online.target systemd-journald.service - - [Service] - Type=simple - ExecStart=/opt/bin/busybox syslogd -n -R syslog.undercloud.local -P 514 - Restart=always - RestartSec=2 - - [Install] - WantedBy=multi-user.target - # --- Boot entrypoint: only this target is enabled at boot --- - name: undercloud-bootstrap.target enabled: true diff --git a/terraform/terraform.tfstate b/terraform/terraform.tfstate index 6715e69..0fde4b0 100644 --- a/terraform/terraform.tfstate +++ b/terraform/terraform.tfstate @@ -1,7 +1,7 @@ { "version": 4, "terraform_version": "1.12.2", - "serial": 1294, + "serial": 1362, "lineage": "d92c42be-29f9-bad9-ef9a-3dc952ff5fa5", "outputs": {}, "resources": [ @@ -14,10 +14,10 @@ { "schema_version": 0, "attributes": { - "content": "variant: flatcar\nversion: 1.1.0\n\npasswd:\n users:\n - name: core\n ssh_authorized_keys:\n - \"ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIHHEAlPo3v4U67Y3411pTjIMkQxwlFWdXrBJkSzXenDH flatcar@undercloud\"\n\nstorage:\n directories:\n - path: /opt/bin\n overwrite: true\n mode: 0755\n - path: /opt/cni/bin\n overwrite: true\n mode: 755\n - path: /etc/kubernetes/manifests\n #overwrite: true\n mode: 0755\n - path: /etc/install-calico\n overwrite: true\n mode: 0755\n - path: /var/lib/undercloud-stamps\n mode: 0755\n\n files:\n - path: /etc/hostname\n mode: 0644\n contents:\n inline: |\n control-plane1\n\n - path: /etc/systemd/network/00-eth.network\n mode: 0644\n contents:\n inline: |\n [Match]\n Name=eth*\n \n [Network]\n Address=fd00:0:0:2::91/64\n Address=2001:470:7116:2::91/64\n Gateway=2001:470:7116:2::3\n DNS=fd00:0:0:1::1\n Address=10.0.2.91/24\n Gateway=10.0.2.3\n DNS=10.0.1.1\n Domains=undercloud.local\n IPv6AcceptRA=no\n IPv6PrivacyExtensions=no\n\n - path: /etc/hosts\n mode: 0644\n overwrite: true\n contents:\n inline: |\n 127.0.0.1 localhost\n ::1 localhost ip6-localhost ip6-loopback\n\n 2001:470:7116:2::91 control-plane1.undercloud.local control-plane1\n 2001:470:7116:2::92 control-plane2.undercloud.local control-plane2\n 2001:470:7116:2::93 control-plane3.undercloud.local control-plane3\n 2001:470:7116:2::101 worker1.undercloud.local worker1\n 2001:470:7116:2::102 worker2.undercloud.local worker2\n 2001:470:7116:2::103 worker3.undercloud.local worker3\n\n fd00:0:0:2::91 control-plane1.undercloud.local control-plane1\n fd00:0:0:2::92 control-plane2.undercloud.local control-plane2\n fd00:0:0:2::93 control-plane3.undercloud.local control-plane3\n fd00:0:0:2::101 worker1.undercloud.local worker1\n fd00:0:0:2::102 worker2.undercloud.local worker2\n fd00:0:0:2::103 worker3.undercloud.local worker3\n\n 10.0.2.91 control-plane1.undercloud.local control-plane1\n 10.0.2.92 control-plane2.undercloud.local control-plane2\n 10.0.2.93 control-plane3.undercloud.local control-plane3\n 10.0.2.101 worker1.undercloud.local worker1\n 10.0.2.102 worker2.undercloud.local worker2\n 10.0.2.103 worker3.undercloud.local worker3\n\n - path: /etc/motd\n mode: 0644\n overwrite: true\n contents:\n inline: |\n *******************************************************************\n * AUTHORIZED ACCESS ONLY *\n * *\n * This system is part of a secured infrastructure. *\n * All activities are monitored and logged. *\n * Unauthorized access or misuse is strictly prohibited and *\n * may result in disciplinary and legal action. *\n *******************************************************************\n\n --------------------------------------------------------------------------------\n kubernetes controle plane Node\n\n Manage via:\n kubectl (kubectl)\n calico (calicoctl)\n velero - backup (velero)\n argocd https://argocd-server.argocd.svc.k8aux.undercloud.cf/\n --------------------------------------------------------------------------------\n\n - path: /etc/sysctl.d/99-k8s.conf\n mode: 0644\n contents:\n inline: |\n net.ipv4.ip_forward = 1\n net.ipv6.ip_forward = 1\n net.ipv6.conf.all.forwarding = 1\n net.ipv4.conf.all.forwarding = 1\n net.bridge.bridge-nf-call-iptables = 1\n net.bridge.bridge-nf-call-ip6tables = 1\n net.netfilter.nf_conntrack_max = 1000000\n net.ipv4.conf.all.rp_filter = 0\n net.ipv6.conf.all.disable_ipv6 = 0\n vm.overcommit_memory = 1\n fs.inotify.max_user_watches = 524288\n fs.inotify.max_user_instances = 4096\n kernel.panic = 10\n kernel.panic_on_oops = 1\n\n - path: /etc/flatcar/update.conf\n overwrite: true\n mode: 0420\n contents:\n inline: |\n REBOOT_STRATEGY=off\n\n - path: /opt/bin/kubeadm\n mode: 0755\n contents:\n source: \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubeadm\"\n\n - path: /opt/bin/kubelet\n mode: 0755\n contents:\n source: \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubelet\"\n\n - path: /opt/bin/kubectl\n mode: 0755\n contents:\n source: \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubectl\"\n\n - path: /opt/bin/calicoctl\n mode: 0755\n contents:\n source: \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/calicoctl\"\n\n - path: /opt/bin/velero\n mode: 0755\n contents:\n source: \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/velero\"\n\n - path: /etc/kubernetes/kubeadm-init.yaml\n mode: 0644\n contents:\n inline: |\n apiVersion: kubeadm.k8s.io/v1beta3\n kind: InitConfiguration\n bootstrapTokens:\n - token: \"kvg1hc.t3rewovrps426rof\"\n description: \"default kubeadm bootstrap token\"\n ttl: \"0\"\n nodeRegistration:\n name: control-plane1\n criSocket: unix:///run/containerd/containerd.sock\n kubeletExtraArgs:\n node-ip: \"2001:470:7116:2::91\"\n cluster-dns: \"10.0.91.53,2001:470:7116:f:1::53\"\n volume-plugin-dir: \"/opt/libexec/kubernetes/kubelet-plugins/volume/exec/\"\n localAPIEndpoint:\n advertiseAddress: \"2001:470:7116:2::91\"\n bindPort: 6443\n certificateKey: \"fee7c3e5cfcac7e4774c6efca0464a42d897f30f7300340d6578b5cfb4a3d34b\"\n ---\n apiVersion: kubeadm.k8s.io/v1beta3\n kind: ClusterConfiguration\n controlPlaneEndpoint: \"[fd00:0:0:2::100]:6443\"\n networking:\n podSubnet: \"2001:470:7116:a::/64,10.0.10.0/24\"\n serviceSubnet: \"2001:470:7116:f:1::/108,10.0.91.0/24\"\n dnsDomain: \"k8s.undercloud.local\"\n controllerManager:\n extraArgs:\n flex-volume-plugin-dir: \"/opt/libexec/kubernetes/kubelet-plugins/volume/exec/\"\n bind-address: '::'\n apiServer:\n extraArgs:\n enable-aggregator-routing: \"true\"\n proxy-client-cert-file: /etc/kubernetes/pki/front-proxy-client.crt\n proxy-client-key-file: /etc/kubernetes/pki/front-proxy-client.key\n requestheader-client-ca-file: /etc/kubernetes/pki/front-proxy-ca.crt\n requestheader-allowed-names: front-proxy-client\n requestheader-extra-headers-prefix: X-Remote-Extra-\n requestheader-group-headers: X-Remote-Group\n requestheader-username-headers: X-Remote-User\n\n ---\n apiVersion: kubelet.config.k8s.io/v1beta1\n kind: KubeletConfiguration\n address: \"::\"\n healthzBindAddress: \"::\"\n clusterDomain: \"k8s.undercloud.local\"\n clusterDNS:\n - \"2001:470:7116:f:1::53\"\n - \"10.0.91.53\"\n volumePluginDir: /opt/libexec/kubernetes/kubelet-plugins/volume/exec\n cgroupDriver: \"systemd\"\n authentication:\n anonymous:\n enabled: true\n webhook:\n enabled: true\n authorization:\n mode: Webhook\n\n - path: /etc/kubernetes/addons/kube-dns-fixed-svc.yaml\n mode: 0644\n contents:\n inline: |\n apiVersion: v1\n kind: Service\n metadata:\n name: kube-dns\n namespace: kube-system\n labels:\n k8s-app: kube-dns\n spec:\n type: ClusterIP\n ipFamilyPolicy: RequireDualStack\n ipFamilies: [IPv6, IPv4]\n clusterIP: 2001:470:7116:f:1::53\n clusterIPs:\n - 2001:470:7116:f:1::53\n - 10.0.91.53\n ports:\n - name: dns\n port: 53\n protocol: UDP\n targetPort: 53\n - name: dns-tcp\n port: 53\n protocol: TCP\n targetPort: 53\n - name: metrics\n port: 9153\n protocol: TCP\n targetPort: 9153\n selector:\n k8s-app: kube-dns\n\n\nsystemd:\n units:\n - name: modules-load.service\n enabled: true\n contents: |\n [Unit]\n Description=Load necessary kernel modules\n Before=containerd.service kubeadm-init.service\n\n [Service]\n Type=oneshot\n\n ExecStart=/usr/bin/modprobe br_netfilter\n ExecStart=/usr/bin/modprobe overlay\n RemainAfterExit=yes\n\n [Install]\n WantedBy=multi-user.target\n\n - name: systemd-networkd-wait-online.service\n enabled: true\n\n - name: containerd.service\n enabled: true\n contents: |\n [Unit]\n Description=containerd container runtime\n After=network.target modules-load.service\n\n [Service]\n #StandardOutput=journal+console\n #StandardError=journal+console\n\n ExecStart=/usr/bin/containerd\n Restart=always\n RestartSec=5\n Delegate=yes\n KillMode=process\n OOMScoreAdjust=-999\n\n [Install]\n WantedBy=multi-user.target\n\n - name: set-timezone.service\n enabled: true\n contents: |\n [Unit]\n Description=Set Timezone\n After=network-online.target\n Wants=network-online.target\n [Service]\n StandardOutput=journal+console\n StandardError=journal+console\n\n ExecStart=/bin/sh -c 'echo \"setting timezone to Europe/Berlin\"'\n StandardOutput=journal+console\n StandardError=journal+console\n Type=oneshot\n Restart=on-failure\n ExecStart=/usr/bin/timedatectl set-timezone Europe/Berlin\n ExecStart=/usr/bin/timedatectl set-ntp true \n [Install]\n WantedBy=multi-user.target\n\n - name: kubelet.service\n enabled: true\n contents: |\n [Unit]\n Description=kubelet, the Kubernetes Node Agent\n Documentation=https://kubernets.io/docs/home\n Wants=network-online.target\n After=network-online.target\n [Service]\n #StandardOutput=journal+console\n #StandardError=journal+console\n\n #EnvironmentFile=/run/metadata/coreos\n Environment=\"KUBELET_KUBECONFIG_ARGS=--bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf\"\n Environment=\"KUBELET_CONFIG_ARGS=--config=/var/lib/kubelet/config.yaml\"\n # This is a file that \"kubeadm init\" and \"kubeadm join\" generates at runtime, populating the KUBELET_KUBEADM_ARGS variable dynamically\n EnvironmentFile=-/var/lib/kubelet/kubeadm-flags.env\n ExecStart=/opt/bin/kubelet $KUBELET_KUBECONFIG_ARGS $KUBELET_CONFIG_ARGS $KUBELET_KUBEADM_ARGS $KUBELET_EXTRA_ARGS\n Restart=always\n StartLimitInterval=0\n RestartSec=10\n [Install]\n WantedBy=multi-user.target\n\n - name: kubeadm-init.service\n enabled: true\n contents: |\n [Unit]\n Description=Kubeadm Init Cluster\n After=network-online.target containerd.service kubelet.service\n Wants=network-online.target\n ConditionPathExists=!/etc/kubernetes/kubelet.conf\n\n [Service]\n Type=oneshot\n StandardOutput=journal+console\n StandardError=journal+console\n\n ExecStart=/bin/sh -c 'echo \"kubeadm-init.service started...\"'\n\n # Environment\n Environment=KUBECONFIG=/etc/kubernetes/admin.conf\n Environment=DATASTORE_TYPE=kubernetes\n Environment=PATH=/usr/bin/:/usr/sbin:/opt/bin:/opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent-uds/\n \n ExecStartPre=/bin/sleep 30s\n ExecStart=/bin/sh -c 'echo \"running kubeadm init...\"'\n ExecStart=/opt/bin/kubeadm init --upload-certs --config=/etc/kubernetes/kubeadm-init.yaml\n \n # copy files for kubectl\n ExecStart=/bin/sh -c 'echo \"copying files (admin.conf) to core home folder.\"'\n ExecStartPost=/usr/bin/mkdir -p /home/core/.kube\n ExecStartPost=/usr/bin/cp -i /etc/kubernetes/admin.conf /home/core/.kube/config\n ExecStartPost=/usr/bin/chown core:core /home/core/.kube/config\n \n ExecStartPost=/usr/bin/systemctl disable kubeadm-init.service\n Restart=on-failure\n RestartSec=120s\n [Install]\n WantedBy=multi-user.target\n - name: install-calico.service\n enabled: true\n contents: |\n [Unit]\n Wants=kubeadm-init.service\n After=kubeadm-init.service\n ConditionPathExists=!/var/lib/undercloud-stamps/install-calico.done\n\n\n [Service]\n StandardOutput=journal+console\n StandardError=journal+console\n\n ExecStart=/bin/sh -c 'echo \"install.calico.service started...\"'\n Environment=KUBECONFIG=/etc/kubernetes/admin.conf\n Environment=DATASTORE_TYPE=kubernetes\n Environment=PATH=/usr/bin/:/usr/sbin:/opt/bin\n Type=oneshot\n StandardOutput=journal+console\n StandardError=journal+console\n ExecStart=/bin/sh -c 'echo \"witing 30s...\"'\n ExecStart=/bin/sleep 30s\n ExecStart=/bin/sh -c 'echo \"create calico namespace...\"'\n ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/namespace.yaml\n ExecStart=/bin/sh -c 'echo \"install tigera operator...\"'\n ExecStart=-/opt/bin/kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/operator-crds.yaml\n ExecStart=-/opt/bin/kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/tigera-operator.yaml\n ExecStart=/bin/sh -c 'echo \"witing 60s...\"'\n ExecStart=/bin/sleep 60s\n ExecStart=/bin/sh -c 'echo \"witing for tigera operator... (20mini max)\"'\n ExecStart=/opt/bin/kubectl wait deployment -n tigera-operator tigera-operator --for condition=Available=True --timeout=1200s\n ExecStart=/bin/sh -c 'echo \"create clico custom ressources...\"'\n ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/custom-resources.yaml\n \n ExecStart=/bin/sh -c 'echo \"witing 3m..\"'\n ExecStart=/bin/sleep 3m\n #ExecStart=/bin/sh -c 'echo \"apply calico (calico-apiserver)...\"'\n #ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/calico.yaml\n #ExecStart=/bin/sh -c 'echo \"witing 1m...\"'\n #ExecStart=/bin/sleep 2m\n ExecStart=/bin/sh -c 'echo \"witing calico-apiserver... (20mini max)\"'\n ExecStart=/opt/bin/kubectl wait deployment -n calico-apiserver calico-apiserver --for condition=Available=True --timeout=1200s\n ExecStart=/bin/sh -c 'echo \"witing 120s...\"'\n ExecStart=/bin/sleep 2m\n ExecStart=/bin/sh -c 'echo \"apply calico-peers...\"'\n ExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/calico-peer.yaml\n ExecStart=/bin/sh -c 'echo \"witing 60s...\"'\n ExecStart=/bin/sleep 1m\n ExecStart=/bin/sh -c 'echo \"apply calico-ippools...\"'\n ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/ippools.yaml\n \n #ExecStart=/bin/sh -c 'echo \"witing for whisker..\"'\n #ExecStart=/opt/bin/kubectl wait deployment -n calico-system whisker --for condition=Available=True --timeout=1200s\n #ExecStart=/bin/sh -c 'echo \"port-forward -n calico-system service/whisker 8081:8081\"'\n #ExecStart=/opt/bin/kubectl port-forward -n calico-system service/whisker 8081:8081\n \n ExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-calico.done\n ExecStart=/usr/bin/systemctl disable install-calico.service\n #RemainAfterExit=true\n Restart=on-failure\n RestartSec=120s\n [Install]\n WantedBy=multi-user.target\n - name: install-ceph.service\n enabled: true\n contents: |\n [Unit]\n Wants=kubeadm-init.service\n After=kubeadm-init.service\n ConditionPathExists=!/var/lib/undercloud-stamps/install-ceph.done\n\n [Service]\n StandardOutput=journal+console\n StandardError=journal+console\n\n ExecStart=/bin/sh -c 'echo \"install.ceph.service started...\"'\n Environment=KUBECONFIG=/etc/kubernetes/admin.conf\n Environment=DATASTORE_TYPE=kubernetes\n Environment=PATH=/usr/bin/:/usr/sbin:/opt/bin\n Type=oneshot\n \n StandardOutput=journal+console\n StandardError=journal+console\n ExecStart=/bin/sh -c 'echo \"witing 30s...\"'\n ExecStart=/bin/sleep 30s\n ExecStart=/bin/sh -c 'echo \"create ceph namespace...\"'\n ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/namespace.yaml\n \n ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-nodeplugin-rbac.yaml\n ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-provisioner-rbac.yaml\n\n ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/secrets.yaml\n\n ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/ceph-conf.yaml\n ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-cephfsplugin-provisioner.yaml\n ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-cephfsplugin.yaml\n ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-config-map.yaml\n ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-encryption-kms-config.yaml\n \n ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csidriver.yaml\n \n ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/storage-class.yaml\n \n ExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-ceph.done\n ExecStart=/usr/bin/systemctl disable install-ceph.service\n #RemainAfterExit=true\n Restart=on-failure\n RestartSec=120s\n [Install]\n WantedBy=multi-user.target\n - name: install-gitea.service\n enabled: true\n contents: |\n [Unit]\n Wants=install-ceph.service\n After=install-ceph.service\n ConditionPathExists=!/var/lib/undercloud-stamps/install-gitea.done\n \n [Service]\n StandardOutput=journal+console\n StandardError=journal+console\n Environment=KUBECONFIG=/etc/kubernetes/admin.conf\n Environment=DATASTORE_TYPE=kubernetes\n Environment=PATH=/usr/bin/:/usr/sbin:/opt/bin\n Type=oneshot\n \n ExecStart=/opt/bin/kubectl wait deployment -n ceph csi-cephfsplugin-provisioner --for condition=Available=True --timeout=1200s\n ExecStart=/bin/sleep 4m\n ExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/namespace.yaml\n ExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/secrets.yaml\n ExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/db.yaml\n ExecStart=/bin/sleep 60s\n ExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/adminer.yaml\n ExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/gitea.yaml\n ExecStart=/bin/sleep 3m\n ExecStart=/opt/bin/kubectl exec deploy/gitea -n gitea -- /bin/startup.sh\n\n ExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-gitea.done\n ExecStart=/usr/bin/systemctl disable install-gitea.service\n Restart=on-failure\n RestartSec=120s\n [Install]\n WantedBy=multi-user.target\n - name: install-argocd.service\n enabled: true\n contents: |\n [Unit]\n Wants=install-calico.service\n After=install-calico.service\n ConditionPathExists=!/var/lib/undercloud-stamps/install-argocd.done\n [Service]\n StandardOutput=journal+console\n StandardError=journal+console\n\n Environment=KUBECONFIG=/etc/kubernetes/admin.conf\n Environment=DATASTORE_TYPE=kubernetes\n Environment=PATH=/usr/bin/:/usr/sbin:/opt/bin\n Type=oneshot\n\n ExecStart=/opt/bin/kubectl wait deployment -n kube-system coredns --for condition=Available=True --timeout=600s\n \n ExecStart=/bin/sleep 1m\n ExecStart=/opt/bin/kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/namespace.yaml\n ExecStart=/opt/bin/kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/install.yaml\n ExecStart=/opt/bin/kubectl wait deployment -n argocd argocd-server --for condition=Available=True --timeout=600s\n\n ExecStart=/opt/bin/kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/repo.yaml\n ExecStart=/bin/sleep 10s\n ExecStart=/opt/bin/kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/apps.yaml\n\n ExecStart=/bin/sleep 10s\n ExecStart=/opt/bin/kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/ingress.yaml\n\n\n #ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/repos/k8aux-bootstrap.yaml\n #ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/repos/k8aux-apps.yaml\n #ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/apps/argocd.yaml\n #ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/apps/calico.yaml\n #ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/apps/rook-ceph.yaml\n #ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/apps/gitea.yaml\n #ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/argocd/argocd-secret.yaml\n ##ExecStart=/bin/sleep 10m\n #ExecStart=/opt/bin/kubectl wait deployment -n gitea gitea --for condition=Available=True --timeout=4800s\n #ExecStart=/bin/sleep 10m\n #ExecStart=/opt/bin/kubectl apply -n argocd -f http://gitea.gitea.svc.k8aux.undercloud.cf:3000/undercloud/k8aux-apps/raw/branch/main/app-of-apps/app-of-apps.yaml\n \n ExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-argocd.done\n ExecStart=/usr/bin/systemctl disable install-argocd.service\n Restart=on-failure\n RestartSec=120s\n [Install]\n WantedBy=multi-user.target\n - name: pin-service-ips.service\n enabled: true\n contents: |\n [Unit]\n Description=Pin fixed dual-stack ClusterIPs for kube-dns, argocd-server and whisker\n After=install-argocd.service install-calico.service kubeadm-init.service network-online.target\n Wants=install-argocd.service install-calico.service kubeadm-init.service network-online.target\n ConditionPathExists=!/var/lib/undercloud-stamps/pin-service-ips.done\n\n\n [Service]\n Type=oneshot\n StandardOutput=journal+console\n StandardError=journal+console\n Environment=KUBECONFIG=/etc/kubernetes/admin.conf\n Environment=PATH=/usr/bin:/usr/sbin:/opt/bin\n ExecStart=/bin/sh -eu -c '\\\n echo \"[pin-service-ips] waiting for API...\" ; \\\n for i in $(seq 1 120); do kubectl get --raw=/readyz \u003e/dev/null 2\u003e\u00261 \u0026\u0026 break; sleep 2; done ; \\\n echo \"[pin-service-ips] ensure namespaces exist...\" ; \\\n kubectl get ns kube-system \u003e/dev/null ; \\\n kubectl get ns argocd \u003e/dev/null 2\u003e\u00261 || kubectl create ns argocd ; \\\n kubectl get ns calico-system \u003e/dev/null ; \\\n echo \"[pin-service-ips] wait for coredns/argocd readiness (best effort)...\" ; \\\n kubectl -n kube-system wait deploy coredns --for=condition=Available=True --timeout=300s || true ; \\\n kubectl -n argocd wait deploy argocd-server --for=condition=Available=True --timeout=600s || true ; \\\n echo \"[pin-service-ips] replace Services with fixed ClusterIPs...\" ; \\\n kubectl -n kube-system delete svc kube-dns --ignore-not-found ; \\\n kubectl apply -f /etc/kubernetes/addons/kube-dns-fixed-svc.yaml ; \\\n kubectl -n argocd delete svc argocd-server --ignore-not-found ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/service.yaml ; \\\n kubectl -n calico-system delete svc whisker --ignore-not-found || true ; \\\n kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/whisker.yaml || true ; \\\n echo \"[pin-service-ips] done.\" \\\n '\n\n ExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/pin-service-ips.done\n Restart=on-failure\n RestartSec=120s\n [Install]\n WantedBy=multi-user.target\n\n", - "id": "1444300035", + "content": "variant: flatcar\nversion: 1.1.0\n\npasswd:\n users:\n - name: core\n ssh_authorized_keys:\n - \"ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIHHEAlPo3v4U67Y3411pTjIMkQxwlFWdXrBJkSzXenDH flatcar@undercloud\"\n\nstorage:\n directories:\n - path: /opt/bin\n overwrite: true\n mode: 0755\n - path: /opt/cni/bin\n overwrite: true\n mode: 755\n - path: /etc/kubernetes/manifests\n #overwrite: true\n mode: 0755\n - path: /etc/install-calico\n overwrite: true\n mode: 0755\n - path: /var/lib/undercloud-stamps\n mode: 0755\n\n files:\n - path: /etc/hostname\n mode: 0644\n contents:\n inline: |\n control-plane1\n\n - path: /etc/systemd/network/00-eth.network\n mode: 0644\n contents:\n inline: |\n [Match]\n Name=eth*\n \n [Network]\n Address=fd00:0:0:2::91/64\n Address=2001:470:7116:2::91/64\n Gateway=2001:470:7116:2::3\n DNS=fd00:0:0:1::1\n Address=10.0.2.91/24\n Gateway=10.0.2.3\n DNS=10.0.1.1\n Domains=undercloud.local\n IPv6AcceptRA=no\n IPv6PrivacyExtensions=no\n\n - path: /etc/hosts\n mode: 0644\n overwrite: true\n contents:\n inline: |\n 127.0.0.1 localhost\n ::1 localhost ip6-localhost ip6-loopback\n\n 2001:470:7116:2::91 control-plane1.undercloud.local control-plane1\n 2001:470:7116:2::92 control-plane2.undercloud.local control-plane2\n 2001:470:7116:2::93 control-plane3.undercloud.local control-plane3\n 2001:470:7116:2::101 worker1.undercloud.local worker1\n 2001:470:7116:2::102 worker2.undercloud.local worker2\n 2001:470:7116:2::103 worker3.undercloud.local worker3\n\n fd00:0:0:2::91 control-plane1.undercloud.local control-plane1\n fd00:0:0:2::92 control-plane2.undercloud.local control-plane2\n fd00:0:0:2::93 control-plane3.undercloud.local control-plane3\n fd00:0:0:2::101 worker1.undercloud.local worker1\n fd00:0:0:2::102 worker2.undercloud.local worker2\n fd00:0:0:2::103 worker3.undercloud.local worker3\n\n 10.0.2.91 control-plane1.undercloud.local control-plane1\n 10.0.2.92 control-plane2.undercloud.local control-plane2\n 10.0.2.93 control-plane3.undercloud.local control-plane3\n 10.0.2.101 worker1.undercloud.local worker1\n 10.0.2.102 worker2.undercloud.local worker2\n 10.0.2.103 worker3.undercloud.local worker3\n\n - path: /etc/motd\n mode: 0644\n overwrite: true\n contents:\n inline: |\n *******************************************************************\n * AUTHORIZED ACCESS ONLY *\n * *\n * This system is part of a secured infrastructure. *\n * All activities are monitored and logged. *\n * Unauthorized access or misuse is strictly prohibited and *\n * may result in disciplinary and legal action. *\n *******************************************************************\n\n --------------------------------------------------------------------------------\n kubernetes controle plane Node\n\n Manage via:\n kubectl (kubectl)\n calico (calicoctl)\n velero - backup (velero)\n argocd https://argocd-server.argocd.svc.k8aux.undercloud.cf/\n --------------------------------------------------------------------------------\n\n - path: /etc/sysctl.d/99-k8s.conf\n mode: 0644\n contents:\n inline: |\n net.ipv4.ip_forward = 1\n net.ipv6.ip_forward = 1\n net.ipv6.conf.all.forwarding = 1\n net.ipv4.conf.all.forwarding = 1\n net.bridge.bridge-nf-call-iptables = 1\n net.bridge.bridge-nf-call-ip6tables = 1\n net.netfilter.nf_conntrack_max = 1000000\n net.ipv4.conf.all.rp_filter = 0\n net.ipv6.conf.all.disable_ipv6 = 0\n vm.overcommit_memory = 1\n fs.inotify.max_user_watches = 524288\n fs.inotify.max_user_instances = 4096\n kernel.panic = 10\n kernel.panic_on_oops = 1\n\n - path: /etc/flatcar/update.conf\n overwrite: true\n mode: 0420\n contents:\n inline: |\n REBOOT_STRATEGY=off\n\n - path: /opt/bin/kubeadm\n mode: 0755\n contents:\n source: \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubeadm\"\n\n - path: /opt/bin/kubelet\n mode: 0755\n contents:\n source: \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubelet\"\n\n - path: /opt/bin/kubectl\n mode: 0755\n contents:\n source: \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubectl\"\n\n - path: /opt/bin/calicoctl\n mode: 0755\n contents:\n source: \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/calicoctl\"\n\n - path: /opt/bin/velero\n mode: 0755\n contents:\n source: \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/velero\"\n \n - path: /opt/bin/busybox\n mode: 0755\n contents:\n source: \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/busybox\"\n\n - path: /etc/systemd/journald.conf.d/10-forward-to-syslog.conf\n mode: 0644\n contents:\n inline: |\n [Journal]\n ForwardToSyslog=yes\n\n - path: /etc/kubernetes/kubeadm-init.yaml\n mode: 0644\n contents:\n inline: |\n apiVersion: kubeadm.k8s.io/v1beta3\n kind: InitConfiguration\n bootstrapTokens:\n - token: \"kvg1hc.t3rewovrps426rof\"\n description: \"default kubeadm bootstrap token\"\n ttl: \"0\"\n nodeRegistration:\n name: control-plane1\n criSocket: unix:///run/containerd/containerd.sock\n kubeletExtraArgs:\n node-ip: \"2001:470:7116:2::91\"\n cluster-dns: \"10.0.91.53,2001:470:7116:f:1::53\"\n volume-plugin-dir: \"/opt/libexec/kubernetes/kubelet-plugins/volume/exec/\"\n localAPIEndpoint:\n advertiseAddress: \"2001:470:7116:2::91\"\n bindPort: 6443\n certificateKey: \"fee7c3e5cfcac7e4774c6efca0464a42d897f30f7300340d6578b5cfb4a3d34b\"\n ---\n apiVersion: kubeadm.k8s.io/v1beta3\n kind: ClusterConfiguration\n controlPlaneEndpoint: \"[fd00:0:0:2::100]:6443\"\n networking:\n podSubnet: \"2001:470:7116:a::/64,10.0.10.0/24\"\n serviceSubnet: \"2001:470:7116:f:1::/108,10.0.91.0/24\"\n dnsDomain: \"k8s.undercloud.local\"\n controllerManager:\n extraArgs:\n flex-volume-plugin-dir: \"/opt/libexec/kubernetes/kubelet-plugins/volume/exec/\"\n bind-address: '::'\n apiServer:\n extraArgs:\n enable-aggregator-routing: \"true\"\n proxy-client-cert-file: /etc/kubernetes/pki/front-proxy-client.crt\n proxy-client-key-file: /etc/kubernetes/pki/front-proxy-client.key\n requestheader-client-ca-file: /etc/kubernetes/pki/front-proxy-ca.crt\n requestheader-allowed-names: front-proxy-client\n requestheader-extra-headers-prefix: X-Remote-Extra-\n requestheader-group-headers: X-Remote-Group\n requestheader-username-headers: X-Remote-User\n\n ---\n apiVersion: kubelet.config.k8s.io/v1beta1\n kind: KubeletConfiguration\n address: \"::\"\n healthzBindAddress: \"::\"\n clusterDomain: \"k8s.undercloud.local\"\n clusterDNS:\n - \"2001:470:7116:f:1::53\"\n - \"10.0.91.53\"\n volumePluginDir: /opt/libexec/kubernetes/kubelet-plugins/volume/exec\n cgroupDriver: \"systemd\"\n authentication:\n anonymous:\n enabled: true\n webhook:\n enabled: true\n authorization:\n mode: Webhook\n\n - path: /etc/kubernetes/addons/kube-dns-fixed-svc.yaml\n mode: 0644\n contents:\n inline: |\n apiVersion: v1\n kind: Service\n metadata:\n name: kube-dns\n namespace: kube-system\n labels:\n k8s-app: kube-dns\n spec:\n type: ClusterIP\n ipFamilyPolicy: RequireDualStack\n ipFamilies: [IPv6, IPv4]\n clusterIP: 2001:470:7116:f:1::53\n clusterIPs:\n - 2001:470:7116:f:1::53\n - 10.0.91.53\n ports:\n - name: dns\n port: 53\n protocol: UDP\n targetPort: 53\n - name: dns-tcp\n port: 53\n protocol: TCP\n targetPort: 53\n - name: metrics\n port: 9153\n protocol: TCP\n targetPort: 9153\n selector:\n k8s-app: kube-dns\n\nsystemd:\n units:\n - name: syslog-forward.service\n enabled: true\n contents: |\n [Unit]\n Description=Forward journald to central syslog server\n Wants=network-online.target\n After=network-online.target systemd-journald.service\n\n [Service]\n Type=simple\n ExecStart=/opt/bin/busybox syslogd -n -R syslog.undercloud.local -P 514\n Restart=always\n RestartSec=2\n\n [Install]\n WantedBy=multi-user.target\n\n # --- Boot entrypoint: only this target is enabled at boot ---\n - name: undercloud-bootstrap.target\n enabled: true\n contents: |\n [Unit]\n Description=Undercloud Bootstrap Chain\n Wants=network-online.target\n After=network-online.target\n # Start the chain entry\n Wants=containerd.service kubelet.service kubeadm-init.service\n After=containerd.service kubelet.service kubeadm-init.service\n\n [Install]\n WantedBy=multi-user.target\n\n - name: modules-load.service\n enabled: true\n contents: |\n [Unit]\n Description=Load necessary kernel modules\n Before=containerd.service kubeadm-init.service\n\n [Service]\n Type=oneshot\n ExecStart=/usr/bin/modprobe br_netfilter\n ExecStart=/usr/bin/modprobe overlay\n RemainAfterExit=yes\n\n [Install]\n WantedBy=multi-user.target\n\n - name: systemd-networkd-wait-online.service\n enabled: true\n\n - name: containerd.service\n enabled: true\n contents: |\n [Unit]\n Description=containerd container runtime\n After=network.target modules-load.service\n Wants=modules-load.service\n\n [Service]\n ExecStart=/usr/bin/containerd\n Restart=always\n RestartSec=5\n Delegate=yes\n KillMode=process\n OOMScoreAdjust=-999\n\n [Install]\n WantedBy=multi-user.target\n\n - name: set-timezone.service\n enabled: true\n contents: |\n [Unit]\n Description=Set Timezone\n After=network-online.target\n Wants=network-online.target\n\n [Service]\n Type=oneshot\n StandardOutput=journal+console\n StandardError=journal+console\n ExecStart=/bin/sh -c 'echo \"setting timezone to Europe/Berlin\"'\n ExecStart=/usr/bin/timedatectl set-timezone Europe/Berlin\n ExecStart=/usr/bin/timedatectl set-ntp true\n\n [Install]\n WantedBy=multi-user.target\n\n - name: kubelet.service\n enabled: true\n contents: |\n [Unit]\n Description=kubelet, the Kubernetes Node Agent\n Documentation=https://kubernetes.io/docs/home\n Wants=network-online.target\n After=network-online.target containerd.service\n Requires=containerd.service\n\n [Service]\n Environment=\"KUBELET_KUBECONFIG_ARGS=--bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf\"\n Environment=\"KUBELET_CONFIG_ARGS=--config=/var/lib/kubelet/config.yaml\"\n EnvironmentFile=-/var/lib/kubelet/kubeadm-flags.env\n ExecStart=/opt/bin/kubelet $KUBELET_KUBECONFIG_ARGS $KUBELET_CONFIG_ARGS $KUBELET_KUBEADM_ARGS $KUBELET_EXTRA_ARGS\n Restart=always\n StartLimitInterval=0\n RestartSec=10\n\n [Install]\n WantedBy=multi-user.target\n\n # --- Chain step 1 ---\n - name: kubeadm-init.service\n enabled: false\n contents: |\n [Unit]\n Description=Kubeadm Init Cluster\n Wants=network-online.target\n After=network-online.target containerd.service kubelet.service\n Requires=containerd.service kubelet.service\n ConditionPathExists=!/etc/kubernetes/kubelet.conf\n\n [Service]\n Type=oneshot\n StandardOutput=journal+console\n StandardError=journal+console\n\n Environment=KUBECONFIG=/etc/kubernetes/admin.conf\n Environment=DATASTORE_TYPE=kubernetes\n Environment=PATH=/usr/bin/:/usr/sbin:/opt/bin:/opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent-uds/\n\n ExecStart=/bin/sh -eu -c '\\\n echo \"[kubeadm-init] started...\" ; \\\n echo \"[kubeadm-init] waiting for containerd socket...\" ; \\\n for i in $(seq 1 60); do test -S /run/containerd/containerd.sock \u0026\u0026 break; sleep 1; done ; \\\n echo \"[kubeadm-init] running kubeadm init...\" ; \\\n /opt/bin/kubeadm init --upload-certs --config=/etc/kubernetes/kubeadm-init.yaml ; \\\n echo \"[kubeadm-init] copying kubeconfig to core...\" ; \\\n mkdir -p /home/core/.kube ; \\\n cp -f /etc/kubernetes/admin.conf /home/core/.kube/config ; \\\n chown core:core /home/core/.kube/config ; \\\n echo \"[kubeadm-init] done.\" \\\n '\n\n # strictly start next step (serialization)\n ExecStartPost=/usr/bin/systemctl start install-calico.service\n\n [Install]\n WantedBy=undercloud-bootstrap.target\n\n # --- Chain step 2 ---\n - name: install-calico.service\n enabled: false\n contents: |\n [Unit]\n Description=Install Calico\n Requires=kubeadm-init.service\n After=kubeadm-init.service\n ConditionPathExists=!/var/lib/undercloud-stamps/install-calico.done\n\n [Service]\n Type=oneshot\n StandardOutput=journal+console\n StandardError=journal+console\n Environment=KUBECONFIG=/etc/kubernetes/admin.conf\n Environment=DATASTORE_TYPE=kubernetes\n Environment=PATH=/usr/bin/:/usr/sbin:/opt/bin\n\n ExecStart=/bin/sh -eu -c '\\\n echo \"[calico] waiting for API /readyz...\" ; \\\n for i in $(seq 1 180); do kubectl get --raw=/readyz \u003e/dev/null 2\u003e\u00261 \u0026\u0026 break; sleep 2; done ; \\\n echo \"[calico] create namespace + operator...\" ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/namespace.yaml ; \\\n kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/operator-crds.yaml || true ; \\\n kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/tigera-operator.yaml || true ; \\\n echo \"[calico] wait for tigera-operator...\" ; \\\n kubectl wait deployment -n tigera-operator tigera-operator --for condition=Available=True --timeout=1200s ; \\\n echo \"[calico] apply custom resources...\" ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/custom-resources.yaml ; \\\n echo \"[calico] wait for calico-apiserver...\" ; \\\n kubectl wait deployment -n calico-apiserver calico-apiserver --for condition=Available=True --timeout=1200s ; \\\n echo \"[calico] apply peers + pools...\" ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/calico-peer.yaml ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/ippools.yaml ; \\\n echo \"[calico] done.\" \\\n '\n\n ExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-calico.done\n ExecStartPost=/usr/bin/systemctl start install-ceph.service\n\n [Install]\n WantedBy=undercloud-bootstrap.target\n\n # --- Chain step 3 ---\n - name: install-ceph.service\n enabled: false\n contents: |\n [Unit]\n Description=Install Ceph CSI\n Requires=install-calico.service\n After=install-calico.service\n ConditionPathExists=!/var/lib/undercloud-stamps/install-ceph.done\n\n [Service]\n Type=oneshot\n StandardOutput=journal+console\n StandardError=journal+console\n Environment=KUBECONFIG=/etc/kubernetes/admin.conf\n Environment=DATASTORE_TYPE=kubernetes\n Environment=PATH=/usr/bin/:/usr/sbin:/opt/bin\n\n ExecStart=/bin/sh -eu -c '\\\n echo \"[ceph] waiting for API /readyz...\" ; \\\n for i in $(seq 1 180); do kubectl get --raw=/readyz \u003e/dev/null 2\u003e\u00261 \u0026\u0026 break; sleep 2; done ; \\\n echo \"[ceph] apply manifests...\" ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/namespace.yaml ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-nodeplugin-rbac.yaml ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-provisioner-rbac.yaml ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/secrets.yaml ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/ceph-conf.yaml ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-cephfsplugin-provisioner.yaml ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-cephfsplugin.yaml ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-config-map.yaml ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-encryption-kms-config.yaml ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csidriver.yaml ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/storage-class.yaml ; \\\n echo \"[ceph] done.\" \\\n '\n\n ExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-ceph.done\n ExecStartPost=/usr/bin/systemctl start install-gitea.service\n\n [Install]\n WantedBy=undercloud-bootstrap.target\n\n # --- Chain step 4 ---\n - name: install-gitea.service\n enabled: false\n contents: |\n [Unit]\n Description=Install Gitea\n Requires=install-ceph.service\n After=install-ceph.service\n ConditionPathExists=!/var/lib/undercloud-stamps/install-gitea.done\n\n [Service]\n Type=oneshot\n StandardOutput=journal+console\n StandardError=journal+console\n Environment=KUBECONFIG=/etc/kubernetes/admin.conf\n Environment=DATASTORE_TYPE=kubernetes\n Environment=PATH=/usr/bin/:/usr/sbin:/opt/bin\n\n ExecStart=/bin/sh -eu -c '\\\n echo \"[gitea] wait for ceph provisioner...\" ; \\\n kubectl wait deployment -n ceph csi-cephfsplugin-provisioner --for condition=Available=True --timeout=1200s ; \\\n echo \"[gitea] apply manifests...\" ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/namespace.yaml ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/secrets.yaml ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/db.yaml ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/adminer.yaml ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/gitea.yaml ; \\\n echo \"[gitea] wait for gitea deployment...\" ; \\\n kubectl -n gitea wait deployment gitea --for=condition=Available=True --timeout=1200s ; \\\n echo \"[gitea] run startup...\" ; \\\n kubectl exec deploy/gitea -n gitea -- /bin/startup.sh ; \\\n echo \"[gitea] done.\" \\\n '\n\n ExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-gitea.done\n ExecStartPost=/usr/bin/systemctl start install-argocd.service\n\n [Install]\n WantedBy=undercloud-bootstrap.target\n\n # --- Chain step 5 ---\n - name: install-argocd.service\n enabled: false\n contents: |\n [Unit]\n Description=Install ArgoCD\n Requires=install-calico.service install-gitea.service\n After=install-calico.service install-gitea.service\n ConditionPathExists=!/var/lib/undercloud-stamps/install-argocd.done\n\n [Service]\n Type=oneshot\n StandardOutput=journal+console\n StandardError=journal+console\n Environment=KUBECONFIG=/etc/kubernetes/admin.conf\n Environment=DATASTORE_TYPE=kubernetes\n Environment=PATH=/usr/bin/:/usr/sbin:/opt/bin\n\n ExecStart=/bin/sh -eu -c '\\\n echo \"[argocd] wait for coredns...\" ; \\\n kubectl -n kube-system wait deploy coredns --for=condition=Available=True --timeout=1200s ; \\\n echo \"[argocd] install...\" ; \\\n kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/namespace.yaml ; \\\n kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/install.yaml ; \\\n kubectl -n argocd wait deploy argocd-server --for=condition=Available=True --timeout=1200s ; \\\n kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/repo.yaml ; \\\n kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/apps.yaml ; \\\n kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/ingress.yaml ; \\\n echo \"[argocd] done.\" \\\n '\n\n ExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-argocd.done\n ExecStartPost=/usr/bin/systemctl start pin-service-ips.service\n\n [Install]\n WantedBy=undercloud-bootstrap.target\n\n # --- Chain step 6 (final) ---\n - name: pin-service-ips.service\n enabled: false\n contents: |\n [Unit]\n Description=Pin fixed dual-stack ClusterIPs for kube-dns, argocd-server and whisker\n Requires=install-argocd.service install-calico.service kubeadm-init.service\n After=install-argocd.service install-calico.service kubeadm-init.service network-online.target\n Wants=network-online.target\n ConditionPathExists=!/var/lib/undercloud-stamps/pin-service-ips.done\n\n [Service]\n Type=oneshot\n StandardOutput=journal+console\n StandardError=journal+console\n Environment=KUBECONFIG=/etc/kubernetes/admin.conf\n Environment=PATH=/usr/bin:/usr/sbin:/opt/bin\n ExecStart=/bin/sh -eu -c '\\\n echo \"[pin-service-ips] waiting for API...\" ; \\\n for i in $(seq 1 120); do kubectl get --raw=/readyz \u003e/dev/null 2\u003e\u00261 \u0026\u0026 break; sleep 2; done ; \\\n echo \"[pin-service-ips] ensure namespaces exist...\" ; \\\n kubectl get ns kube-system \u003e/dev/null ; \\\n kubectl get ns argocd \u003e/dev/null 2\u003e\u00261 || kubectl create ns argocd ; \\\n kubectl get ns calico-system \u003e/dev/null ; \\\n echo \"[pin-service-ips] wait for coredns/argocd readiness (best effort)...\" ; \\\n kubectl -n kube-system wait deploy coredns --for=condition=Available=True --timeout=300s || true ; \\\n kubectl -n argocd wait deploy argocd-server --for=condition=Available=True --timeout=600s || true ; \\\n echo \"[pin-service-ips] replace Services with fixed ClusterIPs...\" ; \\\n kubectl -n kube-system delete svc kube-dns --ignore-not-found ; \\\n kubectl apply -f /etc/kubernetes/addons/kube-dns-fixed-svc.yaml ; \\\n kubectl -n argocd delete svc argocd-server --ignore-not-found ; \\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/service.yaml ; \\\n kubectl -n calico-system delete svc whisker --ignore-not-found || true ; \\\n kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/whisker.yaml || true ; \\\n echo \"[pin-service-ips] done.\" \\\n '\n\n ExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/pin-service-ips.done\n\n [Install]\n WantedBy=undercloud-bootstrap.target\n", + "id": "1808288070", "pretty_print": true, - "rendered": "{\n \"ignition\": {\n \"config\": {\n \"replace\": {\n \"verification\": {}\n }\n },\n \"proxy\": {},\n \"security\": {\n \"tls\": {}\n },\n \"timeouts\": {},\n \"version\": \"3.4.0\"\n },\n \"kernelArguments\": {},\n \"passwd\": {\n \"users\": [\n {\n \"name\": \"core\",\n \"sshAuthorizedKeys\": [\n \"ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIHHEAlPo3v4U67Y3411pTjIMkQxwlFWdXrBJkSzXenDH flatcar@undercloud\"\n ]\n }\n ]\n },\n \"storage\": {\n \"directories\": [\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/opt/bin\",\n \"user\": {},\n \"mode\": 493\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/opt/cni/bin\",\n \"user\": {},\n \"mode\": 755\n },\n {\n \"group\": {},\n \"path\": \"/etc/kubernetes/manifests\",\n \"user\": {},\n \"mode\": 493\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/install-calico\",\n \"user\": {},\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/var/lib/undercloud-stamps\",\n \"user\": {},\n \"mode\": 493\n }\n ],\n \"files\": [\n {\n \"group\": {},\n \"path\": \"/etc/hostname\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"\",\n \"source\": \"data:,control-plane1%0A\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"path\": \"/etc/systemd/network/00-eth.network\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/2SMMWuFMBRG9/yUDjE3FcVABsFSOlSkHcUhJLco1aQkUeu/f+h76PC407mH77SfKuq+I7WaUGLsXwhpa4yr878dKY3xGIL8MYyJ/bgQBSRZehrOGIg0ZyIHyC79riKuanvWr6Sqv68eCAFnCxhllNMCEn4VHs/77gCgQCo3qcEGOVuDXo9uNnR0Wo3ko1myUmv8i1+ltO7gxg+L0tvbf0QbBmfDLm4BAAD//4CQfvX4AAAA\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/hosts\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/5ST7W6DIBRA//MUvEDJ/Vhqdt/GqcuWEjGszV5/wVEjlBa5/sLLOYknEakzYMCg1tq6obdf7ueqRML5f7a3+ns5n/KTWz764aIUAaC8dSAd4llI5B21Htx89c6eFtvPE5rbPE5+sO42mlWT7QsOyhxUcVDBwZmDKw5+dCCg/nX+MvnCR8RFiaK4pGcUlSiOS35GsVKfI4CEJ5YO01g7UVBBUY+dKLigqLfeK0LmMNXUKUQ76EXpFOId9Co0hp+DzNq4MfAdpfawd5Tbg0Z0bVkNuV2mAwG3y3wg3F8AAAD//+yy88FYBAAA\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/motd\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/6yRwYobMQyG7/MUOmYX4lzL3sJ2oYV2A01yaG+KrMyIeOxBkodOn74kk0MPaSg0/8WWkD4+4+f/T/MMf8l6v/u0+fb5x9tHWL++vm23sHn/8v3G4B3GP+fC2HViYJM59yAGA6pDOQKCMVXlCJKPiuZayatyuMVYpwRILqO4sAEqQ1+yeDnvY46QSttyDPc89hmrd0Xl13mHiM2gKPRi1fhsZq5CniYYtHRyEL+y/2D0OIGy1eQgGaIYyZAko06zBbc4i5Ycbns84G+b5YPTnOqBNbOzAZXsWhLDkDAzvJfITfMVM7YMo+DLZZY8weJ6eWoIk1CBxXxeWiMn1gJLOCCd6gCLufHUoLaFInTug72sVnO5NNaRNcxVsJHC6QPWn6HmyEqp1BjouHr8u38HAAD//04R3qpqAwAA\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"path\": \"/etc/sysctl.d/99-k8s.conf\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/4zQX27DIAwG8PeeIhcIyqos6h56FuQS01kFGxmWtrefsmXZHzE13wsS+iH7g7EYSlNvKFkvegUdm785Nk+7xQ2bnRP2BkIwiyY+V1y/xZ2UxjMuR8u+dRBCS6nAKWB+7IYFfjnG4ikUVMPeOmEuCu5iI9zWud1HKltqsp9vf7TuKq1HyvNQO99+uykamVCdxEjFRoyi99ov+myIpZC/mwg3+5ZR7RWKe53rru553+8PhyomzgXYrfzY9N3LsLugMgaTgMk1/2Ru/wtaYSuScm3R9wAAAP//X0LFgj8CAAA=\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/flatcar/update.conf\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"\",\n \"source\": \"data:,REBOOT_STRATEGY%3Doff%0A\",\n \"verification\": {}\n },\n \"mode\": 272\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/kubeadm\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubeadm\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/kubelet\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubelet\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/kubectl\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubectl\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/calicoctl\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/calicoctl\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/velero\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/velero\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/etc/kubernetes/kubeadm-init.yaml\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/6xV3W7jNhO951MQvNmbUD+WYie8y7dZfFgsWgSb/gFFLyhyJBOiSXVIOfY+fUHJNqKugwZFbcCQoTlHhzPnjORgfgEMxjtB+7EBqXdZfxcy4/N92UCUFemN04J+diZ+9K413YgyGu9I430MEeXwk+/BBUEo5TSma0FZv+/KrcpihfDi9ziEerVG3zJCKaUagkIzxOmpTEMrRxvPj6cX3plshsRoBWUFI85r+AqdSRUTAaHUyR0IqryL6C0frHRQEkoVmmeveoiCjs4cRJ7nOLo81UnjAPWryyx41RM6ibAQPx0iygfsplNRmh7KzSAoWxVFKepNITZluRYrIe7LWaCyY4iAXLsgKCuLrMjuy+y2ulkiWlEKcVvNmL234w74YMfOOK4NCspyP8TcmgYOoPKkBh1ECPlJ2Kk45DM2n8oYsV5J+/D0+ZPTgzcuJtlS7wGjCfCgNUIIb6tvjNNPHqOg67quiEqw1igZ4QscBWUtwEZVcKtaJdUG6s2mVmtolSzqdS3rlb6737RV0W6qoqjqQq9vN3fNrWqbWla6qhtGOOfk3Vb7OLdy6bbTdJ/ScC+npOz3VheFSN+VEGVR/CHSERhxEF889sZ1qRWD189j45IT/tYDKUS+rm+meaWffFWnjgTAvVHwBmgaYl4WdzfnOZ9x2oVHv5NmisBdyEanAZX1o86mEbHzOSzgD9LJDjDpg6XdWgsH/t+ZI1GmEXN5NsIHIT6keTwD7q8qACcbC1x2HUIno0eOfoypm5RFHGEmHdAfjlxZAy7y5BreGguC5hAX8obe5C16F/lrRKYwfk/Tw/HE8l6aHo4TDcKfI4S4BakBL6rkezXJi54lkbTWv4DmackEQb8XcAUzdZPPfwIfEFpzEPQ3/hV2PgKftgu/guvQj8MZJ+j8ucD+n+5eQY0BMKlbAC+onwMguRpACzFTU8iWOSxPOfwy1yxzePEQE4KRLUgbt9/+Z5x+WNw4rcN/CsOp6sfn+e3B3lyW/PVOZWQ2+NNk98cUjX+RDKKmdj+iSRmgLBxDhJ1mRI5xCy6mBXh6wUjn3XHnx0U6tKApCoTSF2i23vfXbiYuj+bbhWrnNQj664wgfwUAAP//F9AABn8HAAA=\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"path\": \"/etc/kubernetes/addons/kube-dns-fixed-svc.yaml\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/4yRwU7rMBBF9/6K+YCXKn6Uls62FVJ3FgU2iMXUGZAVJzb2JFL+HoWmEBYV9fLOuUejMUX3zCm70CL0WtWurRAOnHpnWTUsVJEQKoCWGkaouyMXVZunIEey5zQPWbhRAJ6O7PPYAajvckExzoo5sh1nMkRG2PouC6e9UQAu3lPj/GCCd3ZAeOCPziXedeQPQrb+QRxnhJe96Vf/YG/65asCsGcTwv+y1Lhcl7jWeoVvqBFvb+bItFxxkRxnulyUi41efAUxJPlunU5xusL4xiHCVASIKUiwwSM87cyUCaV3FjMHZ55CbPzL9bi9wtWwJGd/77XRV9smNLNnKyFd+sLPAAAA//8MLaPANAIAAA==\",\n \"verification\": {}\n },\n \"mode\": 420\n }\n ]\n },\n \"systemd\": {\n \"units\": [\n {\n \"contents\": \"[Unit]\\nDescription=Load necessary kernel modules\\nBefore=containerd.service kubeadm-init.service\\n\\n[Service]\\nType=oneshot\\n\\nExecStart=/usr/bin/modprobe br_netfilter\\nExecStart=/usr/bin/modprobe overlay\\nRemainAfterExit=yes\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"modules-load.service\"\n },\n {\n \"enabled\": true,\n \"name\": \"systemd-networkd-wait-online.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=containerd container runtime\\nAfter=network.target modules-load.service\\n\\n[Service]\\n#StandardOutput=journal+console\\n#StandardError=journal+console\\n\\nExecStart=/usr/bin/containerd\\nRestart=always\\nRestartSec=5\\nDelegate=yes\\nKillMode=process\\nOOMScoreAdjust=-999\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"containerd.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Set Timezone\\nAfter=network-online.target\\nWants=network-online.target\\n[Service]\\nStandardOutput=journal+console\\nStandardError=journal+console\\n\\nExecStart=/bin/sh -c 'echo \\\"setting timezone to Europe/Berlin\\\"'\\nStandardOutput=journal+console\\nStandardError=journal+console\\nType=oneshot\\nRestart=on-failure\\nExecStart=/usr/bin/timedatectl set-timezone Europe/Berlin\\nExecStart=/usr/bin/timedatectl set-ntp true \\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"set-timezone.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=kubelet, the Kubernetes Node Agent\\nDocumentation=https://kubernets.io/docs/home\\nWants=network-online.target\\nAfter=network-online.target\\n[Service]\\n#StandardOutput=journal+console\\n#StandardError=journal+console\\n\\n#EnvironmentFile=/run/metadata/coreos\\nEnvironment=\\\"KUBELET_KUBECONFIG_ARGS=--bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf\\\"\\nEnvironment=\\\"KUBELET_CONFIG_ARGS=--config=/var/lib/kubelet/config.yaml\\\"\\n# This is a file that \\\"kubeadm init\\\" and \\\"kubeadm join\\\" generates at runtime, populating the KUBELET_KUBEADM_ARGS variable dynamically\\nEnvironmentFile=-/var/lib/kubelet/kubeadm-flags.env\\nExecStart=/opt/bin/kubelet $KUBELET_KUBECONFIG_ARGS $KUBELET_CONFIG_ARGS $KUBELET_KUBEADM_ARGS $KUBELET_EXTRA_ARGS\\nRestart=always\\nStartLimitInterval=0\\nRestartSec=10\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"kubelet.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Kubeadm Init Cluster\\nAfter=network-online.target containerd.service kubelet.service\\nWants=network-online.target\\nConditionPathExists=!/etc/kubernetes/kubelet.conf\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\n\\nExecStart=/bin/sh -c 'echo \\\"kubeadm-init.service started...\\\"'\\n\\n# Environment\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin:/opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent-uds/\\n\\nExecStartPre=/bin/sleep 30s\\nExecStart=/bin/sh -c 'echo \\\"running kubeadm init...\\\"'\\nExecStart=/opt/bin/kubeadm init --upload-certs --config=/etc/kubernetes/kubeadm-init.yaml\\n\\n# copy files for kubectl\\nExecStart=/bin/sh -c 'echo \\\"copying files (admin.conf) to core home folder.\\\"'\\nExecStartPost=/usr/bin/mkdir -p /home/core/.kube\\nExecStartPost=/usr/bin/cp -i /etc/kubernetes/admin.conf /home/core/.kube/config\\nExecStartPost=/usr/bin/chown core:core /home/core/.kube/config\\n\\nExecStartPost=/usr/bin/systemctl disable kubeadm-init.service\\nRestart=on-failure\\nRestartSec=120s\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"kubeadm-init.service\"\n },\n {\n \"contents\": \"[Unit]\\nWants=kubeadm-init.service\\nAfter=kubeadm-init.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-calico.done\\n\\n\\n[Service]\\nStandardOutput=journal+console\\nStandardError=journal+console\\n\\nExecStart=/bin/sh -c 'echo \\\"install.calico.service started...\\\"'\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nExecStart=/bin/sh -c 'echo \\\"witing 30s...\\\"'\\nExecStart=/bin/sleep 30s\\nExecStart=/bin/sh -c 'echo \\\"create calico namespace...\\\"'\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/namespace.yaml\\nExecStart=/bin/sh -c 'echo \\\"install tigera operator...\\\"'\\nExecStart=-/opt/bin/kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/operator-crds.yaml\\nExecStart=-/opt/bin/kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/tigera-operator.yaml\\nExecStart=/bin/sh -c 'echo \\\"witing 60s...\\\"'\\nExecStart=/bin/sleep 60s\\nExecStart=/bin/sh -c 'echo \\\"witing for tigera operator... (20mini max)\\\"'\\nExecStart=/opt/bin/kubectl wait deployment -n tigera-operator tigera-operator --for condition=Available=True --timeout=1200s\\nExecStart=/bin/sh -c 'echo \\\"create clico custom ressources...\\\"'\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/custom-resources.yaml\\n\\nExecStart=/bin/sh -c 'echo \\\"witing 3m..\\\"'\\nExecStart=/bin/sleep 3m\\n#ExecStart=/bin/sh -c 'echo \\\"apply calico (calico-apiserver)...\\\"'\\n#ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/calico.yaml\\n#ExecStart=/bin/sh -c 'echo \\\"witing 1m...\\\"'\\n#ExecStart=/bin/sleep 2m\\nExecStart=/bin/sh -c 'echo \\\"witing calico-apiserver... (20mini max)\\\"'\\nExecStart=/opt/bin/kubectl wait deployment -n calico-apiserver calico-apiserver --for condition=Available=True --timeout=1200s\\nExecStart=/bin/sh -c 'echo \\\"witing 120s...\\\"'\\nExecStart=/bin/sleep 2m\\nExecStart=/bin/sh -c 'echo \\\"apply calico-peers...\\\"'\\nExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/calico-peer.yaml\\nExecStart=/bin/sh -c 'echo \\\"witing 60s...\\\"'\\nExecStart=/bin/sleep 1m\\nExecStart=/bin/sh -c 'echo \\\"apply calico-ippools...\\\"'\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/ippools.yaml\\n\\n#ExecStart=/bin/sh -c 'echo \\\"witing for whisker..\\\"'\\n#ExecStart=/opt/bin/kubectl wait deployment -n calico-system whisker --for condition=Available=True --timeout=1200s\\n#ExecStart=/bin/sh -c 'echo \\\"port-forward -n calico-system service/whisker 8081:8081\\\"'\\n#ExecStart=/opt/bin/kubectl port-forward -n calico-system service/whisker 8081:8081\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-calico.done\\nExecStart=/usr/bin/systemctl disable install-calico.service\\n#RemainAfterExit=true\\nRestart=on-failure\\nRestartSec=120s\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"install-calico.service\"\n },\n {\n \"contents\": \"[Unit]\\nWants=kubeadm-init.service\\nAfter=kubeadm-init.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-ceph.done\\n\\n[Service]\\nStandardOutput=journal+console\\nStandardError=journal+console\\n\\nExecStart=/bin/sh -c 'echo \\\"install.ceph.service started...\\\"'\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\nType=oneshot\\n\\nStandardOutput=journal+console\\nStandardError=journal+console\\nExecStart=/bin/sh -c 'echo \\\"witing 30s...\\\"'\\nExecStart=/bin/sleep 30s\\nExecStart=/bin/sh -c 'echo \\\"create ceph namespace...\\\"'\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/namespace.yaml\\n\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-nodeplugin-rbac.yaml\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-provisioner-rbac.yaml\\n\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/secrets.yaml\\n\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/ceph-conf.yaml\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-cephfsplugin-provisioner.yaml\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-cephfsplugin.yaml\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-config-map.yaml\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-encryption-kms-config.yaml\\n\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csidriver.yaml\\n\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/storage-class.yaml\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-ceph.done\\nExecStart=/usr/bin/systemctl disable install-ceph.service\\n#RemainAfterExit=true\\nRestart=on-failure\\nRestartSec=120s\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"install-ceph.service\"\n },\n {\n \"contents\": \"[Unit]\\nWants=install-ceph.service\\nAfter=install-ceph.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-gitea.done\\n\\n[Service]\\nStandardOutput=journal+console\\nStandardError=journal+console\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\nType=oneshot\\n\\nExecStart=/opt/bin/kubectl wait deployment -n ceph csi-cephfsplugin-provisioner --for condition=Available=True --timeout=1200s\\nExecStart=/bin/sleep 4m\\nExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/namespace.yaml\\nExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/secrets.yaml\\nExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/db.yaml\\nExecStart=/bin/sleep 60s\\nExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/adminer.yaml\\nExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/gitea.yaml\\nExecStart=/bin/sleep 3m\\nExecStart=/opt/bin/kubectl exec deploy/gitea -n gitea -- /bin/startup.sh\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-gitea.done\\nExecStart=/usr/bin/systemctl disable install-gitea.service\\nRestart=on-failure\\nRestartSec=120s\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"install-gitea.service\"\n },\n {\n \"contents\": \"[Unit]\\nWants=install-calico.service\\nAfter=install-calico.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-argocd.done\\n[Service]\\nStandardOutput=journal+console\\nStandardError=journal+console\\n\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\nType=oneshot\\n\\nExecStart=/opt/bin/kubectl wait deployment -n kube-system coredns --for condition=Available=True --timeout=600s\\n\\nExecStart=/bin/sleep 1m\\nExecStart=/opt/bin/kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/namespace.yaml\\nExecStart=/opt/bin/kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/install.yaml\\nExecStart=/opt/bin/kubectl wait deployment -n argocd argocd-server --for condition=Available=True --timeout=600s\\n\\nExecStart=/opt/bin/kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/repo.yaml\\nExecStart=/bin/sleep 10s\\nExecStart=/opt/bin/kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/apps.yaml\\n\\nExecStart=/bin/sleep 10s\\nExecStart=/opt/bin/kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/ingress.yaml\\n\\n\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/repos/k8aux-bootstrap.yaml\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/repos/k8aux-apps.yaml\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/apps/argocd.yaml\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/apps/calico.yaml\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/apps/rook-ceph.yaml\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/apps/gitea.yaml\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/argocd/argocd-secret.yaml\\n##ExecStart=/bin/sleep 10m\\n#ExecStart=/opt/bin/kubectl wait deployment -n gitea gitea --for condition=Available=True --timeout=4800s\\n#ExecStart=/bin/sleep 10m\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://gitea.gitea.svc.k8aux.undercloud.cf:3000/undercloud/k8aux-apps/raw/branch/main/app-of-apps/app-of-apps.yaml\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-argocd.done\\nExecStart=/usr/bin/systemctl disable install-argocd.service\\nRestart=on-failure\\nRestartSec=120s\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"install-argocd.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Pin fixed dual-stack ClusterIPs for kube-dns, argocd-server and whisker\\nAfter=install-argocd.service install-calico.service kubeadm-init.service network-online.target\\nWants=install-argocd.service install-calico.service kubeadm-init.service network-online.target\\nConditionPathExists=!/var/lib/undercloud-stamps/pin-service-ips.done\\n\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=PATH=/usr/bin:/usr/sbin:/opt/bin\\nExecStart=/bin/sh -eu -c '\\\\\\n echo \\\"[pin-service-ips] waiting for API...\\\" ; \\\\\\n for i in $(seq 1 120); do kubectl get --raw=/readyz \\u003e/dev/null 2\\u003e\\u00261 \\u0026\\u0026 break; sleep 2; done ; \\\\\\n echo \\\"[pin-service-ips] ensure namespaces exist...\\\" ; \\\\\\n kubectl get ns kube-system \\u003e/dev/null ; \\\\\\n kubectl get ns argocd \\u003e/dev/null 2\\u003e\\u00261 || kubectl create ns argocd ; \\\\\\n kubectl get ns calico-system \\u003e/dev/null ; \\\\\\n echo \\\"[pin-service-ips] wait for coredns/argocd readiness (best effort)...\\\" ; \\\\\\n kubectl -n kube-system wait deploy coredns --for=condition=Available=True --timeout=300s || true ; \\\\\\n kubectl -n argocd wait deploy argocd-server --for=condition=Available=True --timeout=600s || true ; \\\\\\n echo \\\"[pin-service-ips] replace Services with fixed ClusterIPs...\\\" ; \\\\\\n kubectl -n kube-system delete svc kube-dns --ignore-not-found ; \\\\\\n kubectl apply -f /etc/kubernetes/addons/kube-dns-fixed-svc.yaml ; \\\\\\n kubectl -n argocd delete svc argocd-server --ignore-not-found ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/service.yaml ; \\\\\\n kubectl -n calico-system delete svc whisker --ignore-not-found || true ; \\\\\\n kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/whisker.yaml || true ; \\\\\\n echo \\\"[pin-service-ips] done.\\\" \\\\\\n'\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/pin-service-ips.done\\nRestart=on-failure\\nRestartSec=120s\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"pin-service-ips.service\"\n }\n ]\n }\n}", + "rendered": "{\n \"ignition\": {\n \"config\": {\n \"replace\": {\n \"verification\": {}\n }\n },\n \"proxy\": {},\n \"security\": {\n \"tls\": {}\n },\n \"timeouts\": {},\n \"version\": \"3.4.0\"\n },\n \"kernelArguments\": {},\n \"passwd\": {\n \"users\": [\n {\n \"name\": \"core\",\n \"sshAuthorizedKeys\": [\n \"ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIHHEAlPo3v4U67Y3411pTjIMkQxwlFWdXrBJkSzXenDH flatcar@undercloud\"\n ]\n }\n ]\n },\n \"storage\": {\n \"directories\": [\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/opt/bin\",\n \"user\": {},\n \"mode\": 493\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/opt/cni/bin\",\n \"user\": {},\n \"mode\": 755\n },\n {\n \"group\": {},\n \"path\": \"/etc/kubernetes/manifests\",\n \"user\": {},\n \"mode\": 493\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/install-calico\",\n \"user\": {},\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/var/lib/undercloud-stamps\",\n \"user\": {},\n \"mode\": 493\n }\n ],\n \"files\": [\n {\n \"group\": {},\n \"path\": \"/etc/hostname\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"\",\n \"source\": \"data:,control-plane1%0A\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"path\": \"/etc/systemd/network/00-eth.network\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/2SMMWuFMBRG9/yUDjE3FcVABsFSOlSkHcUhJLco1aQkUeu/f+h76PC407mH77SfKuq+I7WaUGLsXwhpa4yr878dKY3xGIL8MYyJ/bgQBSRZehrOGIg0ZyIHyC79riKuanvWr6Sqv68eCAFnCxhllNMCEn4VHs/77gCgQCo3qcEGOVuDXo9uNnR0Wo3ko1myUmv8i1+ltO7gxg+L0tvbf0QbBmfDLm4BAAD//4CQfvX4AAAA\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/hosts\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/5ST7W6DIBRA//MUvEDJ/Vhqdt/GqcuWEjGszV5/wVEjlBa5/sLLOYknEakzYMCg1tq6obdf7ueqRML5f7a3+ns5n/KTWz764aIUAaC8dSAd4llI5B21Htx89c6eFtvPE5rbPE5+sO42mlWT7QsOyhxUcVDBwZmDKw5+dCCg/nX+MvnCR8RFiaK4pGcUlSiOS35GsVKfI4CEJ5YO01g7UVBBUY+dKLigqLfeK0LmMNXUKUQ76EXpFOId9Co0hp+DzNq4MfAdpfawd5Tbg0Z0bVkNuV2mAwG3y3wg3F8AAAD//+yy88FYBAAA\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/motd\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/6yRwYobMQyG7/MUOmYX4lzL3sJ2oYV2A01yaG+KrMyIeOxBkodOn74kk0MPaSg0/8WWkD4+4+f/T/MMf8l6v/u0+fb5x9tHWL++vm23sHn/8v3G4B3GP+fC2HViYJM59yAGA6pDOQKCMVXlCJKPiuZayatyuMVYpwRILqO4sAEqQ1+yeDnvY46QSttyDPc89hmrd0Xl13mHiM2gKPRi1fhsZq5CniYYtHRyEL+y/2D0OIGy1eQgGaIYyZAko06zBbc4i5Ycbns84G+b5YPTnOqBNbOzAZXsWhLDkDAzvJfITfMVM7YMo+DLZZY8weJ6eWoIk1CBxXxeWiMn1gJLOCCd6gCLufHUoLaFInTug72sVnO5NNaRNcxVsJHC6QPWn6HmyEqp1BjouHr8u38HAAD//04R3qpqAwAA\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"path\": \"/etc/sysctl.d/99-k8s.conf\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/4zQX27DIAwG8PeeIhcIyqos6h56FuQS01kFGxmWtrefsmXZHzE13wsS+iH7g7EYSlNvKFkvegUdm785Nk+7xQ2bnRP2BkIwiyY+V1y/xZ2UxjMuR8u+dRBCS6nAKWB+7IYFfjnG4ikUVMPeOmEuCu5iI9zWud1HKltqsp9vf7TuKq1HyvNQO99+uykamVCdxEjFRoyi99ov+myIpZC/mwg3+5ZR7RWKe53rru553+8PhyomzgXYrfzY9N3LsLugMgaTgMk1/2Ru/wtaYSuScm3R9wAAAP//X0LFgj8CAAA=\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/flatcar/update.conf\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"\",\n \"source\": \"data:,REBOOT_STRATEGY%3Doff%0A\",\n \"verification\": {}\n },\n \"mode\": 272\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/kubeadm\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubeadm\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/kubelet\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubelet\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/kubectl\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubectl\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/calicoctl\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/calicoctl\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/velero\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/velero\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/busybox\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/busybox\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/etc/systemd/journald.conf.d/10-forward-to-syslog.conf\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"\",\n \"source\": \"data:,%5BJournal%5D%0AForwardToSyslog%3Dyes%0A\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"path\": \"/etc/kubernetes/kubeadm-init.yaml\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/6xV3W7jNhO951MQvNmbUD+WYie8y7dZfFgsWgSb/gFFLyhyJBOiSXVIOfY+fUHJNqKugwZFbcCQoTlHhzPnjORgfgEMxjtB+7EBqXdZfxcy4/N92UCUFemN04J+diZ+9K413YgyGu9I430MEeXwk+/BBUEo5TSma0FZv+/KrcpihfDi9ziEerVG3zJCKaUagkIzxOmpTEMrRxvPj6cX3plshsRoBWUFI85r+AqdSRUTAaHUyR0IqryL6C0frHRQEkoVmmeveoiCjs4cRJ7nOLo81UnjAPWryyx41RM6ibAQPx0iygfsplNRmh7KzSAoWxVFKepNITZluRYrIe7LWaCyY4iAXLsgKCuLrMjuy+y2ulkiWlEKcVvNmL234w74YMfOOK4NCspyP8TcmgYOoPKkBh1ECPlJ2Kk45DM2n8oYsV5J+/D0+ZPTgzcuJtlS7wGjCfCgNUIIb6tvjNNPHqOg67quiEqw1igZ4QscBWUtwEZVcKtaJdUG6s2mVmtolSzqdS3rlb6737RV0W6qoqjqQq9vN3fNrWqbWla6qhtGOOfk3Vb7OLdy6bbTdJ/ScC+npOz3VheFSN+VEGVR/CHSERhxEF889sZ1qRWD189j45IT/tYDKUS+rm+meaWffFWnjgTAvVHwBmgaYl4WdzfnOZ9x2oVHv5NmisBdyEanAZX1o86mEbHzOSzgD9LJDjDpg6XdWgsH/t+ZI1GmEXN5NsIHIT6keTwD7q8qACcbC1x2HUIno0eOfoypm5RFHGEmHdAfjlxZAy7y5BreGguC5hAX8obe5C16F/lrRKYwfk/Tw/HE8l6aHo4TDcKfI4S4BakBL6rkezXJi54lkbTWv4DmackEQb8XcAUzdZPPfwIfEFpzEPQ3/hV2PgKftgu/guvQj8MZJ+j8ucD+n+5eQY0BMKlbAC+onwMguRpACzFTU8iWOSxPOfwy1yxzePEQE4KRLUgbt9/+Z5x+WNw4rcN/CsOp6sfn+e3B3lyW/PVOZWQ2+NNk98cUjX+RDKKmdj+iSRmgLBxDhJ1mRI5xCy6mBXh6wUjn3XHnx0U6tKApCoTSF2i23vfXbiYuj+bbhWrnNQj664wgfwUAAP//F9AABn8HAAA=\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"path\": \"/etc/kubernetes/addons/kube-dns-fixed-svc.yaml\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/4yRwU7rMBBF9/6K+YCXKn6Uls62FVJ3FgU2iMXUGZAVJzb2JFL+HoWmEBYV9fLOuUejMUX3zCm70CL0WtWurRAOnHpnWTUsVJEQKoCWGkaouyMXVZunIEey5zQPWbhRAJ6O7PPYAajvckExzoo5sh1nMkRG2PouC6e9UQAu3lPj/GCCd3ZAeOCPziXedeQPQrb+QRxnhJe96Vf/YG/65asCsGcTwv+y1Lhcl7jWeoVvqBFvb+bItFxxkRxnulyUi41efAUxJPlunU5xusL4xiHCVASIKUiwwSM87cyUCaV3FjMHZ55CbPzL9bi9wtWwJGd/77XRV9smNLNnKyFd+sLPAAAA//8MLaPANAIAAA==\",\n \"verification\": {}\n },\n \"mode\": 420\n }\n ]\n },\n \"systemd\": {\n \"units\": [\n {\n \"contents\": \"[Unit]\\nDescription=Forward journald to central syslog server\\nWants=network-online.target\\nAfter=network-online.target systemd-journald.service\\n\\n[Service]\\nType=simple\\nExecStart=/opt/bin/busybox syslogd -n -R syslog.undercloud.local -P 514\\nRestart=always\\nRestartSec=2\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"syslog-forward.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Undercloud Bootstrap Chain\\nWants=network-online.target\\nAfter=network-online.target\\n# Start the chain entry\\nWants=containerd.service kubelet.service kubeadm-init.service\\nAfter=containerd.service kubelet.service kubeadm-init.service\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"undercloud-bootstrap.target\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Load necessary kernel modules\\nBefore=containerd.service kubeadm-init.service\\n\\n[Service]\\nType=oneshot\\nExecStart=/usr/bin/modprobe br_netfilter\\nExecStart=/usr/bin/modprobe overlay\\nRemainAfterExit=yes\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"modules-load.service\"\n },\n {\n \"enabled\": true,\n \"name\": \"systemd-networkd-wait-online.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=containerd container runtime\\nAfter=network.target modules-load.service\\nWants=modules-load.service\\n\\n[Service]\\nExecStart=/usr/bin/containerd\\nRestart=always\\nRestartSec=5\\nDelegate=yes\\nKillMode=process\\nOOMScoreAdjust=-999\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"containerd.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Set Timezone\\nAfter=network-online.target\\nWants=network-online.target\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nExecStart=/bin/sh -c 'echo \\\"setting timezone to Europe/Berlin\\\"'\\nExecStart=/usr/bin/timedatectl set-timezone Europe/Berlin\\nExecStart=/usr/bin/timedatectl set-ntp true\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"set-timezone.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=kubelet, the Kubernetes Node Agent\\nDocumentation=https://kubernetes.io/docs/home\\nWants=network-online.target\\nAfter=network-online.target containerd.service\\nRequires=containerd.service\\n\\n[Service]\\nEnvironment=\\\"KUBELET_KUBECONFIG_ARGS=--bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf\\\"\\nEnvironment=\\\"KUBELET_CONFIG_ARGS=--config=/var/lib/kubelet/config.yaml\\\"\\nEnvironmentFile=-/var/lib/kubelet/kubeadm-flags.env\\nExecStart=/opt/bin/kubelet $KUBELET_KUBECONFIG_ARGS $KUBELET_CONFIG_ARGS $KUBELET_KUBEADM_ARGS $KUBELET_EXTRA_ARGS\\nRestart=always\\nStartLimitInterval=0\\nRestartSec=10\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"kubelet.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Kubeadm Init Cluster\\nWants=network-online.target\\nAfter=network-online.target containerd.service kubelet.service\\nRequires=containerd.service kubelet.service\\nConditionPathExists=!/etc/kubernetes/kubelet.conf\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\n\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin:/opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent-uds/\\n\\nExecStart=/bin/sh -eu -c '\\\\\\n echo \\\"[kubeadm-init] started...\\\" ; \\\\\\n echo \\\"[kubeadm-init] waiting for containerd socket...\\\" ; \\\\\\n for i in $(seq 1 60); do test -S /run/containerd/containerd.sock \\u0026\\u0026 break; sleep 1; done ; \\\\\\n echo \\\"[kubeadm-init] running kubeadm init...\\\" ; \\\\\\n /opt/bin/kubeadm init --upload-certs --config=/etc/kubernetes/kubeadm-init.yaml ; \\\\\\n echo \\\"[kubeadm-init] copying kubeconfig to core...\\\" ; \\\\\\n mkdir -p /home/core/.kube ; \\\\\\n cp -f /etc/kubernetes/admin.conf /home/core/.kube/config ; \\\\\\n chown core:core /home/core/.kube/config ; \\\\\\n echo \\\"[kubeadm-init] done.\\\" \\\\\\n'\\n\\n# strictly start next step (serialization)\\nExecStartPost=/usr/bin/systemctl start install-calico.service\\n\\n[Install]\\nWantedBy=undercloud-bootstrap.target\\n\",\n \"enabled\": false,\n \"name\": \"kubeadm-init.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Install Calico\\nRequires=kubeadm-init.service\\nAfter=kubeadm-init.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-calico.done\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\n\\nExecStart=/bin/sh -eu -c '\\\\\\n echo \\\"[calico] waiting for API /readyz...\\\" ; \\\\\\n for i in $(seq 1 180); do kubectl get --raw=/readyz \\u003e/dev/null 2\\u003e\\u00261 \\u0026\\u0026 break; sleep 2; done ; \\\\\\n echo \\\"[calico] create namespace + operator...\\\" ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/namespace.yaml ; \\\\\\n kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/operator-crds.yaml || true ; \\\\\\n kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/tigera-operator.yaml || true ; \\\\\\n echo \\\"[calico] wait for tigera-operator...\\\" ; \\\\\\n kubectl wait deployment -n tigera-operator tigera-operator --for condition=Available=True --timeout=1200s ; \\\\\\n echo \\\"[calico] apply custom resources...\\\" ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/custom-resources.yaml ; \\\\\\n echo \\\"[calico] wait for calico-apiserver...\\\" ; \\\\\\n kubectl wait deployment -n calico-apiserver calico-apiserver --for condition=Available=True --timeout=1200s ; \\\\\\n echo \\\"[calico] apply peers + pools...\\\" ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/calico-peer.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/ippools.yaml ; \\\\\\n echo \\\"[calico] done.\\\" \\\\\\n'\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-calico.done\\nExecStartPost=/usr/bin/systemctl start install-ceph.service\\n\\n[Install]\\nWantedBy=undercloud-bootstrap.target\\n\",\n \"enabled\": false,\n \"name\": \"install-calico.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Install Ceph CSI\\nRequires=install-calico.service\\nAfter=install-calico.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-ceph.done\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\n\\nExecStart=/bin/sh -eu -c '\\\\\\n echo \\\"[ceph] waiting for API /readyz...\\\" ; \\\\\\n for i in $(seq 1 180); do kubectl get --raw=/readyz \\u003e/dev/null 2\\u003e\\u00261 \\u0026\\u0026 break; sleep 2; done ; \\\\\\n echo \\\"[ceph] apply manifests...\\\" ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/namespace.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-nodeplugin-rbac.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-provisioner-rbac.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/secrets.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/ceph-conf.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-cephfsplugin-provisioner.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-cephfsplugin.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-config-map.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-encryption-kms-config.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csidriver.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/storage-class.yaml ; \\\\\\n echo \\\"[ceph] done.\\\" \\\\\\n'\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-ceph.done\\nExecStartPost=/usr/bin/systemctl start install-gitea.service\\n\\n[Install]\\nWantedBy=undercloud-bootstrap.target\\n\",\n \"enabled\": false,\n \"name\": \"install-ceph.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Install Gitea\\nRequires=install-ceph.service\\nAfter=install-ceph.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-gitea.done\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\n\\nExecStart=/bin/sh -eu -c '\\\\\\n echo \\\"[gitea] wait for ceph provisioner...\\\" ; \\\\\\n kubectl wait deployment -n ceph csi-cephfsplugin-provisioner --for condition=Available=True --timeout=1200s ; \\\\\\n echo \\\"[gitea] apply manifests...\\\" ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/namespace.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/secrets.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/db.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/adminer.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/gitea.yaml ; \\\\\\n echo \\\"[gitea] wait for gitea deployment...\\\" ; \\\\\\n kubectl -n gitea wait deployment gitea --for=condition=Available=True --timeout=1200s ; \\\\\\n echo \\\"[gitea] run startup...\\\" ; \\\\\\n kubectl exec deploy/gitea -n gitea -- /bin/startup.sh ; \\\\\\n echo \\\"[gitea] done.\\\" \\\\\\n'\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-gitea.done\\nExecStartPost=/usr/bin/systemctl start install-argocd.service\\n\\n[Install]\\nWantedBy=undercloud-bootstrap.target\\n\",\n \"enabled\": false,\n \"name\": \"install-gitea.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Install ArgoCD\\nRequires=install-calico.service install-gitea.service\\nAfter=install-calico.service install-gitea.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-argocd.done\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\n\\nExecStart=/bin/sh -eu -c '\\\\\\n echo \\\"[argocd] wait for coredns...\\\" ; \\\\\\n kubectl -n kube-system wait deploy coredns --for=condition=Available=True --timeout=1200s ; \\\\\\n echo \\\"[argocd] install...\\\" ; \\\\\\n kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/namespace.yaml ; \\\\\\n kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/install.yaml ; \\\\\\n kubectl -n argocd wait deploy argocd-server --for=condition=Available=True --timeout=1200s ; \\\\\\n kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/repo.yaml ; \\\\\\n kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/apps.yaml ; \\\\\\n kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/ingress.yaml ; \\\\\\n echo \\\"[argocd] done.\\\" \\\\\\n'\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-argocd.done\\nExecStartPost=/usr/bin/systemctl start pin-service-ips.service\\n\\n[Install]\\nWantedBy=undercloud-bootstrap.target\\n\",\n \"enabled\": false,\n \"name\": \"install-argocd.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Pin fixed dual-stack ClusterIPs for kube-dns, argocd-server and whisker\\nRequires=install-argocd.service install-calico.service kubeadm-init.service\\nAfter=install-argocd.service install-calico.service kubeadm-init.service network-online.target\\nWants=network-online.target\\nConditionPathExists=!/var/lib/undercloud-stamps/pin-service-ips.done\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=PATH=/usr/bin:/usr/sbin:/opt/bin\\nExecStart=/bin/sh -eu -c '\\\\\\n echo \\\"[pin-service-ips] waiting for API...\\\" ; \\\\\\n for i in $(seq 1 120); do kubectl get --raw=/readyz \\u003e/dev/null 2\\u003e\\u00261 \\u0026\\u0026 break; sleep 2; done ; \\\\\\n echo \\\"[pin-service-ips] ensure namespaces exist...\\\" ; \\\\\\n kubectl get ns kube-system \\u003e/dev/null ; \\\\\\n kubectl get ns argocd \\u003e/dev/null 2\\u003e\\u00261 || kubectl create ns argocd ; \\\\\\n kubectl get ns calico-system \\u003e/dev/null ; \\\\\\n echo \\\"[pin-service-ips] wait for coredns/argocd readiness (best effort)...\\\" ; \\\\\\n kubectl -n kube-system wait deploy coredns --for=condition=Available=True --timeout=300s || true ; \\\\\\n kubectl -n argocd wait deploy argocd-server --for=condition=Available=True --timeout=600s || true ; \\\\\\n echo \\\"[pin-service-ips] replace Services with fixed ClusterIPs...\\\" ; \\\\\\n kubectl -n kube-system delete svc kube-dns --ignore-not-found ; \\\\\\n kubectl apply -f /etc/kubernetes/addons/kube-dns-fixed-svc.yaml ; \\\\\\n kubectl -n argocd delete svc argocd-server --ignore-not-found ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/service.yaml ; \\\\\\n kubectl -n calico-system delete svc whisker --ignore-not-found || true ; \\\\\\n kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/whisker.yaml || true ; \\\\\\n echo \\\"[pin-service-ips] done.\\\" \\\\\\n'\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/pin-service-ips.done\\n\\n[Install]\\nWantedBy=undercloud-bootstrap.target\\n\",\n \"enabled\": false,\n \"name\": \"pin-service-ips.service\"\n }\n ]\n }\n}", "snippets": null, "strict": false }, @@ -152,38 +152,6 @@ } ] }, - { - "mode": "managed", - "type": "null_resource", - "name": "wait_for_cluster_ready", - "provider": "provider[\"registry.terraform.io/hashicorp/null\"]", - "instances": [ - { - "schema_version": 0, - "attributes": { - "id": "6387838205690992951", - "triggers": null - }, - "sensitive_attributes": [], - "identity_schema_version": 0, - "dependencies": [ - "data.ct_config.control_plane1_ignition", - "data.ct_config.control_plane2_ignition", - "data.ct_config.control_plane3_ignition", - "null_resource.wait_for_cp1_api", - "null_resource.wait_for_cp2_cp3_api", - "proxmox_virtual_environment_download_file.flatcar_image", - "proxmox_virtual_environment_file.control_plane1_ignition", - "proxmox_virtual_environment_file.control_plane2_ignition", - "proxmox_virtual_environment_file.control_plane3_ignition", - "proxmox_virtual_environment_vm.control_plane1", - "proxmox_virtual_environment_vm.control_plane2", - "proxmox_virtual_environment_vm.control_plane3", - "proxmox_virtual_environment_vm.flatcar_template" - ] - } - ] - }, { "mode": "managed", "type": "null_resource", @@ -191,9 +159,10 @@ "provider": "provider[\"registry.terraform.io/hashicorp/null\"]", "instances": [ { + "status": "tainted", "schema_version": 0, "attributes": { - "id": "1483253127709706838", + "id": "968412366077437715", "triggers": null }, "sensitive_attributes": [], @@ -208,37 +177,6 @@ } ] }, - { - "mode": "managed", - "type": "null_resource", - "name": "wait_for_cp2_cp3_api", - "provider": "provider[\"registry.terraform.io/hashicorp/null\"]", - "instances": [ - { - "schema_version": 0, - "attributes": { - "id": "8200555079622067824", - "triggers": null - }, - "sensitive_attributes": [], - "identity_schema_version": 0, - "dependencies": [ - "data.ct_config.control_plane1_ignition", - "data.ct_config.control_plane2_ignition", - "data.ct_config.control_plane3_ignition", - "null_resource.wait_for_cp1_api", - "proxmox_virtual_environment_download_file.flatcar_image", - "proxmox_virtual_environment_file.control_plane1_ignition", - "proxmox_virtual_environment_file.control_plane2_ignition", - "proxmox_virtual_environment_file.control_plane3_ignition", - "proxmox_virtual_environment_vm.control_plane1", - "proxmox_virtual_environment_vm.control_plane2", - "proxmox_virtual_environment_vm.control_plane3", - "proxmox_virtual_environment_vm.flatcar_template" - ] - } - ] - }, { "mode": "managed", "type": "proxmox_virtual_environment_download_file", @@ -290,7 +228,7 @@ "source_file": [], "source_raw": [ { - "data": "{\n \"ignition\": {\n \"config\": {\n \"replace\": {\n \"verification\": {}\n }\n },\n \"proxy\": {},\n \"security\": {\n \"tls\": {}\n },\n \"timeouts\": {},\n \"version\": \"3.4.0\"\n },\n \"kernelArguments\": {},\n \"passwd\": {\n \"users\": [\n {\n \"name\": \"core\",\n \"sshAuthorizedKeys\": [\n \"ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIHHEAlPo3v4U67Y3411pTjIMkQxwlFWdXrBJkSzXenDH flatcar@undercloud\"\n ]\n }\n ]\n },\n \"storage\": {\n \"directories\": [\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/opt/bin\",\n \"user\": {},\n \"mode\": 493\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/opt/cni/bin\",\n \"user\": {},\n \"mode\": 755\n },\n {\n \"group\": {},\n \"path\": \"/etc/kubernetes/manifests\",\n \"user\": {},\n \"mode\": 493\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/install-calico\",\n \"user\": {},\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/var/lib/undercloud-stamps\",\n \"user\": {},\n \"mode\": 493\n }\n ],\n \"files\": [\n {\n \"group\": {},\n \"path\": \"/etc/hostname\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"\",\n \"source\": \"data:,control-plane1%0A\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"path\": \"/etc/systemd/network/00-eth.network\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/2SMMWuFMBRG9/yUDjE3FcVABsFSOlSkHcUhJLco1aQkUeu/f+h76PC407mH77SfKuq+I7WaUGLsXwhpa4yr878dKY3xGIL8MYyJ/bgQBSRZehrOGIg0ZyIHyC79riKuanvWr6Sqv68eCAFnCxhllNMCEn4VHs/77gCgQCo3qcEGOVuDXo9uNnR0Wo3ko1myUmv8i1+ltO7gxg+L0tvbf0QbBmfDLm4BAAD//4CQfvX4AAAA\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/hosts\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/5ST7W6DIBRA//MUvEDJ/Vhqdt/GqcuWEjGszV5/wVEjlBa5/sLLOYknEakzYMCg1tq6obdf7ueqRML5f7a3+ns5n/KTWz764aIUAaC8dSAd4llI5B21Htx89c6eFtvPE5rbPE5+sO42mlWT7QsOyhxUcVDBwZmDKw5+dCCg/nX+MvnCR8RFiaK4pGcUlSiOS35GsVKfI4CEJ5YO01g7UVBBUY+dKLigqLfeK0LmMNXUKUQ76EXpFOId9Co0hp+DzNq4MfAdpfawd5Tbg0Z0bVkNuV2mAwG3y3wg3F8AAAD//+yy88FYBAAA\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/motd\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/6yRwYobMQyG7/MUOmYX4lzL3sJ2oYV2A01yaG+KrMyIeOxBkodOn74kk0MPaSg0/8WWkD4+4+f/T/MMf8l6v/u0+fb5x9tHWL++vm23sHn/8v3G4B3GP+fC2HViYJM59yAGA6pDOQKCMVXlCJKPiuZayatyuMVYpwRILqO4sAEqQ1+yeDnvY46QSttyDPc89hmrd0Xl13mHiM2gKPRi1fhsZq5CniYYtHRyEL+y/2D0OIGy1eQgGaIYyZAko06zBbc4i5Ycbns84G+b5YPTnOqBNbOzAZXsWhLDkDAzvJfITfMVM7YMo+DLZZY8weJ6eWoIk1CBxXxeWiMn1gJLOCCd6gCLufHUoLaFInTug72sVnO5NNaRNcxVsJHC6QPWn6HmyEqp1BjouHr8u38HAAD//04R3qpqAwAA\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"path\": \"/etc/sysctl.d/99-k8s.conf\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/4zQX27DIAwG8PeeIhcIyqos6h56FuQS01kFGxmWtrefsmXZHzE13wsS+iH7g7EYSlNvKFkvegUdm785Nk+7xQ2bnRP2BkIwiyY+V1y/xZ2UxjMuR8u+dRBCS6nAKWB+7IYFfjnG4ikUVMPeOmEuCu5iI9zWud1HKltqsp9vf7TuKq1HyvNQO99+uykamVCdxEjFRoyi99ov+myIpZC/mwg3+5ZR7RWKe53rru553+8PhyomzgXYrfzY9N3LsLugMgaTgMk1/2Ru/wtaYSuScm3R9wAAAP//X0LFgj8CAAA=\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/flatcar/update.conf\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"\",\n \"source\": \"data:,REBOOT_STRATEGY%3Doff%0A\",\n \"verification\": {}\n },\n \"mode\": 272\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/kubeadm\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubeadm\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/kubelet\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubelet\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/kubectl\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubectl\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/calicoctl\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/calicoctl\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/velero\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/velero\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/etc/kubernetes/kubeadm-init.yaml\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/6xV3W7jNhO951MQvNmbUD+WYie8y7dZfFgsWgSb/gFFLyhyJBOiSXVIOfY+fUHJNqKugwZFbcCQoTlHhzPnjORgfgEMxjtB+7EBqXdZfxcy4/N92UCUFemN04J+diZ+9K413YgyGu9I430MEeXwk+/BBUEo5TSma0FZv+/KrcpihfDi9ziEerVG3zJCKaUagkIzxOmpTEMrRxvPj6cX3plshsRoBWUFI85r+AqdSRUTAaHUyR0IqryL6C0frHRQEkoVmmeveoiCjs4cRJ7nOLo81UnjAPWryyx41RM6ibAQPx0iygfsplNRmh7KzSAoWxVFKepNITZluRYrIe7LWaCyY4iAXLsgKCuLrMjuy+y2ulkiWlEKcVvNmL234w74YMfOOK4NCspyP8TcmgYOoPKkBh1ECPlJ2Kk45DM2n8oYsV5J+/D0+ZPTgzcuJtlS7wGjCfCgNUIIb6tvjNNPHqOg67quiEqw1igZ4QscBWUtwEZVcKtaJdUG6s2mVmtolSzqdS3rlb6737RV0W6qoqjqQq9vN3fNrWqbWla6qhtGOOfk3Vb7OLdy6bbTdJ/ScC+npOz3VheFSN+VEGVR/CHSERhxEF889sZ1qRWD189j45IT/tYDKUS+rm+meaWffFWnjgTAvVHwBmgaYl4WdzfnOZ9x2oVHv5NmisBdyEanAZX1o86mEbHzOSzgD9LJDjDpg6XdWgsH/t+ZI1GmEXN5NsIHIT6keTwD7q8qACcbC1x2HUIno0eOfoypm5RFHGEmHdAfjlxZAy7y5BreGguC5hAX8obe5C16F/lrRKYwfk/Tw/HE8l6aHo4TDcKfI4S4BakBL6rkezXJi54lkbTWv4DmackEQb8XcAUzdZPPfwIfEFpzEPQ3/hV2PgKftgu/guvQj8MZJ+j8ucD+n+5eQY0BMKlbAC+onwMguRpACzFTU8iWOSxPOfwy1yxzePEQE4KRLUgbt9/+Z5x+WNw4rcN/CsOp6sfn+e3B3lyW/PVOZWQ2+NNk98cUjX+RDKKmdj+iSRmgLBxDhJ1mRI5xCy6mBXh6wUjn3XHnx0U6tKApCoTSF2i23vfXbiYuj+bbhWrnNQj664wgfwUAAP//F9AABn8HAAA=\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"path\": \"/etc/kubernetes/addons/kube-dns-fixed-svc.yaml\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/4yRwU7rMBBF9/6K+YCXKn6Uls62FVJ3FgU2iMXUGZAVJzb2JFL+HoWmEBYV9fLOuUejMUX3zCm70CL0WtWurRAOnHpnWTUsVJEQKoCWGkaouyMXVZunIEey5zQPWbhRAJ6O7PPYAajvckExzoo5sh1nMkRG2PouC6e9UQAu3lPj/GCCd3ZAeOCPziXedeQPQrb+QRxnhJe96Vf/YG/65asCsGcTwv+y1Lhcl7jWeoVvqBFvb+bItFxxkRxnulyUi41efAUxJPlunU5xusL4xiHCVASIKUiwwSM87cyUCaV3FjMHZ55CbPzL9bi9wtWwJGd/77XRV9smNLNnKyFd+sLPAAAA//8MLaPANAIAAA==\",\n \"verification\": {}\n },\n \"mode\": 420\n }\n ]\n },\n \"systemd\": {\n \"units\": [\n {\n \"contents\": \"[Unit]\\nDescription=Load necessary kernel modules\\nBefore=containerd.service kubeadm-init.service\\n\\n[Service]\\nType=oneshot\\n\\nExecStart=/usr/bin/modprobe br_netfilter\\nExecStart=/usr/bin/modprobe overlay\\nRemainAfterExit=yes\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"modules-load.service\"\n },\n {\n \"enabled\": true,\n \"name\": \"systemd-networkd-wait-online.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=containerd container runtime\\nAfter=network.target modules-load.service\\n\\n[Service]\\n#StandardOutput=journal+console\\n#StandardError=journal+console\\n\\nExecStart=/usr/bin/containerd\\nRestart=always\\nRestartSec=5\\nDelegate=yes\\nKillMode=process\\nOOMScoreAdjust=-999\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"containerd.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Set Timezone\\nAfter=network-online.target\\nWants=network-online.target\\n[Service]\\nStandardOutput=journal+console\\nStandardError=journal+console\\n\\nExecStart=/bin/sh -c 'echo \\\"setting timezone to Europe/Berlin\\\"'\\nStandardOutput=journal+console\\nStandardError=journal+console\\nType=oneshot\\nRestart=on-failure\\nExecStart=/usr/bin/timedatectl set-timezone Europe/Berlin\\nExecStart=/usr/bin/timedatectl set-ntp true \\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"set-timezone.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=kubelet, the Kubernetes Node Agent\\nDocumentation=https://kubernets.io/docs/home\\nWants=network-online.target\\nAfter=network-online.target\\n[Service]\\n#StandardOutput=journal+console\\n#StandardError=journal+console\\n\\n#EnvironmentFile=/run/metadata/coreos\\nEnvironment=\\\"KUBELET_KUBECONFIG_ARGS=--bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf\\\"\\nEnvironment=\\\"KUBELET_CONFIG_ARGS=--config=/var/lib/kubelet/config.yaml\\\"\\n# This is a file that \\\"kubeadm init\\\" and \\\"kubeadm join\\\" generates at runtime, populating the KUBELET_KUBEADM_ARGS variable dynamically\\nEnvironmentFile=-/var/lib/kubelet/kubeadm-flags.env\\nExecStart=/opt/bin/kubelet $KUBELET_KUBECONFIG_ARGS $KUBELET_CONFIG_ARGS $KUBELET_KUBEADM_ARGS $KUBELET_EXTRA_ARGS\\nRestart=always\\nStartLimitInterval=0\\nRestartSec=10\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"kubelet.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Kubeadm Init Cluster\\nAfter=network-online.target containerd.service kubelet.service\\nWants=network-online.target\\nConditionPathExists=!/etc/kubernetes/kubelet.conf\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\n\\nExecStart=/bin/sh -c 'echo \\\"kubeadm-init.service started...\\\"'\\n\\n# Environment\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin:/opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent-uds/\\n\\nExecStartPre=/bin/sleep 30s\\nExecStart=/bin/sh -c 'echo \\\"running kubeadm init...\\\"'\\nExecStart=/opt/bin/kubeadm init --upload-certs --config=/etc/kubernetes/kubeadm-init.yaml\\n\\n# copy files for kubectl\\nExecStart=/bin/sh -c 'echo \\\"copying files (admin.conf) to core home folder.\\\"'\\nExecStartPost=/usr/bin/mkdir -p /home/core/.kube\\nExecStartPost=/usr/bin/cp -i /etc/kubernetes/admin.conf /home/core/.kube/config\\nExecStartPost=/usr/bin/chown core:core /home/core/.kube/config\\n\\nExecStartPost=/usr/bin/systemctl disable kubeadm-init.service\\nRestart=on-failure\\nRestartSec=120s\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"kubeadm-init.service\"\n },\n {\n \"contents\": \"[Unit]\\nWants=kubeadm-init.service\\nAfter=kubeadm-init.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-calico.done\\n\\n\\n[Service]\\nStandardOutput=journal+console\\nStandardError=journal+console\\n\\nExecStart=/bin/sh -c 'echo \\\"install.calico.service started...\\\"'\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nExecStart=/bin/sh -c 'echo \\\"witing 30s...\\\"'\\nExecStart=/bin/sleep 30s\\nExecStart=/bin/sh -c 'echo \\\"create calico namespace...\\\"'\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/namespace.yaml\\nExecStart=/bin/sh -c 'echo \\\"install tigera operator...\\\"'\\nExecStart=-/opt/bin/kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/operator-crds.yaml\\nExecStart=-/opt/bin/kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/tigera-operator.yaml\\nExecStart=/bin/sh -c 'echo \\\"witing 60s...\\\"'\\nExecStart=/bin/sleep 60s\\nExecStart=/bin/sh -c 'echo \\\"witing for tigera operator... (20mini max)\\\"'\\nExecStart=/opt/bin/kubectl wait deployment -n tigera-operator tigera-operator --for condition=Available=True --timeout=1200s\\nExecStart=/bin/sh -c 'echo \\\"create clico custom ressources...\\\"'\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/custom-resources.yaml\\n\\nExecStart=/bin/sh -c 'echo \\\"witing 3m..\\\"'\\nExecStart=/bin/sleep 3m\\n#ExecStart=/bin/sh -c 'echo \\\"apply calico (calico-apiserver)...\\\"'\\n#ExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/calico.yaml\\n#ExecStart=/bin/sh -c 'echo \\\"witing 1m...\\\"'\\n#ExecStart=/bin/sleep 2m\\nExecStart=/bin/sh -c 'echo \\\"witing calico-apiserver... (20mini max)\\\"'\\nExecStart=/opt/bin/kubectl wait deployment -n calico-apiserver calico-apiserver --for condition=Available=True --timeout=1200s\\nExecStart=/bin/sh -c 'echo \\\"witing 120s...\\\"'\\nExecStart=/bin/sleep 2m\\nExecStart=/bin/sh -c 'echo \\\"apply calico-peers...\\\"'\\nExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/calico-peer.yaml\\nExecStart=/bin/sh -c 'echo \\\"witing 60s...\\\"'\\nExecStart=/bin/sleep 1m\\nExecStart=/bin/sh -c 'echo \\\"apply calico-ippools...\\\"'\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/ippools.yaml\\n\\n#ExecStart=/bin/sh -c 'echo \\\"witing for whisker..\\\"'\\n#ExecStart=/opt/bin/kubectl wait deployment -n calico-system whisker --for condition=Available=True --timeout=1200s\\n#ExecStart=/bin/sh -c 'echo \\\"port-forward -n calico-system service/whisker 8081:8081\\\"'\\n#ExecStart=/opt/bin/kubectl port-forward -n calico-system service/whisker 8081:8081\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-calico.done\\nExecStart=/usr/bin/systemctl disable install-calico.service\\n#RemainAfterExit=true\\nRestart=on-failure\\nRestartSec=120s\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"install-calico.service\"\n },\n {\n \"contents\": \"[Unit]\\nWants=kubeadm-init.service\\nAfter=kubeadm-init.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-ceph.done\\n\\n[Service]\\nStandardOutput=journal+console\\nStandardError=journal+console\\n\\nExecStart=/bin/sh -c 'echo \\\"install.ceph.service started...\\\"'\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\nType=oneshot\\n\\nStandardOutput=journal+console\\nStandardError=journal+console\\nExecStart=/bin/sh -c 'echo \\\"witing 30s...\\\"'\\nExecStart=/bin/sleep 30s\\nExecStart=/bin/sh -c 'echo \\\"create ceph namespace...\\\"'\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/namespace.yaml\\n\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-nodeplugin-rbac.yaml\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-provisioner-rbac.yaml\\n\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/secrets.yaml\\n\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/ceph-conf.yaml\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-cephfsplugin-provisioner.yaml\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-cephfsplugin.yaml\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-config-map.yaml\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-encryption-kms-config.yaml\\n\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csidriver.yaml\\n\\nExecStart=-/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/storage-class.yaml\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-ceph.done\\nExecStart=/usr/bin/systemctl disable install-ceph.service\\n#RemainAfterExit=true\\nRestart=on-failure\\nRestartSec=120s\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"install-ceph.service\"\n },\n {\n \"contents\": \"[Unit]\\nWants=install-ceph.service\\nAfter=install-ceph.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-gitea.done\\n\\n[Service]\\nStandardOutput=journal+console\\nStandardError=journal+console\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\nType=oneshot\\n\\nExecStart=/opt/bin/kubectl wait deployment -n ceph csi-cephfsplugin-provisioner --for condition=Available=True --timeout=1200s\\nExecStart=/bin/sleep 4m\\nExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/namespace.yaml\\nExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/secrets.yaml\\nExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/db.yaml\\nExecStart=/bin/sleep 60s\\nExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/adminer.yaml\\nExecStart=/opt/bin/kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/gitea.yaml\\nExecStart=/bin/sleep 3m\\nExecStart=/opt/bin/kubectl exec deploy/gitea -n gitea -- /bin/startup.sh\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-gitea.done\\nExecStart=/usr/bin/systemctl disable install-gitea.service\\nRestart=on-failure\\nRestartSec=120s\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"install-gitea.service\"\n },\n {\n \"contents\": \"[Unit]\\nWants=install-calico.service\\nAfter=install-calico.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-argocd.done\\n[Service]\\nStandardOutput=journal+console\\nStandardError=journal+console\\n\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\nType=oneshot\\n\\nExecStart=/opt/bin/kubectl wait deployment -n kube-system coredns --for condition=Available=True --timeout=600s\\n\\nExecStart=/bin/sleep 1m\\nExecStart=/opt/bin/kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/namespace.yaml\\nExecStart=/opt/bin/kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/install.yaml\\nExecStart=/opt/bin/kubectl wait deployment -n argocd argocd-server --for condition=Available=True --timeout=600s\\n\\nExecStart=/opt/bin/kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/repo.yaml\\nExecStart=/bin/sleep 10s\\nExecStart=/opt/bin/kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/apps.yaml\\n\\nExecStart=/bin/sleep 10s\\nExecStart=/opt/bin/kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/ingress.yaml\\n\\n\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/repos/k8aux-bootstrap.yaml\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/repos/k8aux-apps.yaml\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/apps/argocd.yaml\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/apps/calico.yaml\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/apps/rook-ceph.yaml\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/apps/gitea.yaml\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://aux-balancer.undercloud.cf:3000/undercloud/k8aux-bootstrap/raw/branch/main/argocd/argocd-secret.yaml\\n##ExecStart=/bin/sleep 10m\\n#ExecStart=/opt/bin/kubectl wait deployment -n gitea gitea --for condition=Available=True --timeout=4800s\\n#ExecStart=/bin/sleep 10m\\n#ExecStart=/opt/bin/kubectl apply -n argocd -f http://gitea.gitea.svc.k8aux.undercloud.cf:3000/undercloud/k8aux-apps/raw/branch/main/app-of-apps/app-of-apps.yaml\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-argocd.done\\nExecStart=/usr/bin/systemctl disable install-argocd.service\\nRestart=on-failure\\nRestartSec=120s\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"install-argocd.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Pin fixed dual-stack ClusterIPs for kube-dns, argocd-server and whisker\\nAfter=install-argocd.service install-calico.service kubeadm-init.service network-online.target\\nWants=install-argocd.service install-calico.service kubeadm-init.service network-online.target\\nConditionPathExists=!/var/lib/undercloud-stamps/pin-service-ips.done\\n\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=PATH=/usr/bin:/usr/sbin:/opt/bin\\nExecStart=/bin/sh -eu -c '\\\\\\n echo \\\"[pin-service-ips] waiting for API...\\\" ; \\\\\\n for i in $(seq 1 120); do kubectl get --raw=/readyz \\u003e/dev/null 2\\u003e\\u00261 \\u0026\\u0026 break; sleep 2; done ; \\\\\\n echo \\\"[pin-service-ips] ensure namespaces exist...\\\" ; \\\\\\n kubectl get ns kube-system \\u003e/dev/null ; \\\\\\n kubectl get ns argocd \\u003e/dev/null 2\\u003e\\u00261 || kubectl create ns argocd ; \\\\\\n kubectl get ns calico-system \\u003e/dev/null ; \\\\\\n echo \\\"[pin-service-ips] wait for coredns/argocd readiness (best effort)...\\\" ; \\\\\\n kubectl -n kube-system wait deploy coredns --for=condition=Available=True --timeout=300s || true ; \\\\\\n kubectl -n argocd wait deploy argocd-server --for=condition=Available=True --timeout=600s || true ; \\\\\\n echo \\\"[pin-service-ips] replace Services with fixed ClusterIPs...\\\" ; \\\\\\n kubectl -n kube-system delete svc kube-dns --ignore-not-found ; \\\\\\n kubectl apply -f /etc/kubernetes/addons/kube-dns-fixed-svc.yaml ; \\\\\\n kubectl -n argocd delete svc argocd-server --ignore-not-found ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/service.yaml ; \\\\\\n kubectl -n calico-system delete svc whisker --ignore-not-found || true ; \\\\\\n kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/whisker.yaml || true ; \\\\\\n echo \\\"[pin-service-ips] done.\\\" \\\\\\n'\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/pin-service-ips.done\\nRestart=on-failure\\nRestartSec=120s\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"pin-service-ips.service\"\n }\n ]\n }\n}", + "data": "{\n \"ignition\": {\n \"config\": {\n \"replace\": {\n \"verification\": {}\n }\n },\n \"proxy\": {},\n \"security\": {\n \"tls\": {}\n },\n \"timeouts\": {},\n \"version\": \"3.4.0\"\n },\n \"kernelArguments\": {},\n \"passwd\": {\n \"users\": [\n {\n \"name\": \"core\",\n \"sshAuthorizedKeys\": [\n \"ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIHHEAlPo3v4U67Y3411pTjIMkQxwlFWdXrBJkSzXenDH flatcar@undercloud\"\n ]\n }\n ]\n },\n \"storage\": {\n \"directories\": [\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/opt/bin\",\n \"user\": {},\n \"mode\": 493\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/opt/cni/bin\",\n \"user\": {},\n \"mode\": 755\n },\n {\n \"group\": {},\n \"path\": \"/etc/kubernetes/manifests\",\n \"user\": {},\n \"mode\": 493\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/install-calico\",\n \"user\": {},\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/var/lib/undercloud-stamps\",\n \"user\": {},\n \"mode\": 493\n }\n ],\n \"files\": [\n {\n \"group\": {},\n \"path\": \"/etc/hostname\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"\",\n \"source\": \"data:,control-plane1%0A\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"path\": \"/etc/systemd/network/00-eth.network\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/2SMMWuFMBRG9/yUDjE3FcVABsFSOlSkHcUhJLco1aQkUeu/f+h76PC407mH77SfKuq+I7WaUGLsXwhpa4yr878dKY3xGIL8MYyJ/bgQBSRZehrOGIg0ZyIHyC79riKuanvWr6Sqv68eCAFnCxhllNMCEn4VHs/77gCgQCo3qcEGOVuDXo9uNnR0Wo3ko1myUmv8i1+ltO7gxg+L0tvbf0QbBmfDLm4BAAD//4CQfvX4AAAA\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/hosts\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/5ST7W6DIBRA//MUvEDJ/Vhqdt/GqcuWEjGszV5/wVEjlBa5/sLLOYknEakzYMCg1tq6obdf7ueqRML5f7a3+ns5n/KTWz764aIUAaC8dSAd4llI5B21Htx89c6eFtvPE5rbPE5+sO42mlWT7QsOyhxUcVDBwZmDKw5+dCCg/nX+MvnCR8RFiaK4pGcUlSiOS35GsVKfI4CEJ5YO01g7UVBBUY+dKLigqLfeK0LmMNXUKUQ76EXpFOId9Co0hp+DzNq4MfAdpfawd5Tbg0Z0bVkNuV2mAwG3y3wg3F8AAAD//+yy88FYBAAA\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/motd\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/6yRwYobMQyG7/MUOmYX4lzL3sJ2oYV2A01yaG+KrMyIeOxBkodOn74kk0MPaSg0/8WWkD4+4+f/T/MMf8l6v/u0+fb5x9tHWL++vm23sHn/8v3G4B3GP+fC2HViYJM59yAGA6pDOQKCMVXlCJKPiuZayatyuMVYpwRILqO4sAEqQ1+yeDnvY46QSttyDPc89hmrd0Xl13mHiM2gKPRi1fhsZq5CniYYtHRyEL+y/2D0OIGy1eQgGaIYyZAko06zBbc4i5Ycbns84G+b5YPTnOqBNbOzAZXsWhLDkDAzvJfITfMVM7YMo+DLZZY8weJ6eWoIk1CBxXxeWiMn1gJLOCCd6gCLufHUoLaFInTug72sVnO5NNaRNcxVsJHC6QPWn6HmyEqp1BjouHr8u38HAAD//04R3qpqAwAA\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"path\": \"/etc/sysctl.d/99-k8s.conf\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/4zQX27DIAwG8PeeIhcIyqos6h56FuQS01kFGxmWtrefsmXZHzE13wsS+iH7g7EYSlNvKFkvegUdm785Nk+7xQ2bnRP2BkIwiyY+V1y/xZ2UxjMuR8u+dRBCS6nAKWB+7IYFfjnG4ikUVMPeOmEuCu5iI9zWud1HKltqsp9vf7TuKq1HyvNQO99+uykamVCdxEjFRoyi99ov+myIpZC/mwg3+5ZR7RWKe53rru553+8PhyomzgXYrfzY9N3LsLugMgaTgMk1/2Ru/wtaYSuScm3R9wAAAP//X0LFgj8CAAA=\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"overwrite\": true,\n \"path\": \"/etc/flatcar/update.conf\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"\",\n \"source\": \"data:,REBOOT_STRATEGY%3Doff%0A\",\n \"verification\": {}\n },\n \"mode\": 272\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/kubeadm\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubeadm\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/kubelet\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubelet\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/kubectl\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/kubectl\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/calicoctl\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/calicoctl\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/velero\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/velero\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/opt/bin/busybox\",\n \"user\": {},\n \"contents\": {\n \"source\": \"http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/k8s-binaries/busybox\",\n \"verification\": {}\n },\n \"mode\": 493\n },\n {\n \"group\": {},\n \"path\": \"/etc/systemd/journald.conf.d/10-forward-to-syslog.conf\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"\",\n \"source\": \"data:,%5BJournal%5D%0AForwardToSyslog%3Dyes%0A\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"path\": \"/etc/kubernetes/kubeadm-init.yaml\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/6xV3W7jNhO951MQvNmbUD+WYie8y7dZfFgsWgSb/gFFLyhyJBOiSXVIOfY+fUHJNqKugwZFbcCQoTlHhzPnjORgfgEMxjtB+7EBqXdZfxcy4/N92UCUFemN04J+diZ+9K413YgyGu9I430MEeXwk+/BBUEo5TSma0FZv+/KrcpihfDi9ziEerVG3zJCKaUagkIzxOmpTEMrRxvPj6cX3plshsRoBWUFI85r+AqdSRUTAaHUyR0IqryL6C0frHRQEkoVmmeveoiCjs4cRJ7nOLo81UnjAPWryyx41RM6ibAQPx0iygfsplNRmh7KzSAoWxVFKepNITZluRYrIe7LWaCyY4iAXLsgKCuLrMjuy+y2ulkiWlEKcVvNmL234w74YMfOOK4NCspyP8TcmgYOoPKkBh1ECPlJ2Kk45DM2n8oYsV5J+/D0+ZPTgzcuJtlS7wGjCfCgNUIIb6tvjNNPHqOg67quiEqw1igZ4QscBWUtwEZVcKtaJdUG6s2mVmtolSzqdS3rlb6737RV0W6qoqjqQq9vN3fNrWqbWla6qhtGOOfk3Vb7OLdy6bbTdJ/ScC+npOz3VheFSN+VEGVR/CHSERhxEF889sZ1qRWD189j45IT/tYDKUS+rm+meaWffFWnjgTAvVHwBmgaYl4WdzfnOZ9x2oVHv5NmisBdyEanAZX1o86mEbHzOSzgD9LJDjDpg6XdWgsH/t+ZI1GmEXN5NsIHIT6keTwD7q8qACcbC1x2HUIno0eOfoypm5RFHGEmHdAfjlxZAy7y5BreGguC5hAX8obe5C16F/lrRKYwfk/Tw/HE8l6aHo4TDcKfI4S4BakBL6rkezXJi54lkbTWv4DmackEQb8XcAUzdZPPfwIfEFpzEPQ3/hV2PgKftgu/guvQj8MZJ+j8ucD+n+5eQY0BMKlbAC+onwMguRpACzFTU8iWOSxPOfwy1yxzePEQE4KRLUgbt9/+Z5x+WNw4rcN/CsOp6sfn+e3B3lyW/PVOZWQ2+NNk98cUjX+RDKKmdj+iSRmgLBxDhJ1mRI5xCy6mBXh6wUjn3XHnx0U6tKApCoTSF2i23vfXbiYuj+bbhWrnNQj664wgfwUAAP//F9AABn8HAAA=\",\n \"verification\": {}\n },\n \"mode\": 420\n },\n {\n \"group\": {},\n \"path\": \"/etc/kubernetes/addons/kube-dns-fixed-svc.yaml\",\n \"user\": {},\n \"contents\": {\n \"compression\": \"gzip\",\n \"source\": \"data:;base64,H4sIAAAAAAAC/4yRwU7rMBBF9/6K+YCXKn6Uls62FVJ3FgU2iMXUGZAVJzb2JFL+HoWmEBYV9fLOuUejMUX3zCm70CL0WtWurRAOnHpnWTUsVJEQKoCWGkaouyMXVZunIEey5zQPWbhRAJ6O7PPYAajvckExzoo5sh1nMkRG2PouC6e9UQAu3lPj/GCCd3ZAeOCPziXedeQPQrb+QRxnhJe96Vf/YG/65asCsGcTwv+y1Lhcl7jWeoVvqBFvb+bItFxxkRxnulyUi41efAUxJPlunU5xusL4xiHCVASIKUiwwSM87cyUCaV3FjMHZ55CbPzL9bi9wtWwJGd/77XRV9smNLNnKyFd+sLPAAAA//8MLaPANAIAAA==\",\n \"verification\": {}\n },\n \"mode\": 420\n }\n ]\n },\n \"systemd\": {\n \"units\": [\n {\n \"contents\": \"[Unit]\\nDescription=Forward journald to central syslog server\\nWants=network-online.target\\nAfter=network-online.target systemd-journald.service\\n\\n[Service]\\nType=simple\\nExecStart=/opt/bin/busybox syslogd -n -R syslog.undercloud.local -P 514\\nRestart=always\\nRestartSec=2\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"syslog-forward.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Undercloud Bootstrap Chain\\nWants=network-online.target\\nAfter=network-online.target\\n# Start the chain entry\\nWants=containerd.service kubelet.service kubeadm-init.service\\nAfter=containerd.service kubelet.service kubeadm-init.service\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"undercloud-bootstrap.target\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Load necessary kernel modules\\nBefore=containerd.service kubeadm-init.service\\n\\n[Service]\\nType=oneshot\\nExecStart=/usr/bin/modprobe br_netfilter\\nExecStart=/usr/bin/modprobe overlay\\nRemainAfterExit=yes\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"modules-load.service\"\n },\n {\n \"enabled\": true,\n \"name\": \"systemd-networkd-wait-online.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=containerd container runtime\\nAfter=network.target modules-load.service\\nWants=modules-load.service\\n\\n[Service]\\nExecStart=/usr/bin/containerd\\nRestart=always\\nRestartSec=5\\nDelegate=yes\\nKillMode=process\\nOOMScoreAdjust=-999\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"containerd.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Set Timezone\\nAfter=network-online.target\\nWants=network-online.target\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nExecStart=/bin/sh -c 'echo \\\"setting timezone to Europe/Berlin\\\"'\\nExecStart=/usr/bin/timedatectl set-timezone Europe/Berlin\\nExecStart=/usr/bin/timedatectl set-ntp true\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"set-timezone.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=kubelet, the Kubernetes Node Agent\\nDocumentation=https://kubernetes.io/docs/home\\nWants=network-online.target\\nAfter=network-online.target containerd.service\\nRequires=containerd.service\\n\\n[Service]\\nEnvironment=\\\"KUBELET_KUBECONFIG_ARGS=--bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf\\\"\\nEnvironment=\\\"KUBELET_CONFIG_ARGS=--config=/var/lib/kubelet/config.yaml\\\"\\nEnvironmentFile=-/var/lib/kubelet/kubeadm-flags.env\\nExecStart=/opt/bin/kubelet $KUBELET_KUBECONFIG_ARGS $KUBELET_CONFIG_ARGS $KUBELET_KUBEADM_ARGS $KUBELET_EXTRA_ARGS\\nRestart=always\\nStartLimitInterval=0\\nRestartSec=10\\n\\n[Install]\\nWantedBy=multi-user.target\\n\",\n \"enabled\": true,\n \"name\": \"kubelet.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Kubeadm Init Cluster\\nWants=network-online.target\\nAfter=network-online.target containerd.service kubelet.service\\nRequires=containerd.service kubelet.service\\nConditionPathExists=!/etc/kubernetes/kubelet.conf\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\n\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin:/opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent-uds/\\n\\nExecStart=/bin/sh -eu -c '\\\\\\n echo \\\"[kubeadm-init] started...\\\" ; \\\\\\n echo \\\"[kubeadm-init] waiting for containerd socket...\\\" ; \\\\\\n for i in $(seq 1 60); do test -S /run/containerd/containerd.sock \\u0026\\u0026 break; sleep 1; done ; \\\\\\n echo \\\"[kubeadm-init] running kubeadm init...\\\" ; \\\\\\n /opt/bin/kubeadm init --upload-certs --config=/etc/kubernetes/kubeadm-init.yaml ; \\\\\\n echo \\\"[kubeadm-init] copying kubeconfig to core...\\\" ; \\\\\\n mkdir -p /home/core/.kube ; \\\\\\n cp -f /etc/kubernetes/admin.conf /home/core/.kube/config ; \\\\\\n chown core:core /home/core/.kube/config ; \\\\\\n echo \\\"[kubeadm-init] done.\\\" \\\\\\n'\\n\\n# strictly start next step (serialization)\\nExecStartPost=/usr/bin/systemctl start install-calico.service\\n\\n[Install]\\nWantedBy=undercloud-bootstrap.target\\n\",\n \"enabled\": false,\n \"name\": \"kubeadm-init.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Install Calico\\nRequires=kubeadm-init.service\\nAfter=kubeadm-init.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-calico.done\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\n\\nExecStart=/bin/sh -eu -c '\\\\\\n echo \\\"[calico] waiting for API /readyz...\\\" ; \\\\\\n for i in $(seq 1 180); do kubectl get --raw=/readyz \\u003e/dev/null 2\\u003e\\u00261 \\u0026\\u0026 break; sleep 2; done ; \\\\\\n echo \\\"[calico] create namespace + operator...\\\" ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/namespace.yaml ; \\\\\\n kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/operator-crds.yaml || true ; \\\\\\n kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/tigera-operator.yaml || true ; \\\\\\n echo \\\"[calico] wait for tigera-operator...\\\" ; \\\\\\n kubectl wait deployment -n tigera-operator tigera-operator --for condition=Available=True --timeout=1200s ; \\\\\\n echo \\\"[calico] apply custom resources...\\\" ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/custom-resources.yaml ; \\\\\\n echo \\\"[calico] wait for calico-apiserver...\\\" ; \\\\\\n kubectl wait deployment -n calico-apiserver calico-apiserver --for condition=Available=True --timeout=1200s ; \\\\\\n echo \\\"[calico] apply peers + pools...\\\" ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/calico-peer.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/ippools.yaml ; \\\\\\n echo \\\"[calico] done.\\\" \\\\\\n'\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-calico.done\\nExecStartPost=/usr/bin/systemctl start install-ceph.service\\n\\n[Install]\\nWantedBy=undercloud-bootstrap.target\\n\",\n \"enabled\": false,\n \"name\": \"install-calico.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Install Ceph CSI\\nRequires=install-calico.service\\nAfter=install-calico.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-ceph.done\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\n\\nExecStart=/bin/sh -eu -c '\\\\\\n echo \\\"[ceph] waiting for API /readyz...\\\" ; \\\\\\n for i in $(seq 1 180); do kubectl get --raw=/readyz \\u003e/dev/null 2\\u003e\\u00261 \\u0026\\u0026 break; sleep 2; done ; \\\\\\n echo \\\"[ceph] apply manifests...\\\" ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/namespace.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-nodeplugin-rbac.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-provisioner-rbac.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/secrets.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/ceph-conf.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-cephfsplugin-provisioner.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-cephfsplugin.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-config-map.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csi-encryption-kms-config.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/csidriver.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/ceph/storage-class.yaml ; \\\\\\n echo \\\"[ceph] done.\\\" \\\\\\n'\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-ceph.done\\nExecStartPost=/usr/bin/systemctl start install-gitea.service\\n\\n[Install]\\nWantedBy=undercloud-bootstrap.target\\n\",\n \"enabled\": false,\n \"name\": \"install-ceph.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Install Gitea\\nRequires=install-ceph.service\\nAfter=install-ceph.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-gitea.done\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\n\\nExecStart=/bin/sh -eu -c '\\\\\\n echo \\\"[gitea] wait for ceph provisioner...\\\" ; \\\\\\n kubectl wait deployment -n ceph csi-cephfsplugin-provisioner --for condition=Available=True --timeout=1200s ; \\\\\\n echo \\\"[gitea] apply manifests...\\\" ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/namespace.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/secrets.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/db.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/adminer.yaml ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/gitea/gitea.yaml ; \\\\\\n echo \\\"[gitea] wait for gitea deployment...\\\" ; \\\\\\n kubectl -n gitea wait deployment gitea --for=condition=Available=True --timeout=1200s ; \\\\\\n echo \\\"[gitea] run startup...\\\" ; \\\\\\n kubectl exec deploy/gitea -n gitea -- /bin/startup.sh ; \\\\\\n echo \\\"[gitea] done.\\\" \\\\\\n'\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-gitea.done\\nExecStartPost=/usr/bin/systemctl start install-argocd.service\\n\\n[Install]\\nWantedBy=undercloud-bootstrap.target\\n\",\n \"enabled\": false,\n \"name\": \"install-gitea.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Install ArgoCD\\nRequires=install-calico.service install-gitea.service\\nAfter=install-calico.service install-gitea.service\\nConditionPathExists=!/var/lib/undercloud-stamps/install-argocd.done\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=DATASTORE_TYPE=kubernetes\\nEnvironment=PATH=/usr/bin/:/usr/sbin:/opt/bin\\n\\nExecStart=/bin/sh -eu -c '\\\\\\n echo \\\"[argocd] wait for coredns...\\\" ; \\\\\\n kubectl -n kube-system wait deploy coredns --for=condition=Available=True --timeout=1200s ; \\\\\\n echo \\\"[argocd] install...\\\" ; \\\\\\n kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/namespace.yaml ; \\\\\\n kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/install.yaml ; \\\\\\n kubectl -n argocd wait deploy argocd-server --for=condition=Available=True --timeout=1200s ; \\\\\\n kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/repo.yaml ; \\\\\\n kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/apps.yaml ; \\\\\\n kubectl apply -n argocd -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/ingress.yaml ; \\\\\\n echo \\\"[argocd] done.\\\" \\\\\\n'\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/install-argocd.done\\nExecStartPost=/usr/bin/systemctl start pin-service-ips.service\\n\\n[Install]\\nWantedBy=undercloud-bootstrap.target\\n\",\n \"enabled\": false,\n \"name\": \"install-argocd.service\"\n },\n {\n \"contents\": \"[Unit]\\nDescription=Pin fixed dual-stack ClusterIPs for kube-dns, argocd-server and whisker\\nRequires=install-argocd.service install-calico.service kubeadm-init.service\\nAfter=install-argocd.service install-calico.service kubeadm-init.service network-online.target\\nWants=network-online.target\\nConditionPathExists=!/var/lib/undercloud-stamps/pin-service-ips.done\\n\\n[Service]\\nType=oneshot\\nStandardOutput=journal+console\\nStandardError=journal+console\\nEnvironment=KUBECONFIG=/etc/kubernetes/admin.conf\\nEnvironment=PATH=/usr/bin:/usr/sbin:/opt/bin\\nExecStart=/bin/sh -eu -c '\\\\\\n echo \\\"[pin-service-ips] waiting for API...\\\" ; \\\\\\n for i in $(seq 1 120); do kubectl get --raw=/readyz \\u003e/dev/null 2\\u003e\\u00261 \\u0026\\u0026 break; sleep 2; done ; \\\\\\n echo \\\"[pin-service-ips] ensure namespaces exist...\\\" ; \\\\\\n kubectl get ns kube-system \\u003e/dev/null ; \\\\\\n kubectl get ns argocd \\u003e/dev/null 2\\u003e\\u00261 || kubectl create ns argocd ; \\\\\\n kubectl get ns calico-system \\u003e/dev/null ; \\\\\\n echo \\\"[pin-service-ips] wait for coredns/argocd readiness (best effort)...\\\" ; \\\\\\n kubectl -n kube-system wait deploy coredns --for=condition=Available=True --timeout=300s || true ; \\\\\\n kubectl -n argocd wait deploy argocd-server --for=condition=Available=True --timeout=600s || true ; \\\\\\n echo \\\"[pin-service-ips] replace Services with fixed ClusterIPs...\\\" ; \\\\\\n kubectl -n kube-system delete svc kube-dns --ignore-not-found ; \\\\\\n kubectl apply -f /etc/kubernetes/addons/kube-dns-fixed-svc.yaml ; \\\\\\n kubectl -n argocd delete svc argocd-server --ignore-not-found ; \\\\\\n kubectl apply -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/argocd/service.yaml ; \\\\\\n kubectl -n calico-system delete svc whisker --ignore-not-found || true ; \\\\\\n kubectl create -f http://git.undercloud.local:3000/Undercloud/undercloud-infrastructure/raw/branch/main/calico-config/whisker.yaml || true ; \\\\\\n echo \\\"[pin-service-ips] done.\\\" \\\\\\n'\\n\\nExecStartPost=/usr/bin/touch /var/lib/undercloud-stamps/pin-service-ips.done\\n\\n[Install]\\nWantedBy=undercloud-bootstrap.target\\n\",\n \"enabled\": false,\n \"name\": \"pin-service-ips.service\"\n }\n ]\n }\n}", "file_name": "control-plane1-ignition-user-data", "resize": 0 } @@ -610,14 +548,14 @@ [ "2001:470:7116:2::91", "fd00:0:0:2::91", - "fe80::be24:11ff:fe0e:f6cf" + "fe80::be24:11ff:fe25:70c3" ] ], "keyboard_layout": "en-us", "kvm_arguments": null, "mac_addresses": [ "00:00:00:00:00:00", - "BC:24:11:0E:F6:CF" + "BC:24:11:25:70:C3" ], "machine": null, "memory": [ @@ -637,7 +575,7 @@ "disconnected": false, "enabled": true, "firewall": false, - "mac_address": "BC:24:11:0E:F6:CF", + "mac_address": "BC:24:11:25:70:C3", "model": "virtio", "mtu": 0, "queues": 0, @@ -706,358 +644,6 @@ } ] }, - { - "mode": "managed", - "type": "proxmox_virtual_environment_vm", - "name": "control_plane2", - "provider": "provider[\"registry.terraform.io/bpg/proxmox\"]", - "instances": [ - { - "schema_version": 0, - "attributes": { - "acpi": true, - "agent": [], - "amd_sev": [], - "audio_device": [], - "bios": "seabios", - "boot_order": [ - "virtio0" - ], - "cdrom": [], - "clone": [ - { - "datastore_id": "", - "full": true, - "node_name": "hyper1", - "retries": 1, - "vm_id": 999 - } - ], - "cpu": [ - { - "affinity": "", - "architecture": "", - "cores": 2, - "flags": null, - "hotplugged": 0, - "limit": 0, - "numa": false, - "sockets": 1, - "type": "host", - "units": 1024 - } - ], - "description": "kubernetes control-plane2", - "disk": [], - "efi_disk": [], - "hook_script_file_id": null, - "hostpci": [], - "id": "1002", - "initialization": [ - { - "datastore_id": "Pool1", - "dns": [], - "interface": "ide2", - "ip_config": [], - "meta_data_file_id": "", - "network_data_file_id": "", - "type": "", - "user_account": [], - "user_data_file_id": "cephfs:snippets/control-plane2-ignition-user-data", - "vendor_data_file_id": "" - } - ], - "ipv4_addresses": [ - [ - "127.0.0.1" - ], - [ - "10.0.2.92" - ] - ], - "ipv6_addresses": [ - [ - "::1" - ], - [ - "2001:470:7116:2::92", - "fd00:0:0:2::92", - "fe80::be24:11ff:fe4b:b983" - ] - ], - "keyboard_layout": "en-us", - "kvm_arguments": null, - "mac_addresses": [ - "00:00:00:00:00:00", - "BC:24:11:4B:B9:83" - ], - "machine": null, - "memory": [ - { - "dedicated": 3072, - "floating": 3072, - "hugepages": "", - "keep_hugepages": false, - "shared": 0 - } - ], - "migrate": false, - "name": "control-plane2", - "network_device": [ - { - "bridge": "vmbr0", - "disconnected": false, - "enabled": true, - "firewall": false, - "mac_address": "BC:24:11:4B:B9:83", - "model": "virtio", - "mtu": 0, - "queues": 0, - "rate_limit": 0, - "trunks": "", - "vlan_id": 0 - } - ], - "network_interface_names": [ - "lo", - "eth0" - ], - "node_name": "hyper2", - "numa": [], - "on_boot": true, - "operating_system": [], - "pool_id": null, - "protection": false, - "reboot": false, - "reboot_after_update": true, - "rng": [], - "scsi_hardware": "virtio-scsi-pci", - "serial_device": [], - "smbios": [], - "started": true, - "startup": [ - { - "down_delay": -1, - "order": 1, - "up_delay": -1 - } - ], - "stop_on_destroy": false, - "tablet_device": true, - "tags": [ - "control-plane", - "flatcar", - "kubernetes", - "terraform" - ], - "template": false, - "timeout_clone": 1800, - "timeout_create": 1800, - "timeout_migrate": 1800, - "timeout_move_disk": 1800, - "timeout_reboot": 1800, - "timeout_shutdown_vm": 1800, - "timeout_start_vm": 1800, - "timeout_stop_vm": 300, - "tpm_state": [], - "usb": [], - "vga": [], - "virtiofs": [], - "vm_id": 1002, - "watchdog": [] - }, - "sensitive_attributes": [], - "identity_schema_version": 0, - "private": "bnVsbA==", - "dependencies": [ - "data.ct_config.control_plane1_ignition", - "data.ct_config.control_plane2_ignition", - "null_resource.wait_for_cp1_api", - "proxmox_virtual_environment_download_file.flatcar_image", - "proxmox_virtual_environment_file.control_plane1_ignition", - "proxmox_virtual_environment_file.control_plane2_ignition", - "proxmox_virtual_environment_vm.control_plane1", - "proxmox_virtual_environment_vm.flatcar_template" - ] - } - ] - }, - { - "mode": "managed", - "type": "proxmox_virtual_environment_vm", - "name": "control_plane3", - "provider": "provider[\"registry.terraform.io/bpg/proxmox\"]", - "instances": [ - { - "schema_version": 0, - "attributes": { - "acpi": true, - "agent": [], - "amd_sev": [], - "audio_device": [], - "bios": "seabios", - "boot_order": [ - "virtio0" - ], - "cdrom": [], - "clone": [ - { - "datastore_id": "", - "full": true, - "node_name": "hyper1", - "retries": 1, - "vm_id": 999 - } - ], - "cpu": [ - { - "affinity": "", - "architecture": "", - "cores": 2, - "flags": null, - "hotplugged": 0, - "limit": 0, - "numa": false, - "sockets": 1, - "type": "host", - "units": 1024 - } - ], - "description": "kubernetes control-plane3", - "disk": [], - "efi_disk": [], - "hook_script_file_id": null, - "hostpci": [], - "id": "1003", - "initialization": [ - { - "datastore_id": "Pool1", - "dns": [], - "interface": "ide2", - "ip_config": [], - "meta_data_file_id": "", - "network_data_file_id": "", - "type": "", - "user_account": [], - "user_data_file_id": "cephfs:snippets/control-plane3-ignition-user-data", - "vendor_data_file_id": "" - } - ], - "ipv4_addresses": [ - [ - "127.0.0.1" - ], - [ - "10.0.2.93" - ] - ], - "ipv6_addresses": [ - [ - "::1" - ], - [ - "2001:470:7116:2::93", - "fd00:0:0:2::93", - "fe80::be24:11ff:fecc:133d" - ] - ], - "keyboard_layout": "en-us", - "kvm_arguments": null, - "mac_addresses": [ - "00:00:00:00:00:00", - "BC:24:11:CC:13:3D" - ], - "machine": null, - "memory": [ - { - "dedicated": 3072, - "floating": 3072, - "hugepages": "", - "keep_hugepages": false, - "shared": 0 - } - ], - "migrate": false, - "name": "control-plane3", - "network_device": [ - { - "bridge": "vmbr0", - "disconnected": false, - "enabled": true, - "firewall": false, - "mac_address": "BC:24:11:CC:13:3D", - "model": "virtio", - "mtu": 0, - "queues": 0, - "rate_limit": 0, - "trunks": "", - "vlan_id": 0 - } - ], - "network_interface_names": [ - "lo", - "eth0" - ], - "node_name": "hyper3", - "numa": [], - "on_boot": true, - "operating_system": [], - "pool_id": null, - "protection": false, - "reboot": false, - "reboot_after_update": true, - "rng": [], - "scsi_hardware": "virtio-scsi-pci", - "serial_device": [], - "smbios": [], - "started": true, - "startup": [ - { - "down_delay": -1, - "order": 1, - "up_delay": -1 - } - ], - "stop_on_destroy": false, - "tablet_device": true, - "tags": [ - "control-plane", - "flatcar", - "kubernetes", - "terraform" - ], - "template": false, - "timeout_clone": 1800, - "timeout_create": 1800, - "timeout_migrate": 1800, - "timeout_move_disk": 1800, - "timeout_reboot": 1800, - "timeout_shutdown_vm": 1800, - "timeout_start_vm": 1800, - "timeout_stop_vm": 300, - "tpm_state": [], - "usb": [], - "vga": [], - "virtiofs": [], - "vm_id": 1003, - "watchdog": [] - }, - "sensitive_attributes": [], - "identity_schema_version": 0, - "private": "bnVsbA==", - "dependencies": [ - "data.ct_config.control_plane1_ignition", - "data.ct_config.control_plane3_ignition", - "null_resource.wait_for_cp1_api", - "proxmox_virtual_environment_download_file.flatcar_image", - "proxmox_virtual_environment_file.control_plane1_ignition", - "proxmox_virtual_environment_file.control_plane3_ignition", - "proxmox_virtual_environment_vm.control_plane1", - "proxmox_virtual_environment_vm.flatcar_template" - ] - } - ] - }, { "mode": "managed", "type": "proxmox_virtual_environment_vm", @@ -1143,7 +729,7 @@ "keyboard_layout": "en-us", "kvm_arguments": "", "mac_addresses": [ - "BC:24:11:AC:1F:D9" + "BC:24:11:78:08:35" ], "machine": "", "memory": [ @@ -1163,7 +749,7 @@ "disconnected": false, "enabled": true, "firewall": false, - "mac_address": "BC:24:11:AC:1F:D9", + "mac_address": "BC:24:11:78:08:35", "model": "virtio", "mtu": 0, "queues": 0, @@ -1218,742 +804,6 @@ ] } ] - }, - { - "mode": "managed", - "type": "proxmox_virtual_environment_vm", - "name": "worker1", - "provider": "provider[\"registry.terraform.io/bpg/proxmox\"]", - "instances": [ - { - "schema_version": 0, - "attributes": { - "acpi": true, - "agent": [], - "amd_sev": [], - "audio_device": [], - "bios": "seabios", - "boot_order": [ - "virtio0" - ], - "cdrom": [], - "clone": [ - { - "datastore_id": "", - "full": true, - "node_name": "hyper1", - "retries": 1, - "vm_id": 999 - } - ], - "cpu": [ - { - "affinity": "", - "architecture": "", - "cores": 2, - "flags": null, - "hotplugged": 0, - "limit": 0, - "numa": false, - "sockets": 1, - "type": "host", - "units": 1024 - } - ], - "description": "kubernetes worker1", - "disk": [], - "efi_disk": [], - "hook_script_file_id": null, - "hostpci": [], - "id": "1011", - "initialization": [ - { - "datastore_id": "Pool1", - "dns": [], - "interface": "ide2", - "ip_config": [], - "meta_data_file_id": "", - "network_data_file_id": "", - "type": "", - "user_account": [], - "user_data_file_id": "cephfs:snippets/worker1-ignition-user-data", - "vendor_data_file_id": "" - } - ], - "ipv4_addresses": [ - [ - "127.0.0.1" - ], - [ - "10.0.2.101" - ] - ], - "ipv6_addresses": [ - [ - "::1" - ], - [ - "2001:470:7116:2::101", - "fd00:0:0:2::101", - "fe80::be24:11ff:feab:b526" - ] - ], - "keyboard_layout": "en-us", - "kvm_arguments": null, - "mac_addresses": [ - "00:00:00:00:00:00", - "BC:24:11:AB:B5:26" - ], - "machine": null, - "memory": [ - { - "dedicated": 10240, - "floating": 10240, - "hugepages": "", - "keep_hugepages": false, - "shared": 0 - } - ], - "migrate": false, - "name": "worker1", - "network_device": [ - { - "bridge": "vmbr0", - "disconnected": false, - "enabled": true, - "firewall": false, - "mac_address": "BC:24:11:AB:B5:26", - "model": "virtio", - "mtu": 0, - "queues": 0, - "rate_limit": 0, - "trunks": "", - "vlan_id": 0 - } - ], - "network_interface_names": [ - "lo", - "eth0" - ], - "node_name": "hyper1", - "numa": [], - "on_boot": true, - "operating_system": [], - "pool_id": null, - "protection": false, - "reboot": false, - "reboot_after_update": true, - "rng": [], - "scsi_hardware": "virtio-scsi-pci", - "serial_device": [], - "smbios": [], - "started": true, - "startup": [ - { - "down_delay": -1, - "order": 1, - "up_delay": -1 - } - ], - "stop_on_destroy": false, - "tablet_device": true, - "tags": [ - "flatcar", - "kubernetes", - "terraform", - "worker" - ], - "template": false, - "timeout_clone": 1800, - "timeout_create": 1800, - "timeout_migrate": 1800, - "timeout_move_disk": 1800, - "timeout_reboot": 1800, - "timeout_shutdown_vm": 1800, - "timeout_start_vm": 1800, - "timeout_stop_vm": 300, - "tpm_state": [], - "usb": [], - "vga": [], - "virtiofs": [], - "vm_id": 1011, - "watchdog": [] - }, - "sensitive_attributes": [], - "identity_schema_version": 0, - "private": "bnVsbA==", - "dependencies": [ - "data.ct_config.control_plane1_ignition", - "data.ct_config.control_plane2_ignition", - "data.ct_config.control_plane3_ignition", - "data.ct_config.worker1_ignition", - "null_resource.wait_for_cluster_ready", - "null_resource.wait_for_cp1_api", - "null_resource.wait_for_cp2_cp3_api", - "proxmox_virtual_environment_download_file.flatcar_image", - "proxmox_virtual_environment_file.control_plane1_ignition", - "proxmox_virtual_environment_file.control_plane2_ignition", - "proxmox_virtual_environment_file.control_plane3_ignition", - "proxmox_virtual_environment_file.worker1_ignition", - "proxmox_virtual_environment_vm.control_plane1", - "proxmox_virtual_environment_vm.control_plane2", - "proxmox_virtual_environment_vm.control_plane3", - "proxmox_virtual_environment_vm.flatcar_template" - ] - } - ] - }, - { - "mode": "managed", - "type": "proxmox_virtual_environment_vm", - "name": "worker2", - "provider": "provider[\"registry.terraform.io/bpg/proxmox\"]", - "instances": [ - { - "schema_version": 0, - "attributes": { - "acpi": true, - "agent": [], - "amd_sev": [], - "audio_device": [], - "bios": "seabios", - "boot_order": [ - "virtio0" - ], - "cdrom": [], - "clone": [ - { - "datastore_id": "", - "full": true, - "node_name": "hyper1", - "retries": 1, - "vm_id": 999 - } - ], - "cpu": [ - { - "affinity": "", - "architecture": "", - "cores": 2, - "flags": null, - "hotplugged": 0, - "limit": 0, - "numa": false, - "sockets": 1, - "type": "host", - "units": 1024 - } - ], - "description": "kubernetes worker2", - "disk": [], - "efi_disk": [], - "hook_script_file_id": null, - "hostpci": [], - "id": "1012", - "initialization": [ - { - "datastore_id": "Pool1", - "dns": [], - "interface": "ide2", - "ip_config": [], - "meta_data_file_id": "", - "network_data_file_id": "", - "type": "", - "user_account": [], - "user_data_file_id": "cephfs:snippets/worker2-ignition-user-data", - "vendor_data_file_id": "" - } - ], - "ipv4_addresses": [ - [ - "127.0.0.1" - ], - [ - "10.0.2.102" - ] - ], - "ipv6_addresses": [ - [ - "::1" - ], - [ - "2001:470:7116:2::102", - "fd00:0:0:2::102", - "fe80::be24:11ff:fec8:39fd" - ] - ], - "keyboard_layout": "en-us", - "kvm_arguments": null, - "mac_addresses": [ - "00:00:00:00:00:00", - "BC:24:11:C8:39:FD" - ], - "machine": null, - "memory": [ - { - "dedicated": 10240, - "floating": 10240, - "hugepages": "", - "keep_hugepages": false, - "shared": 0 - } - ], - "migrate": false, - "name": "worker2", - "network_device": [ - { - "bridge": "vmbr0", - "disconnected": false, - "enabled": true, - "firewall": false, - "mac_address": "BC:24:11:C8:39:FD", - "model": "virtio", - "mtu": 0, - "queues": 0, - "rate_limit": 0, - "trunks": "", - "vlan_id": 0 - } - ], - "network_interface_names": [ - "lo", - "eth0" - ], - "node_name": "hyper2", - "numa": [], - "on_boot": true, - "operating_system": [], - "pool_id": null, - "protection": false, - "reboot": false, - "reboot_after_update": true, - "rng": [], - "scsi_hardware": "virtio-scsi-pci", - "serial_device": [], - "smbios": [], - "started": true, - "startup": [ - { - "down_delay": -1, - "order": 1, - "up_delay": -1 - } - ], - "stop_on_destroy": false, - "tablet_device": true, - "tags": [ - "flatcar", - "kubernetes", - "terraform", - "worker" - ], - "template": false, - "timeout_clone": 1800, - "timeout_create": 1800, - "timeout_migrate": 1800, - "timeout_move_disk": 1800, - "timeout_reboot": 1800, - "timeout_shutdown_vm": 1800, - "timeout_start_vm": 1800, - "timeout_stop_vm": 300, - "tpm_state": [], - "usb": [], - "vga": [], - "virtiofs": [], - "vm_id": 1012, - "watchdog": [] - }, - "sensitive_attributes": [], - "identity_schema_version": 0, - "private": "bnVsbA==", - "dependencies": [ - "data.ct_config.control_plane1_ignition", - "data.ct_config.control_plane2_ignition", - "data.ct_config.control_plane3_ignition", - "data.ct_config.worker2_ignition", - "null_resource.wait_for_cluster_ready", - "null_resource.wait_for_cp1_api", - "null_resource.wait_for_cp2_cp3_api", - "proxmox_virtual_environment_download_file.flatcar_image", - "proxmox_virtual_environment_file.control_plane1_ignition", - "proxmox_virtual_environment_file.control_plane2_ignition", - "proxmox_virtual_environment_file.control_plane3_ignition", - "proxmox_virtual_environment_file.worker2_ignition", - "proxmox_virtual_environment_vm.control_plane1", - "proxmox_virtual_environment_vm.control_plane2", - "proxmox_virtual_environment_vm.control_plane3", - "proxmox_virtual_environment_vm.flatcar_template" - ] - } - ] - }, - { - "mode": "managed", - "type": "proxmox_virtual_environment_vm", - "name": "worker3", - "provider": "provider[\"registry.terraform.io/bpg/proxmox\"]", - "instances": [ - { - "schema_version": 0, - "attributes": { - "acpi": true, - "agent": [], - "amd_sev": [], - "audio_device": [], - "bios": "seabios", - "boot_order": [ - "virtio0" - ], - "cdrom": [], - "clone": [ - { - "datastore_id": "", - "full": true, - "node_name": "hyper1", - "retries": 1, - "vm_id": 999 - } - ], - "cpu": [ - { - "affinity": "", - "architecture": "", - "cores": 2, - "flags": null, - "hotplugged": 0, - "limit": 0, - "numa": false, - "sockets": 1, - "type": "host", - "units": 1024 - } - ], - "description": "kubernetes worker3", - "disk": [], - "efi_disk": [], - "hook_script_file_id": null, - "hostpci": [], - "id": "1013", - "initialization": [ - { - "datastore_id": "Pool1", - "dns": [], - "interface": "ide2", - "ip_config": [], - "meta_data_file_id": "", - "network_data_file_id": "", - "type": "", - "user_account": [], - "user_data_file_id": "cephfs:snippets/worker3-ignition-user-data", - "vendor_data_file_id": "" - } - ], - "ipv4_addresses": [ - [ - "127.0.0.1" - ], - [ - "10.0.2.103" - ] - ], - "ipv6_addresses": [ - [ - "::1" - ], - [ - "2001:470:7116:2::103", - "fd00:0:0:2::103", - "fe80::be24:11ff:febf:b7b5" - ] - ], - "keyboard_layout": "en-us", - "kvm_arguments": null, - "mac_addresses": [ - "00:00:00:00:00:00", - "BC:24:11:BF:B7:B5" - ], - "machine": null, - "memory": [ - { - "dedicated": 10240, - "floating": 10240, - "hugepages": "", - "keep_hugepages": false, - "shared": 0 - } - ], - "migrate": false, - "name": "worker3", - "network_device": [ - { - "bridge": "vmbr0", - "disconnected": false, - "enabled": true, - "firewall": false, - "mac_address": "BC:24:11:BF:B7:B5", - "model": "virtio", - "mtu": 0, - "queues": 0, - "rate_limit": 0, - "trunks": "", - "vlan_id": 0 - } - ], - "network_interface_names": [ - "lo", - "eth0" - ], - "node_name": "hyper3", - "numa": [], - "on_boot": true, - "operating_system": [], - "pool_id": null, - "protection": false, - "reboot": false, - "reboot_after_update": true, - "rng": [], - "scsi_hardware": "virtio-scsi-pci", - "serial_device": [], - "smbios": [], - "started": true, - "startup": [ - { - "down_delay": -1, - "order": 1, - "up_delay": -1 - } - ], - "stop_on_destroy": false, - "tablet_device": true, - "tags": [ - "flatcar", - "kubernetes", - "terraform", - "worker" - ], - "template": false, - "timeout_clone": 1800, - "timeout_create": 1800, - "timeout_migrate": 1800, - "timeout_move_disk": 1800, - "timeout_reboot": 1800, - "timeout_shutdown_vm": 1800, - "timeout_start_vm": 1800, - "timeout_stop_vm": 300, - "tpm_state": [], - "usb": [], - "vga": [], - "virtiofs": [], - "vm_id": 1013, - "watchdog": [] - }, - "sensitive_attributes": [], - "identity_schema_version": 0, - "private": "bnVsbA==", - "dependencies": [ - "data.ct_config.control_plane1_ignition", - "data.ct_config.control_plane2_ignition", - "data.ct_config.control_plane3_ignition", - "data.ct_config.worker3_ignition", - "null_resource.wait_for_cluster_ready", - "null_resource.wait_for_cp1_api", - "null_resource.wait_for_cp2_cp3_api", - "proxmox_virtual_environment_download_file.flatcar_image", - "proxmox_virtual_environment_file.control_plane1_ignition", - "proxmox_virtual_environment_file.control_plane2_ignition", - "proxmox_virtual_environment_file.control_plane3_ignition", - "proxmox_virtual_environment_file.worker3_ignition", - "proxmox_virtual_environment_vm.control_plane1", - "proxmox_virtual_environment_vm.control_plane2", - "proxmox_virtual_environment_vm.control_plane3", - "proxmox_virtual_environment_vm.flatcar_template" - ] - } - ] - }, - { - "mode": "managed", - "type": "proxmox_virtual_environment_vm", - "name": "worker4", - "provider": "provider[\"registry.terraform.io/bpg/proxmox\"]", - "instances": [ - { - "schema_version": 0, - "attributes": { - "acpi": true, - "agent": [], - "amd_sev": [], - "audio_device": [], - "bios": "seabios", - "boot_order": [ - "virtio0" - ], - "cdrom": [], - "clone": [ - { - "datastore_id": "", - "full": true, - "node_name": "hyper1", - "retries": 1, - "vm_id": 999 - } - ], - "cpu": [ - { - "affinity": "", - "architecture": "", - "cores": 2, - "flags": null, - "hotplugged": 0, - "limit": 0, - "numa": false, - "sockets": 1, - "type": "host", - "units": 1024 - } - ], - "description": "kubernetes worker4", - "disk": [], - "efi_disk": [], - "hook_script_file_id": null, - "hostpci": [], - "id": "1014", - "initialization": [ - { - "datastore_id": "Pool1", - "dns": [], - "interface": "ide2", - "ip_config": [], - "meta_data_file_id": "", - "network_data_file_id": "", - "type": "", - "user_account": [], - "user_data_file_id": "cephfs:snippets/worker4-ignition-user-data", - "vendor_data_file_id": "" - } - ], - "ipv4_addresses": [ - [ - "127.0.0.1" - ], - [ - "10.0.2.104" - ] - ], - "ipv6_addresses": [ - [ - "::1" - ], - [ - "2001:470:7116:2::104", - "fd00:0:0:2::104", - "fe80::be24:11ff:fe67:14d8" - ] - ], - "keyboard_layout": "en-us", - "kvm_arguments": null, - "mac_addresses": [ - "00:00:00:00:00:00", - "BC:24:11:67:14:D8" - ], - "machine": null, - "memory": [ - { - "dedicated": 12000, - "floating": 12000, - "hugepages": "", - "keep_hugepages": false, - "shared": 0 - } - ], - "migrate": false, - "name": "worker4", - "network_device": [ - { - "bridge": "vmbr0", - "disconnected": false, - "enabled": true, - "firewall": false, - "mac_address": "BC:24:11:67:14:D8", - "model": "virtio", - "mtu": 0, - "queues": 0, - "rate_limit": 0, - "trunks": "", - "vlan_id": 0 - } - ], - "network_interface_names": [ - "lo", - "eth0" - ], - "node_name": "pbs", - "numa": [], - "on_boot": true, - "operating_system": [], - "pool_id": null, - "protection": false, - "reboot": false, - "reboot_after_update": true, - "rng": [], - "scsi_hardware": "virtio-scsi-pci", - "serial_device": [], - "smbios": [], - "started": true, - "startup": [ - { - "down_delay": -1, - "order": 1, - "up_delay": -1 - } - ], - "stop_on_destroy": false, - "tablet_device": true, - "tags": [ - "flatcar", - "kubernetes", - "terraform", - "worker" - ], - "template": false, - "timeout_clone": 1800, - "timeout_create": 1800, - "timeout_migrate": 1800, - "timeout_move_disk": 1800, - "timeout_reboot": 1800, - "timeout_shutdown_vm": 1800, - "timeout_start_vm": 1800, - "timeout_stop_vm": 300, - "tpm_state": [], - "usb": [], - "vga": [], - "virtiofs": [], - "vm_id": 1014, - "watchdog": [] - }, - "sensitive_attributes": [], - "identity_schema_version": 0, - "private": "bnVsbA==", - "dependencies": [ - "data.ct_config.control_plane1_ignition", - "data.ct_config.control_plane2_ignition", - "data.ct_config.control_plane3_ignition", - "data.ct_config.worker4_ignition", - "null_resource.wait_for_cluster_ready", - "null_resource.wait_for_cp1_api", - "null_resource.wait_for_cp2_cp3_api", - "proxmox_virtual_environment_download_file.flatcar_image", - "proxmox_virtual_environment_file.control_plane1_ignition", - "proxmox_virtual_environment_file.control_plane2_ignition", - "proxmox_virtual_environment_file.control_plane3_ignition", - "proxmox_virtual_environment_file.worker4_ignition", - "proxmox_virtual_environment_vm.control_plane1", - "proxmox_virtual_environment_vm.control_plane2", - "proxmox_virtual_environment_vm.control_plane3", - "proxmox_virtual_environment_vm.flatcar_template" - ] - } - ] } ], "check_results": null diff --git a/terraform/terraform.tfstate.backup b/terraform/terraform.tfstate.backup index d7810ab..fa56db7 100644 --- a/terraform/terraform.tfstate.backup +++ b/terraform/terraform.tfstate.backup @@ -1,7 +1,7 @@ { "version": 4, "terraform_version": "1.12.2", - "serial": 1274, + "serial": 1350, "lineage": "d92c42be-29f9-bad9-ef9a-3dc952ff5fa5", "outputs": {}, "resources": [],