From 8743165707dfd005876c55756dd67d54e820e259 Mon Sep 17 00:00:00 2001 From: JJGadgets Date: Wed, 19 Jun 2024 13:58:47 +0800 Subject: [PATCH] fix(rook-ceph-cluster/biohazard): switch VLANs --- .../core/storage/rook-ceph/cluster/biohazard/hr.yaml | 9 +++++++-- 1 file changed, 7 insertions(+), 2 deletions(-) diff --git a/kube/deploy/core/storage/rook-ceph/cluster/biohazard/hr.yaml b/kube/deploy/core/storage/rook-ceph/cluster/biohazard/hr.yaml index e427ccf7..c1014d67 100644 --- a/kube/deploy/core/storage/rook-ceph/cluster/biohazard/hr.yaml +++ b/kube/deploy/core/storage/rook-ceph/cluster/biohazard/hr.yaml @@ -53,7 +53,9 @@ spec: network: provider: host addressRanges: - public: ["${IP_VLAN_CEPH_CIDR}"] + #public: ["${IP_VLAN_CEPH_CIDR}"] # TODO: maybe switch back to separate VLAN when 10GbE switch becomes managed? + #tagged VLANs don't work with this setup: Ceph nodes --> TL-ST1008F --> GLC-TE --> FortiSwitch 1GbE, but somehow `s/GLC-TE --> FortiSwitch/10G PVE OVS bridge with OPNsense VM/g` works lol + public: ["${IP_ROUTER_VLAN_K8S_CIDR}"] connections: requireMsgr2: true encryption: @@ -103,7 +105,9 @@ spec: tolerations: - key: "node-role.kubernetes.io/control-plane" operator: "Exists" - resources: + - key: "node-role.kubernetes.io/unschedulable" + operator: "Exists" + resources: # defaults are very aggressive mgr: requests: cpu: "150m" @@ -346,6 +350,7 @@ spec: gateway: labels: # netpols tailscale.com/expose: "true" + s3.home.arpa/rgw: "biohazard" instances: 2 hostNetwork: false port: 6953