Compare commits

...

10 Commits

10 changed files with 70 additions and 13061 deletions

View File

@ -0,0 +1,5 @@
apiVersion: kustomize.config.k8s.io/v1beta1
kind: Kustomization
resources:
- gotk-components.yaml
- gotk-sync.yaml

View File

@ -19,16 +19,16 @@ resources:
- ../../soft/pgadmin.yaml - ../../soft/pgadmin.yaml
- ../../soft/coroot.yaml - ../../soft/coroot.yaml
# - install/ingress-nginx.yaml - install/ingress-nginx.yaml
# - install/metrics-server.yaml - install/metrics-server.yaml
# - install/cert-manager.yaml - install/cert-manager.yaml
# - install/piraeus.yaml - install/piraeus.yaml
# - install/nfs-provisioner.yaml # - install/nfs-provisioner.yaml
# - install/monitoring.yaml - install/monitoring.yaml
# - install/loki.yaml # - install/loki.yaml
# - install/promtail.yaml # - install/promtail.yaml
# - install/kyverno.yaml # - install/kyverno.yaml
# - install/cpng.yaml - install/cpng.yaml
# - install/vswh.yaml # - install/vswh.yaml
# - install/stakater.yaml # - install/stakater.yaml
# - install/keycloak.yaml # - install/keycloak.yaml

View File

@ -2,6 +2,7 @@ apiVersion: postgresql.cnpg.io/v1
kind: Cluster kind: Cluster
metadata: metadata:
name: cluster-example-custom name: cluster-example-custom
namespace: default
spec: spec:
instances: 1 instances: 1
postgresql: postgresql:

View File

@ -4,7 +4,7 @@ metadata:
name: data-volume name: data-volume
namespace: default namespace: default
spec: spec:
storageClassName: piraeus-storage storageClassName: piraeus-storage-replicated
resources: resources:
requests: requests:
storage: 1Gi storage: 1Gi

View File

@ -66,20 +66,22 @@ talosctl --context talos-default kubeconfig ~/.kube/talos-default.yaml --nodes 1
## Общие команды ## Общие команды
```bash ```bash
talosctl --context talos-default cluster show CONTEXT=talos-demo
talosctl --context talos-default processes -w --nodes 10.5.0.2
talosctl --context talos-default --nodes 10.5.0.2 usage
talosctl --context talos-default --nodes 10.5.0.2 pcap
talosctl --context talos-default --nodes 10.5.0.2 netstat -nlpt
talosctl --context talos-default --nodes 10.5.0.2 edit machineconfig talosctl --context ${CONTEXT} cluster show
talosctl --context ${CONTEXT} processes -w --nodes 10.5.0.2
talosctl --context ${CONTEXT} --nodes 10.5.0.2 usage
talosctl --context ${CONTEXT} --nodes 10.5.0.2 pcap
talosctl --context ${CONTEXT} --nodes 10.5.0.2 netstat -nlpt
talosctl --context talos-default --nodes 10.5.0.2 get rd talosctl --context ${CONTEXT} --nodes 10.5.0.2 edit machineconfig
talosctl --context talos-default --nodes 10.5.0.2 get routes
talosctl --context talos-default --nodes 10.5.0.2 get mounts talosctl --context ${CONTEXT} --nodes 10.5.0.2 get rd
talosctl --context talos-default --nodes 10.5.0.2 get cpu talosctl --context ${CONTEXT} --nodes 10.5.0.2 get routes
talosctl --context talos-default --nodes 10.5.0.2 get systemdisk talosctl --context ${CONTEXT} --nodes 10.5.0.2 get mounts
talosctl --context talos-default --nodes 10.5.0.2 get etcdmember talosctl --context ${CONTEXT} --nodes 10.5.0.2 get cpu
talosctl --context ${CONTEXT} --nodes 10.5.0.2 get systemdisk
talosctl --context ${CONTEXT} --nodes 10.5.0.2 get etcdmember
``` ```
## Удаление ## Удаление

View File

@ -4,4 +4,4 @@ metadata:
name: pool name: pool
spec: spec:
blocks: blocks:
- cidr: 192.168.23.22/32 - cidr: 192.168.3.201/32

View File

@ -23,7 +23,7 @@ cgroup:
autoMount: autoMount:
enabled: false enabled: false
hostRoot: /sys/fs/cgroup hostRoot: /sys/fs/cgroup
k8sServiceHost: 192.168.23.21 k8sServiceHost: 192.168.3.200
k8sServicePort: 6443 k8sServicePort: 6443
l2announcements: l2announcements:
enabled: true enabled: true

33
talos-k8s/cluster.yaml Normal file
View File

@ -0,0 +1,33 @@
clusterName: talos-demo
k8sVersion: 1.33.2
image: factory.talos.dev/metal-installer/6adc7e7fba27948460e2231e5272e88b85159da3f3db980551976bf9898ff64b:v1.10.4
iface: ens18
cpCount: 3
workerCount: 2
gateway: 192.168.3.1
netmask: "24"
dns1: 192.168.3.1
dns2: 192.168.3.1
ntp1: 1.ru.pool.ntp.org
ntp2: 2.ru.pool.ntp.org
ntp3: 3.ru.pool.ntp.org
useVIP: true
vipIP: 192.168.3.200
useExtBalancer: false
extBalancerIP: ""
disk: /dev/sda
useDRBD: true
useZFS: false
useSPL: false
useVFIOPCI: false
useVFIOIOMMU: false
useOVS: false
useMirrors: true
useMaxPods: false
cpIPs:
- 192.168.3.53
- 192.168.3.54
- 192.168.3.55
workerIPs:
- 192.168.3.58
- 192.168.3.59

View File

@ -2,10 +2,10 @@
VERB=${1:-"create"} VERB=${1:-"create"}
nodeip=pve1-nsk.bildme.ru nodeip=pve1.p8o.ru
nodeid=pve1-nsk nodeid=pve1
poolid=test-vm poolid=test
vmid_template=777 vmid_template=9995
vmid=80 vmid=80
vm_name=t8s-demo vm_name=t8s-demo
@ -37,16 +37,16 @@ echo "${VERB} stand"
[[ $VERB == "create" ]] && { [[ $VERB == "create" ]] && {
for i in $(seq 1 "$controllers") for i in $(seq 1 "$controllers")
do do
ssh root@${nodeip} "pvesh create /nodes/${nodeid}/qemu/${vmid_template}/clone --newid ${vmid}${i} --full false --name ${vm_name}-cp-${i}" ssh root@${nodeip} "pvesh create /nodes/${nodeid}/qemu/${vmid_template}/clone --newid ${vmid}${i} --full false --name ${vm_name}-cp-${i}.p8o.ru"
ssh root@${nodeip} "pvesh set /nodes/${nodeid}/qemu/${vmid}${i}/config --cores 2 --vcpus 2 --memory 6144" ssh root@${nodeip} "pvesh set /nodes/${nodeid}/qemu/${vmid}${i}/config --cores 4 --vcpus 4 --memory 6144"
ssh root@${nodeip} "pvesh set /pools/${poolid} --vms "${vmid}${i}"" ssh root@${nodeip} "pvesh set /pools/${poolid} --vms "${vmid}${i}""
ssh root@${nodeip} "pvesh create /nodes/${nodeid}/qemu/${vmid}${i}/status/start" ssh root@${nodeip} "pvesh create /nodes/${nodeid}/qemu/${vmid}${i}/status/start"
done done
for i in $(seq 1 "$workers") for i in $(seq 1 "$workers")
do do
ssh root@${nodeip} "pvesh create /nodes/${nodeid}/qemu/${vmid_template}/clone --newid ${vmid}$((i + 4)) --full false --name ${vm_name}-w-${i}" ssh root@${nodeip} "pvesh create /nodes/${nodeid}/qemu/${vmid_template}/clone --newid ${vmid}$((i + 4)) --full false --name ${vm_name}-w-${i}.p8o.ru"
ssh root@${nodeip} "pvesh set /nodes/${nodeid}/qemu/${vmid}$((i + 4))/config --cores 2 --vcpus 4 --memory 12288" ssh root@${nodeip} "pvesh set /nodes/${nodeid}/qemu/${vmid}$((i + 4))/config --cores 4 --vcpus 4 --memory 12288"
ssh root@${nodeip} "pvesh set /pools/${poolid} --vms "${vmid}$((i + 4))"" ssh root@${nodeip} "pvesh set /pools/${poolid} --vms "${vmid}$((i + 4))""
ssh root@${nodeip} "pvesh create /nodes/${nodeid}/qemu/${vmid}$((i + 4))/status/start" ssh root@${nodeip} "pvesh create /nodes/${nodeid}/qemu/${vmid}$((i + 4))/status/start"
done done