Compare commits
70 Commits
benvin/exp
...
develop
| Author | SHA1 | Date | |
|---|---|---|---|
| 0377c40a07 | |||
| 8bb40dadce | |||
| bc769aa1df | |||
| 4e652ccbe6 | |||
| 8c24c6582f | |||
| 6bfc63ca31 | |||
| 69dc9e8f66 | |||
| c4d28d52bc | |||
| 6219855fb1 | |||
| 7215a6f534 | |||
| 88efdbcdd3 | |||
| 3c114371e0 | |||
| 1077bdcbc1 | |||
| 4e928585f5 | |||
| dbe1398218 | |||
| 9f5b1cec82 | |||
| 383bbb0507 | |||
| 6f51bffeaa | |||
| 57870658b5 | |||
| f8caa71f34 | |||
| a2c56c9e46 | |||
| 40d8e924ee | |||
| 0aec795aec | |||
| 9854403b02 | |||
| 6400c89853 | |||
| 9eff241003 | |||
| 35614060bd | |||
| 1b0fd10fd7 | |||
| 2c9fb3d86a | |||
| 559c453906 | |||
| 5b0365c096 | |||
| 1e7dfb9d9d | |||
| 9dd74013ea | |||
| 92a48b4113 | |||
| 78adef0eee | |||
| 81f289a185 | |||
| a2a8edb731 | |||
| e129d1cf7a | |||
| e95a59b88a | |||
| 8bed80eac8 | |||
| 5ba483c68a | |||
| 766233c3e5 | |||
| 98b866fce7 | |||
| e724326d43 | |||
| d8b354558d | |||
| fac90c66db | |||
| efbbb6bcb1 | |||
| 16e654fdd7 | |||
| 66d8815e16 | |||
| a9c959d924 | |||
| b224cfb516 | |||
| 4c9204858e | |||
| 571a9b25a7 | |||
| 762f415d2d | |||
| 4e77fb7ee7 | |||
| 6e4bc9fbc7 | |||
| 012e842d7d | |||
| 98a433d366 | |||
| fcd1b049d6 | |||
| 938a6ac990 | |||
| 0665873dc8 | |||
| ae4eb3a5eb | |||
| 65fb52da55 | |||
| d97cbfd570 | |||
| 8f5d102945 | |||
| 62aade77ff | |||
| 83bb3e1085 | |||
| 92728047e7 | |||
| f4af5e7b64 | |||
| 308d97d783 |
@ -1,24 +0,0 @@
|
||||
name: Build
|
||||
|
||||
on:
|
||||
pull_request:
|
||||
|
||||
jobs:
|
||||
precommit:
|
||||
runs-on: almalinux-8
|
||||
container:
|
||||
image: git.unkin.net/unkin/almalinux9-actionsdind:latest
|
||||
options: --privileged
|
||||
|
||||
steps:
|
||||
- name: Checkout code
|
||||
uses: actions/checkout@v3
|
||||
|
||||
- name: Install requirements
|
||||
run: |
|
||||
dnf groupinstall -y "Development Tools" -y
|
||||
dnf install rubygems ruby-devel gcc make redhat-rpm-config glibc-headers glibc-devel -y
|
||||
|
||||
- name: Pre-Commit All Files
|
||||
run: |
|
||||
uvx pre-commit run --all-files
|
||||
10
.woodpecker/pre-commit.yaml
Normal file
10
.woodpecker/pre-commit.yaml
Normal file
@ -0,0 +1,10 @@
|
||||
when:
|
||||
- event: pull_request
|
||||
|
||||
steps:
|
||||
- name: pre-commit
|
||||
image: git.unkin.net/unkin/almalinux9-base:latest
|
||||
commands:
|
||||
- dnf groupinstall -y "Development Tools" -y
|
||||
- dnf install uv rubygems ruby-devel gcc make redhat-rpm-config glibc-headers glibc-devel libffi libffi-devel -y
|
||||
- uvx pre-commit run --all-files
|
||||
@ -19,6 +19,7 @@ mod 'puppetlabs-haproxy', '8.2.0'
|
||||
mod 'puppetlabs-java', '11.1.0'
|
||||
mod 'puppetlabs-reboot', '5.1.0'
|
||||
mod 'puppetlabs-docker', '10.2.0'
|
||||
mod 'puppetlabs-mailalias_core', '1.2.0'
|
||||
|
||||
# puppet
|
||||
mod 'puppet-python', '7.4.0'
|
||||
@ -43,6 +44,8 @@ mod 'puppet-letsencrypt', '11.1.0'
|
||||
mod 'puppet-rundeck', '9.2.0'
|
||||
mod 'puppet-redis', '11.1.0'
|
||||
mod 'puppet-nodejs', '11.0.0'
|
||||
mod 'puppet-postfix', '5.1.0'
|
||||
mod 'puppet-alternatives', '6.0.0'
|
||||
|
||||
# other
|
||||
mod 'saz-sudo', '9.0.2'
|
||||
@ -60,6 +63,7 @@ mod 'rehan-mkdir', '2.0.0'
|
||||
mod 'tailoredautomation-patroni', '2.0.0'
|
||||
mod 'ssm-crypto_policies', '0.3.3'
|
||||
mod 'thias-sysctl', '1.0.8'
|
||||
mod 'cirrax-dovecot', '1.3.3'
|
||||
|
||||
mod 'bind',
|
||||
:git => 'https://git.service.au-syd1.consul/unkinben/puppet-bind.git',
|
||||
|
||||
@ -28,6 +28,98 @@ Always refer back to the official documentation at https://docs.ceph.com/en/late
|
||||
sudo ceph fs set mediafs max_mds 2
|
||||
```
|
||||
|
||||
|
||||
## managing cephfs with subvolumes
|
||||
|
||||
Create erasure code profiles. The K and M values are equivalent to the number of data disks (K) and parity disks (M) in RAID5, RAID6, etc.
|
||||
|
||||
sudo ceph osd erasure-code-profile set ec_6_2 k=6 m=2
|
||||
sudo ceph osd erasure-code-profile set ec_4_1 k=4 m=1
|
||||
|
||||
Create data pools using the erasure-code-profile, set some required options
|
||||
|
||||
sudo ceph osd pool create cephfs_data_ssd_ec_6_2 erasure ec_6_2
|
||||
sudo ceph osd pool set cephfs_data_ssd_ec_6_2 allow_ec_overwrites true
|
||||
sudo ceph osd pool set cephfs_data_ssd_ec_6_2 bulk true
|
||||
|
||||
sudo ceph osd pool create cephfs_data_ssd_ec_4_1 erasure ec_4_1
|
||||
sudo ceph osd pool set cephfs_data_ssd_ec_4_1 allow_ec_overwrites true
|
||||
sudo ceph osd pool set cephfs_data_ssd_ec_4_1 bulk true
|
||||
|
||||
Add the pool to the fs `cephfs`
|
||||
|
||||
sudo ceph fs add_data_pool cephfs cephfs_data_ssd_ec_6_2
|
||||
sudo ceph fs add_data_pool cephfs cephfs_data_ssd_ec_4_1
|
||||
|
||||
Create a subvolumegroup using the new data pool
|
||||
|
||||
sudo ceph fs subvolumegroup create cephfs csi_ssd_ec_6_2 --pool_layout cephfs_data_ssd_ec_6_2
|
||||
sudo ceph fs subvolumegroup create cephfs csi_ssd_ec_4_1 --pool_layout cephfs_data_ssd_ec_4_1
|
||||
|
||||
All together:
|
||||
|
||||
sudo ceph osd erasure-code-profile set ec_6_2 k=6 m=2
|
||||
sudo ceph osd pool create cephfs_data_ssd_ec_6_2 erasure ec_6_2
|
||||
sudo ceph osd pool set cephfs_data_ssd_ec_6_2 allow_ec_overwrites true
|
||||
sudo ceph osd pool set cephfs_data_ssd_ec_6_2 bulk true
|
||||
sudo ceph fs add_data_pool cephfs cephfs_data_ssd_ec_6_2
|
||||
sudo ceph fs subvolumegroup create cephfs csi_ssd_ec_6_2 --pool_layout cephfs_data_ssd_ec_6_2
|
||||
|
||||
sudo ceph osd erasure-code-profile set ec_4_1 k=4 m=1
|
||||
sudo ceph osd pool create cephfs_data_ssd_ec_4_1 erasure ec_4_1
|
||||
sudo ceph osd pool set cephfs_data_ssd_ec_4_1 allow_ec_overwrites true
|
||||
sudo ceph osd pool set cephfs_data_ssd_ec_4_1 bulk true
|
||||
sudo ceph fs add_data_pool cephfs cephfs_data_ssd_ec_4_1
|
||||
sudo ceph fs subvolumegroup create cephfs csi_ssd_ec_4_1 --pool_layout cephfs_data_ssd_ec_4_1
|
||||
|
||||
Create a key with access to the new subvolume groups. Check if the user already exists first:
|
||||
|
||||
sudo ceph auth get client.kubernetes-cephfs
|
||||
|
||||
If it doesnt:
|
||||
|
||||
sudo ceph auth get-or-create client.kubernetes-cephfs \
|
||||
mgr 'allow rw' \
|
||||
osd 'allow rw tag cephfs metadata=cephfs, allow rw tag cephfs data=cephfs' \
|
||||
mds 'allow r fsname=cephfs path=/volumes, allow rws fsname=cephfs path=/volumes/csi_ssd_ec_6_2, allow rws fsname=cephfs path=/volumes/csi_ssd_ec_4_1' \
|
||||
mon 'allow r fsname=cephfs'
|
||||
|
||||
If it does, use `sudo ceph auth caps client.kubernetes-cephfs ...` instead to update existing capabilities.
|
||||
|
||||
## removing a cephfs subvolumegroup from cephfs
|
||||
|
||||
This will cleanup the subvolumegroup, and subvolumes if they exist, then remove the pool.
|
||||
|
||||
Check for subvolumegroups first, then for subvolumes in it
|
||||
|
||||
sudo ceph fs subvolumegroup ls cephfs
|
||||
sudo ceph fs subvolume ls cephfs --group_name csi_raid6
|
||||
|
||||
|
||||
If subvolumes exist, remove each one-by-one:
|
||||
|
||||
sudo ceph fs subvolume rm cephfs <subvol_name> --group_name csi_raid6
|
||||
|
||||
If you have snapshots, remove snapshots first:
|
||||
|
||||
sudo ceph fs subvolume snapshot ls cephfs <subvol_name> --group_name csi_raid6
|
||||
sudo ceph fs subvolume snapshot rm cephfs <subvol_name> <snap_name> --group_name csi_raid6
|
||||
|
||||
Once the group is empty, remove it:
|
||||
|
||||
sudo ceph fs subvolumegroup rm cephfs csi_raid6
|
||||
|
||||
If it complains it’s not empty, go back as there’s still a subvolume or snapshot.
|
||||
|
||||
If you added it with `ceph fs add_data_pool`. Undo with `rm_data_pool`:
|
||||
|
||||
sudo ceph fs rm_data_pool cephfs cephfs_data_csi_raid6
|
||||
|
||||
After it’s detached from CephFS, you can delete it.
|
||||
|
||||
sudo ceph osd pool rm cephfs_data_csi_raid6 cephfs_data_csi_raid6 --yes-i-really-really-mean-it
|
||||
|
||||
|
||||
## creating authentication tokens
|
||||
|
||||
- this will create a client keyring named media
|
||||
@ -58,3 +150,78 @@ this will overwrite the current capabilities of a given client.user
|
||||
mon 'allow r' \
|
||||
mds 'allow rw path=/' \
|
||||
osd 'allow rw pool=media_data'
|
||||
|
||||
## adding a new osd on new node
|
||||
|
||||
create the ceph conf (automate this?)
|
||||
|
||||
cat <<EOF | sudo tee /etc/ceph/ceph.conf
|
||||
[global]
|
||||
auth_client_required = cephx
|
||||
auth_cluster_required = cephx
|
||||
auth_service_required = cephx
|
||||
fsid = de96a98f-3d23-465a-a899-86d3d67edab8
|
||||
mon_allow_pool_delete = true
|
||||
mon_initial_members = prodnxsr0009,prodnxsr0010,prodnxsr0011,prodnxsr0012,prodnxsr0013
|
||||
mon_host = 198.18.23.9,198.18.23.10,198.18.23.11,198.18.23.12,198.18.23.13
|
||||
ms_bind_ipv4 = true
|
||||
ms_bind_ipv6 = false
|
||||
osd_crush_chooseleaf_type = 1
|
||||
osd_pool_default_min_size = 2
|
||||
osd_pool_default_size = 3
|
||||
osd_pool_default_pg_num = 128
|
||||
public_network = 198.18.23.1/32,198.18.23.2/32,198.18.23.3/32,198.18.23.4/32,198.18.23.5/32,198.18.23.6/32,198.18.23.7/32,198.18.23.8/32,198.18.23.9/32,198.18.23.10/32,198.18.23.11/32,198.18.23.12/32,198.18.23.13/32
|
||||
EOF
|
||||
|
||||
ssh to one of the monitor hosts, then transfer the keys required
|
||||
|
||||
sudo cat /etc/ceph/ceph.client.admin.keyring | ssh prodnxsr0003 'sudo tee /etc/ceph/ceph.client.admin.keyring'
|
||||
sudo cat /var/lib/ceph/bootstrap-osd/ceph.keyring | ssh prodnxsr0003 'sudo tee /var/lib/ceph/bootstrap-osd/ceph.keyring'
|
||||
|
||||
assuming we are adding /dev/sda to the cluster, first zap the disk to remove partitions/lvm/metadata
|
||||
|
||||
sudo ceph-volume lvm zap /dev/sda --destroy
|
||||
|
||||
then add it to the cluster
|
||||
|
||||
sudo ceph-volume lvm create --data /dev/sda
|
||||
|
||||
## removing an osd
|
||||
|
||||
check what OSD IDs were on this host (if you know it)
|
||||
|
||||
sudo ceph osd tree
|
||||
|
||||
or check for any DOWN osds
|
||||
|
||||
sudo ceph osd stat
|
||||
sudo ceph health detail
|
||||
|
||||
once you identify the old OSD ID, remove it with these steps, replace X with the actual OSD ID:
|
||||
|
||||
sudo ceph osd out osd.X
|
||||
sudo ceph osd down osd.X
|
||||
sudo ceph osd crush remove osd.X
|
||||
sudo ceph auth del osd.X
|
||||
sudo ceph osd rm osd.X
|
||||
|
||||
|
||||
## maintenance mode for the cluster
|
||||
|
||||
from one node in the cluster disable recovery
|
||||
|
||||
sudo ceph osd set noout
|
||||
sudo ceph osd set nobackfill
|
||||
sudo ceph osd set norecover
|
||||
sudo ceph osd set norebalance
|
||||
sudo ceph osd set nodown
|
||||
sudo ceph osd set pause
|
||||
|
||||
to undo the change, use unset
|
||||
|
||||
sudo ceph osd unset noout
|
||||
sudo ceph osd unset nobackfill
|
||||
sudo ceph osd unset norecover
|
||||
sudo ceph osd unset norebalance
|
||||
sudo ceph osd unset nodown
|
||||
sudo ceph osd unset pause
|
||||
|
||||
@ -29,3 +29,21 @@ these steps are required when adding additional puppet masters, as the subject a
|
||||
|
||||
sudo systemctl start puppetserver
|
||||
sudo cp /root/current_crl.pem /etc/puppetlabs/puppet/ssl/crl.pem
|
||||
|
||||
|
||||
## troubleshooting
|
||||
|
||||
### Issue 1:
|
||||
|
||||
[sysadmin@ausyd1nxvm2056 ~]$ sudo puppet agent -t
|
||||
Error: The CRL issued by 'CN=Puppet CA: prodinf01n01.main.unkin.net' is missing
|
||||
|
||||
Find another puppetserver that IS working, copy the `/etc/puppetlabs/puppet/ssl/crl.pem` to this host, run puppet again.
|
||||
|
||||
|
||||
### Issue 2:
|
||||
|
||||
[sysadmin@ausyd1nxvm2097 ~]$ sudo puppet agent -t
|
||||
Error: Failed to parse CA certificates as PEM
|
||||
|
||||
The puppet-agents CA cert `/etc/puppetlabs/puppet/ssl/certs/ca.pem` is empty or missing. Grab it from any other host. Run puppet again.
|
||||
|
||||
@ -1,6 +1,6 @@
|
||||
---
|
||||
profiles::accounts::sysadmin::password: ENC[PKCS7,MIIBqQYJKoZIhvcNAQcDoIIBmjCCAZYCAQAxggEhMIIBHQIBADAFMAACAQEwDQYJKoZIhvcNAQEBBQAEggEAoS7GyofFaXBNTWU+GtSiz4eCX/9j/sh3fDDRgOgNv1qpcQ87ZlTTenbHo9lxeURxKQ2HVVt7IsrBo/SC/WgipAKnliRkkIvo7nfAs+i+kEE8wakjAs0DcB4mhqtIZRuBkLG2Nay//DcG6cltVkbKEEKmKLMkDFZgTWreOZal8nDljpVe1S8QwtwP4/6hKTef5xsOnrisxuffWTXvwYJhj/VXrjdoH7EhtHGLybzEalglkVHEGft/WrrD/0bwJpmR0RegWI4HTsSvGiHgvf5DZJx8fXPZNPnicGtlfA9ccQPuVo17bY4Qf/WIc1A8Ssv4kHSbNIYJKRymI3UFb0Z4wzBsBgkqhkiG9w0BBwEwHQYJYIZIAWUDBAEqBBBBxDLb6pCGbittkcX6asd/gEBmMcUNupDjSECq5H09YA70eVwWWe0fBqxTxrr2cXCXtRKFvOk8SJmL0xHAWodaLN9+krTWHJcWbAK8JXEPC7rn]
|
||||
profiles::accounts::root::password: ENC[PKCS7,MIIBeQYJKoZIhvcNAQcDoIIBajCCAWYCAQAxggEhMIIBHQIBADAFMAACAQEwDQYJKoZIhvcNAQEBBQAEggEAM79PRxeAZHrDcSm4eSFqU94/LjuSbdUmJWivX/Pa8GumoW2e/PT9nGHW3p98zHthMgCglk52PECQ+TBKjxr+9dTyNK5ePG6ZJEqSHNRqsPGm+kfQj/hlTmq8vOBaFM5GapD1iTHs5JFbGngI56swKBEVXW9+Z37BjQb2xJuyLsu5Bo/tA0BaOKuCtjq1a6E38bOX+nJ+YF1uZgV9ofAEh1YvkcTmnEWYXFRPWd7AaNcWn03V2pfhGqxc+xydak620I47P+FE+qIY72+aQ6tmLU3X9vyA1HLF2Tv572l4a2i+YIk6nAgQdi+hQKznqNL9M9YV+s1AcmcKLT7cfLrjsjA8BgkqhkiG9w0BBwEwHQYJYIZIAWUDBAEqBBCMWrdCWBQgtW3NOEpERwP+gBA3KDiqe4pQq6DwRfsEXQNZ]
|
||||
profiles::accounts::root::password: ENC[PKCS7,MIIB2gYJKoZIhvcNAQcDoIIByzCCAccCAQAxggEhMIIBHQIBADAFMAACAQEwDQYJKoZIhvcNAQEBBQAEggEAIgzGQLoHrm7JSnWG4vdAtxSuETmnqbV7kUsQS8WCRUwFGenDFkps+OGMnOGEHLzMJzXihLfgfdWwTAI4fp48M+zhTMo9TQkdzZqtbFk3+RjV2jDF0wfe4kVUIpReOq+EkaDSkoRSG8V6hWvszhDHUrJBC9eDhomL0f3xNAWxmy5EIX/uMEvg9Ux5YX+E6k2pEIKnHNoVIaWDojlofSIzIqTSS7l3jQtJhs3YqBzLL1DsoF1kdn+Rwl5kcsKkhV+vzl76wEbpYVZW8lu4bFfP6QHMLPcep2tuUDMCDvARRXD7YyZcAtS7aMuqll+BLAszpWxAA7EU2hgvdr6t2uyVCTCBnAYJKoZIhvcNAQcBMB0GCWCGSAFlAwQBKgQQ4D5oDoyE6LPdjpVtGPoJD4BwfnQ9ORjYFPvHQmt+lgU4jMqh6BhqP0VN3lqVfUpOmiVMIqkO/cYtlwVLKEg36TPCHBSpqvhuahSF5saCVr8JY3xWOAmTSgnNjQOPlGrPnYWYbuRLxVRsU+KUkpAzR0c6VN0wYi6bI85Pcv8yHF3UYA==]
|
||||
profiles::consul::client::secret_id_salt: ENC[PKCS7,MIIBmQYJKoZIhvcNAQcDoIIBijCCAYYCAQAxggEhMIIBHQIBADAFMAACAQEwDQYJKoZIhvcNAQEBBQAEggEAS7pNFRX4onccFaR87zB/eFFORuF22j6xqjyeAqqjgEduYhkt6w5kkz+YfUoHUesU0Q6F2p6HrCSZ8yAsx5M25NCiud9P4hIpjKmOZ4zCNO7uhATh4AQDYw3BrdRwfO+c6jOl5wOiNLCfDBJ0sFT3akCvcuPS1xIoRJq4Gyn+uCbOsMbvSl25ld2xKt1/cqs8gc1d8mkpjwWto7t+qZSUFMCehTbehH3G4a3Q5rvfBoNwv42Wbs676BDcCurDaAzHNqE7pDbOWhGuVOBl+q+BU0Ri/CRkGcTViN9fr8Dc9SveVC6EPsMbw+05/8/NlfzQse3KAwQ34nR9tR2PQw5qEzBcBgkqhkiG9w0BBwEwHQYJYIZIAWUDBAEqBBB7LywscQtF7cG2nomfEsu9gDDVqJBFP1jAX2eGZ2crYS5gnBcsRwhc0HNo2/WWdhZprMW+vEJOOGXDelI53NxA3o0=]
|
||||
profiles::consul::token::node_editor::secret_id: ENC[PKCS7,MIIBmQYJKoZIhvcNAQcDoIIBijCCAYYCAQAxggEhMIIBHQIBADAFMAACAQEwDQYJKoZIhvcNAQEBBQAEggEAO8IIF2r18dFf0bVKEwjJUe1TmXHH0AIzsQHxkHwV7d37kvH1cY9rYw0TtdHn7GTxvotJG7GZbWvbunpBs1g2p2RPADiM6TMhbO8mJ0tAWLnMk7bQ221xu8Pc7KceqWmU17dmgNhVCohyfwJNqbA756TlHVgxGA0LtNrKoLOmgKGXAL1VYZoKEQnWq7xOpO+z3e1UfjoO6CvX/Od2hGYfUkHdro8mwRw4GFKzU7XeKFdAMUGpn5rVmY3xe+1ARXwGFaSrTHzk2n85pvwhPRlQ+OwqzyT19Qo2FNeAO6RoCRIFTtqbsjTWPUlseHIhw4Q5bHO1I0Mrlm5IHDESw/22IzBcBgkqhkiG9w0BBwEwHQYJYIZIAWUDBAEqBBCEe9wD72qxnpeq5nCi/d7BgDCP29sDFObkFTabt2uZ/nF9MT1g+QOrrdFKgnG6ThnwH1hwpZPsSVgIs+yRQH8laB4=]
|
||||
profiles::consul::server::acl_tokens_initial_management: ENC[PKCS7,MIIBmQYJKoZIhvcNAQcDoIIBijCCAYYCAQAxggEhMIIBHQIBADAFMAACAQEwDQYJKoZIhvcNAQEBBQAEggEAi1UH7AZirJ1PdxWy+KEgS5ufm0wbn2xy9rkg14hKYpcVjBa4pOZpSLMGMiiUpBIqBytDMZM4ezYa/luktpkBImJbM/TE16beGtsacQGA+9eZk2Tihs9GR2qbAQiu5lLITiDlwNnf0GeWdqHM8CTeD68DczQF320d9U14/k6pG/7z+w/MGLcjsQoSuOFTm42JVn1BI46t1CYSCHMXQc/9Tfs+FzI+vumohI8DxAYBIuyzU5HBX/MntAsvD/yixMJS1pZL9WwgqZJC/wK34rVRB39DpxWf/WROrI+WLuSJwr7WBjaeF9Ju+89WKCgsI53EWhFTj8GgDZm/jqPoE478NjBcBgkqhkiG9w0BBwEwHQYJYIZIAWUDBAEqBBAoACRzJdQKNYXZv6cghFIIgDAzB81DMcuY815nb8POtZpiA06jT/068AoZmSctHoFK/zW9tY229N5r1Tb+WHElqLk=]
|
||||
|
||||
@ -129,6 +129,9 @@ lookup_options:
|
||||
profiles::ceph::client::keyrings:
|
||||
merge:
|
||||
strategy: deep
|
||||
profiles::ceph::conf::config:
|
||||
merge:
|
||||
strategy: deep
|
||||
profiles::nginx::simpleproxy::locations:
|
||||
merge:
|
||||
strategy: deep
|
||||
@ -155,6 +158,24 @@ lookup_options:
|
||||
zfs::datasets:
|
||||
merge:
|
||||
strategy: deep
|
||||
rke2::config_hash:
|
||||
merge:
|
||||
strategy: deep
|
||||
postfix::configs:
|
||||
merge:
|
||||
strategy: deep
|
||||
postfix::maps:
|
||||
merge:
|
||||
strategy: deep
|
||||
postfix::virtuals:
|
||||
merge:
|
||||
strategy: deep
|
||||
stalwart::postgresql_password:
|
||||
convert_to: Sensitive
|
||||
stalwart::s3_secret_key:
|
||||
convert_to: Sensitive
|
||||
stalwart::fallback_admin_password:
|
||||
convert_to: Sensitive
|
||||
|
||||
facts_path: '/opt/puppetlabs/facter/facts.d'
|
||||
|
||||
@ -173,9 +194,6 @@ profiles::ntp::client::peers:
|
||||
- 2.au.pool.ntp.org
|
||||
- 3.au.pool.ntp.org
|
||||
|
||||
profiles::base::puppet_servers:
|
||||
- 'prodinf01n01.main.unkin.net'
|
||||
|
||||
consul::install_method: 'package'
|
||||
consul::manage_repo: false
|
||||
consul::bin_dir: /usr/bin
|
||||
@ -368,6 +386,71 @@ profiles::ceph::client::mons:
|
||||
- 10.18.15.1
|
||||
- 10.18.15.2
|
||||
- 10.18.15.3
|
||||
|
||||
profiles::ceph::conf::config:
|
||||
global:
|
||||
auth_client_required: 'cephx'
|
||||
auth_cluster_required: 'cephx'
|
||||
auth_service_required: 'cephx'
|
||||
fsid: 'de96a98f-3d23-465a-a899-86d3d67edab8'
|
||||
mon_allow_pool_delete: true
|
||||
mon_initial_members: 'prodnxsr0009,prodnxsr0010,prodnxsr0011,prodnxsr0012,prodnxsr0013'
|
||||
mon_host: '198.18.23.9,198.18.23.10,198.18.23.11,198.18.23.12,198.18.23.13'
|
||||
ms_bind_ipv4: true
|
||||
ms_bind_ipv6: false
|
||||
osd_crush_chooseleaf_type: 1
|
||||
osd_pool_default_min_size: 2
|
||||
osd_pool_default_size: 3
|
||||
osd_pool_default_pg_num: 128
|
||||
public_network: >
|
||||
198.18.23.1/32,198.18.23.2/32,198.18.23.3/32,198.18.23.4/32,
|
||||
198.18.23.5/32,198.18.23.6/32,198.18.23.7/32,198.18.23.8/32,
|
||||
198.18.23.9/32,198.18.23.10/32,198.18.23.11/32,198.18.23.12/32,
|
||||
198.18.23.13/32
|
||||
client.rgw.ausyd1nxvm2115:
|
||||
rgw_realm: unkin
|
||||
rgw_zonegroup: au
|
||||
rgw_zone: syd1
|
||||
client.rgw.ausyd1nxvm2116:
|
||||
rgw_realm: unkin
|
||||
rgw_zonegroup: au
|
||||
rgw_zone: syd1
|
||||
client.rgw.ausyd1nxvm2117:
|
||||
rgw_realm: unkin
|
||||
rgw_zonegroup: au
|
||||
rgw_zone: syd1
|
||||
client.rgw.ausyd1nxvm2118:
|
||||
rgw_realm: unkin
|
||||
rgw_zonegroup: au
|
||||
rgw_zone: syd1
|
||||
client.rgw.ausyd1nxvm2119:
|
||||
rgw_realm: unkin
|
||||
rgw_zonegroup: au
|
||||
rgw_zone: syd1
|
||||
mds:
|
||||
keyring: /var/lib/ceph/mds/ceph-$id/keyring
|
||||
mds_standby_replay: true
|
||||
mds.prodnxsr0009-1:
|
||||
host: prodnxsr0009
|
||||
mds.prodnxsr0009-2:
|
||||
host: prodnxsr0009
|
||||
mds.prodnxsr0010-1:
|
||||
host: prodnxsr0010
|
||||
mds.prodnxsr0010-2:
|
||||
host: prodnxsr0010
|
||||
mds.prodnxsr0011-1:
|
||||
host: prodnxsr0011
|
||||
mds.prodnxsr0011-2:
|
||||
host: prodnxsr0011
|
||||
mds.prodnxsr0012-1:
|
||||
host: prodnxsr0012
|
||||
mds.prodnxsr0012-2:
|
||||
host: prodnxsr0012
|
||||
mds.prodnxsr0013-1:
|
||||
host: prodnxsr0013
|
||||
mds.prodnxsr0013-2:
|
||||
host: prodnxsr0013
|
||||
|
||||
#profiles::base::hosts::additional_hosts:
|
||||
# - ip: 198.18.17.9
|
||||
# hostname: prodinf01n09.main.unkin.net
|
||||
|
||||
@ -5,3 +5,5 @@ profiles_dns_upstream_forwarder_unkin:
|
||||
- 198.18.19.15
|
||||
profiles_dns_upstream_forwarder_consul:
|
||||
- 198.18.19.14
|
||||
profiles_dns_upstream_forwarder_k8s:
|
||||
- 198.18.19.20
|
||||
|
||||
@ -10,6 +10,12 @@ profiles::haproxy::dns::vrrp_cnames:
|
||||
- git.unkin.net
|
||||
- fafflix.unkin.net
|
||||
- grafana.unkin.net
|
||||
- dashboard.ceph.unkin.net
|
||||
- mail-webadmin.main.unkin.net
|
||||
- mail-in.main.unkin.net
|
||||
- mail.main.unkin.net
|
||||
- autoconfig.main.unkin.net
|
||||
- autodiscover.main.unkin.net
|
||||
|
||||
profiles::haproxy::mappings:
|
||||
fe_http:
|
||||
@ -27,6 +33,10 @@ profiles::haproxy::mappings:
|
||||
- 'fafflix.unkin.net be_jellyfin'
|
||||
- 'git.unkin.net be_gitea'
|
||||
- 'grafana.unkin.net be_grafana'
|
||||
- 'dashboard.ceph.unkin.net be_ceph_dashboard'
|
||||
- 'mail-webadmin.main.unkin.net be_stalwart_webadmin'
|
||||
- 'autoconfig.main.unkin.net be_stalwart_webadmin'
|
||||
- 'autodiscovery.main.unkin.net be_stalwart_webadmin'
|
||||
fe_https:
|
||||
ensure: present
|
||||
mappings:
|
||||
@ -42,6 +52,10 @@ profiles::haproxy::mappings:
|
||||
- 'fafflix.unkin.net be_jellyfin'
|
||||
- 'git.unkin.net be_gitea'
|
||||
- 'grafana.unkin.net be_grafana'
|
||||
- 'dashboard.ceph.unkin.net be_ceph_dashboard'
|
||||
- 'mail-webadmin.main.unkin.net be_stalwart_webadmin'
|
||||
- 'autoconfig.main.unkin.net be_stalwart_webadmin'
|
||||
- 'autodiscovery.main.unkin.net be_stalwart_webadmin'
|
||||
|
||||
profiles::haproxy::frontends:
|
||||
fe_http:
|
||||
@ -62,6 +76,10 @@ profiles::haproxy::frontends:
|
||||
- 'acl_fafflix req.hdr(host) -i fafflix.unkin.net'
|
||||
- 'acl_gitea req.hdr(host) -i git.unkin.net'
|
||||
- 'acl_grafana req.hdr(host) -i grafana.unkin.net'
|
||||
- 'acl_ceph_dashboard req.hdr(host) -i dashboard.ceph.unkin.net'
|
||||
- 'acl_stalwart_webadmin req.hdr(host) -i mail-webadmin.main.unkin.net'
|
||||
- 'acl_stalwart_webadmin req.hdr(host) -i autoconfig.main.unkin.net'
|
||||
- 'acl_stalwart_webadmin req.hdr(host) -i autodiscovery.main.unkin.net'
|
||||
- 'acl_internalsubnets src 198.18.0.0/16 10.10.12.0/24'
|
||||
use_backend:
|
||||
- "%[req.hdr(host),lower,map(/etc/haproxy/fe_https.map,be_default)]"
|
||||
@ -79,6 +97,8 @@ profiles::haproxy::frontends:
|
||||
- 'set-header X-Frame-Options DENY if acl_fafflix'
|
||||
- 'set-header X-Frame-Options DENY if acl_gitea'
|
||||
- 'set-header X-Frame-Options DENY if acl_grafana'
|
||||
- 'set-header X-Frame-Options DENY if acl_ceph_dashboard'
|
||||
- 'set-header X-Frame-Options DENY if acl_stalwart_webadmin'
|
||||
- 'set-header X-Content-Type-Options nosniff'
|
||||
- 'set-header X-XSS-Protection 1;mode=block'
|
||||
|
||||
@ -262,6 +282,102 @@ profiles::haproxy::backends:
|
||||
redirect: 'scheme https if !{ ssl_fc }'
|
||||
stick-table: 'type ip size 200k expire 30m'
|
||||
stick: 'on src'
|
||||
be_ceph_dashboard:
|
||||
description: Backend for Ceph Dashboard from Mgr instances
|
||||
collect_exported: false # handled in custom function
|
||||
options:
|
||||
balance: roundrobin
|
||||
option:
|
||||
- httpchk GET /
|
||||
- forwardfor
|
||||
- http-keep-alive
|
||||
- prefer-last-server
|
||||
cookie: SRVNAME insert indirect nocache
|
||||
http-reuse: always
|
||||
http-check:
|
||||
- expect status 200
|
||||
http-request:
|
||||
- set-header X-Forwarded-Port %[dst_port]
|
||||
- add-header X-Forwarded-Proto https if { dst_port 9443 }
|
||||
redirect: 'scheme https if !{ ssl_fc }'
|
||||
stick-table: 'type ip size 200k expire 30m'
|
||||
be_stalwart_webadmin:
|
||||
description: Backend for Stalwart Webadmin
|
||||
collect_exported: false # handled in custom function
|
||||
options:
|
||||
balance: roundrobin
|
||||
option:
|
||||
- httpchk GET /
|
||||
- forwardfor
|
||||
- http-keep-alive
|
||||
- prefer-last-server
|
||||
cookie: SRVNAME insert indirect nocache
|
||||
http-reuse: always
|
||||
http-check:
|
||||
- expect status 200
|
||||
http-request:
|
||||
- set-header X-Forwarded-Port %[dst_port]
|
||||
- add-header X-Forwarded-Proto https if { dst_port 9443 }
|
||||
redirect: 'scheme https if !{ ssl_fc }'
|
||||
stick-table: 'type ip size 200k expire 30m'
|
||||
be_stalwart_imap:
|
||||
description: Backend for Stalwart IMAP (STARTTLS)
|
||||
collect_exported: false
|
||||
options:
|
||||
mode: tcp
|
||||
balance: roundrobin
|
||||
option:
|
||||
- tcp-check
|
||||
- prefer-last-server
|
||||
stick-table: 'type ip size 200k expire 30m'
|
||||
stick: 'on src'
|
||||
tcp-check:
|
||||
- connect port 143 send-proxy
|
||||
- expect string "* OK"
|
||||
- send "A001 STARTTLS\r\n"
|
||||
- expect rstring "A001 (OK|2.0.0)"
|
||||
be_stalwart_imaps:
|
||||
description: Backend for Stalwart IMAPS (implicit TLS)
|
||||
collect_exported: false
|
||||
options:
|
||||
mode: tcp
|
||||
balance: roundrobin
|
||||
option:
|
||||
- tcp-check
|
||||
- prefer-last-server
|
||||
stick-table: 'type ip size 200k expire 30m'
|
||||
stick: 'on src'
|
||||
tcp-check:
|
||||
- connect ssl send-proxy
|
||||
- expect string "* OK"
|
||||
be_stalwart_smtp:
|
||||
description: Backend for Stalwart SMTP
|
||||
collect_exported: false
|
||||
options:
|
||||
mode: tcp
|
||||
balance: roundrobin
|
||||
option:
|
||||
- tcp-check
|
||||
- prefer-last-server
|
||||
stick-table: 'type ip size 200k expire 30m'
|
||||
stick: 'on src'
|
||||
tcp-check:
|
||||
- connect port 25 send-proxy
|
||||
- expect string "220 "
|
||||
be_stalwart_submission:
|
||||
description: Backend for Stalwart SMTP Submission
|
||||
collect_exported: false
|
||||
options:
|
||||
mode: tcp
|
||||
balance: roundrobin
|
||||
option:
|
||||
- tcp-check
|
||||
- prefer-last-server
|
||||
stick-table: 'type ip size 200k expire 30m'
|
||||
stick: 'on src'
|
||||
tcp-check:
|
||||
- connect port 587 send-proxy
|
||||
- expect string "220 "
|
||||
|
||||
profiles::haproxy::certlist::enabled: true
|
||||
profiles::haproxy::certlist::certificates:
|
||||
@ -276,6 +392,7 @@ profiles::haproxy::certlist::certificates:
|
||||
- /etc/pki/tls/letsencrypt/fafflix.unkin.net/fullchain_combined.pem
|
||||
- /etc/pki/tls/letsencrypt/git.unkin.net/fullchain_combined.pem
|
||||
- /etc/pki/tls/letsencrypt/grafana.unkin.net/fullchain_combined.pem
|
||||
- /etc/pki/tls/letsencrypt/dashboard.ceph.unkin.net/fullchain_combined.pem
|
||||
- /etc/pki/tls/vault/certificate.pem
|
||||
|
||||
# additional altnames
|
||||
@ -283,6 +400,7 @@ profiles::pki::vault::alt_names:
|
||||
- au-syd1-pve.main.unkin.net
|
||||
- au-syd1-pve-api.main.unkin.net
|
||||
- jellyfin.main.unkin.net
|
||||
- mail-webadmin.main.unkin.net
|
||||
|
||||
# additional cnames
|
||||
profiles::haproxy::dns::cnames:
|
||||
@ -303,3 +421,4 @@ certbot::client::domains:
|
||||
- fafflix.unkin.net
|
||||
- git.unkin.net
|
||||
- grafana.unkin.net
|
||||
- dashboard.ceph.unkin.net
|
||||
|
||||
@ -7,17 +7,4 @@ profiles::packages::include:
|
||||
|
||||
profiles::yum::global::repos:
|
||||
powertools:
|
||||
name: powertools
|
||||
descr: powertools repository
|
||||
target: /etc/yum.repos.d/powertools.repo
|
||||
baseurl: https://edgecache.query.consul/almalinux/%{facts.os.release.full}/PowerTools/%{facts.os.architecture}/os
|
||||
gpgkey: http://edgecache.query.consul/almalinux/RPM-GPG-KEY-AlmaLinux-%{facts.os.release.major}
|
||||
mirrorlist: absent
|
||||
unkin:
|
||||
name: unkin
|
||||
descr: unkin repository
|
||||
target: /etc/yum.repos.d/unkin.repo
|
||||
baseurl: https://git.query.consul/api/packages/unkin/rpm/almalinux/el8
|
||||
gpgkey: https://git.query.consul/api/packages/unkin/rpm/repository.key
|
||||
gpgcheck: false
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
|
||||
@ -3,34 +3,5 @@
|
||||
crypto_policies::policy: 'DEFAULT:SHA1'
|
||||
|
||||
profiles::yum::global::repos:
|
||||
baseos:
|
||||
baseurl: https://packagerepo.service.consul/almalinux/%{facts.os.release.full}/baseos-daily/%{facts.os.architecture}/os/
|
||||
gpgkey: https://packagerepo.service.consul/almalinux/%{facts.os.release.full}/baseos-daily/%{facts.os.architecture}/os//RPM-GPG-KEY-AlmaLinux-9
|
||||
mirrorlist: absent
|
||||
extras:
|
||||
baseurl: https://packagerepo.service.consul/almalinux/%{facts.os.release.full}/extras-daily/%{facts.os.architecture}/os/
|
||||
gpgkey: https://packagerepo.service.consul/almalinux/%{facts.os.release.full}/extras-daily/%{facts.os.architecture}/os//RPM-GPG-KEY-AlmaLinux-9
|
||||
mirrorlist: absent
|
||||
appstream:
|
||||
baseurl: https://packagerepo.service.consul/almalinux/%{facts.os.release.full}/appstream-daily/%{facts.os.architecture}/os/
|
||||
gpgkey: https://packagerepo.service.consul/almalinux/%{facts.os.release.full}/appstream-daily/%{facts.os.architecture}/os//RPM-GPG-KEY-AlmaLinux-9
|
||||
mirrorlist: absent
|
||||
highavailability:
|
||||
baseurl: https://packagerepo.service.consul/almalinux/%{facts.os.release.full}/ha-daily/%{facts.os.architecture}/os/
|
||||
gpgkey: https://packagerepo.service.consul/almalinux/%{facts.os.release.full}/ha-daily/%{facts.os.architecture}/os//RPM-GPG-KEY-AlmaLinux-9
|
||||
mirrorlist: absent
|
||||
crb:
|
||||
name: crb
|
||||
descr: crb repository
|
||||
target: /etc/yum.repos.d/crb.repo
|
||||
baseurl: https://packagerepo.service.consul/almalinux/%{facts.os.release.full}/crb-daily/%{facts.os.architecture}/os/
|
||||
gpgkey: https://packagerepo.service.consul/almalinux/%{facts.os.release.full}/crb-daily/%{facts.os.architecture}/os//RPM-GPG-KEY-AlmaLinux-9
|
||||
mirrorlist: absent
|
||||
unkin:
|
||||
name: unkin
|
||||
descr: unkin repository
|
||||
target: /etc/yum.repos.d/unkin.repo
|
||||
baseurl: https://git.query.consul/api/packages/unkin/rpm/almalinux/el9
|
||||
gpgkey: https://git.query.consul/api/packages/unkin/rpm/repository.key
|
||||
gpgcheck: false
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
|
||||
@ -3,7 +3,8 @@
|
||||
profiles::firewall::firewalld::ensure_package: 'absent'
|
||||
profiles::firewall::firewalld::ensure_service: 'stopped'
|
||||
profiles::firewall::firewalld::enable_service: false
|
||||
profiles::puppet::agent::puppet_version: '7.34.0'
|
||||
profiles::puppet::agent::version: '7.37.2'
|
||||
profiles::puppet::agent::openvox_enable: true
|
||||
|
||||
hiera_include:
|
||||
- profiles::almalinux::base
|
||||
@ -22,43 +23,52 @@ profiles::yum::global::repos:
|
||||
name: baseos
|
||||
descr: baseos repository
|
||||
target: /etc/yum.repos.d/baseos.repo
|
||||
baseurl: https://edgecache.query.consul/almalinux/%{facts.os.release.full}/BaseOS/%{facts.os.architecture}/os
|
||||
gpgkey: http://edgecache.query.consul/almalinux/RPM-GPG-KEY-AlmaLinux-%{facts.os.release.major}
|
||||
baseurl: https://artifactapi.k8s.syd1.au.unkin.net/api/v1/remote/almalinux/%{facts.os.release.full}/BaseOS/%{facts.os.architecture}/os/
|
||||
gpgkey: https://artifactapi.k8s.syd1.au.unkin.net/api/v1/remote/almalinux/%{facts.os.release.full}/BaseOS/%{facts.os.architecture}/os/RPM-GPG-KEY-AlmaLinux-%{facts.os.release.major}
|
||||
mirrorlist: absent
|
||||
extras:
|
||||
name: extras
|
||||
descr: extras repository
|
||||
target: /etc/yum.repos.d/extras.repo
|
||||
baseurl: https://edgecache.query.consul/almalinux/%{facts.os.release.full}/extras/%{facts.os.architecture}/os
|
||||
gpgkey: http://edgecache.query.consul/almalinux/RPM-GPG-KEY-AlmaLinux-%{facts.os.release.major}
|
||||
baseurl: https://artifactapi.k8s.syd1.au.unkin.net/api/v1/remote/almalinux/%{facts.os.release.full}/extras/%{facts.os.architecture}/os/
|
||||
gpgkey: https://artifactapi.k8s.syd1.au.unkin.net/api/v1/remote/almalinux/%{facts.os.release.full}/extras/%{facts.os.architecture}/os/RPM-GPG-KEY-AlmaLinux-%{facts.os.release.major}
|
||||
mirrorlist: absent
|
||||
appstream:
|
||||
name: appstream
|
||||
descr: appstream repository
|
||||
target: /etc/yum.repos.d/appstream.repo
|
||||
baseurl: https://edgecache.query.consul/almalinux/%{facts.os.release.full}/AppStream/%{facts.os.architecture}/os
|
||||
gpgkey: http://edgecache.query.consul/almalinux/RPM-GPG-KEY-AlmaLinux-%{facts.os.release.major}
|
||||
baseurl: https://artifactapi.k8s.syd1.au.unkin.net/api/v1/remote/almalinux/%{facts.os.release.full}/AppStream/%{facts.os.architecture}/os/
|
||||
gpgkey: https://artifactapi.k8s.syd1.au.unkin.net/api/v1/remote/almalinux/%{facts.os.release.full}/AppStream/%{facts.os.architecture}/os/RPM-GPG-KEY-AlmaLinux-%{facts.os.release.major}
|
||||
mirrorlist: absent
|
||||
highavailability:
|
||||
name: highavailability
|
||||
descr: highavailability repository
|
||||
target: /etc/yum.repos.d/highavailability.repo
|
||||
baseurl: https://edgecache.query.consul/almalinux/%{facts.os.release.full}/HighAvailability/%{facts.os.architecture}/os
|
||||
gpgkey: http://edgecache.query.consul/almalinux/RPM-GPG-KEY-AlmaLinux-%{facts.os.release.major}
|
||||
baseurl: https://artifactapi.k8s.syd1.au.unkin.net/api/v1/remote/almalinux/%{facts.os.release.full}/HighAvailability/%{facts.os.architecture}/os/
|
||||
gpgkey: https://artifactapi.k8s.syd1.au.unkin.net/api/v1/remote/almalinux/%{facts.os.release.full}/HighAvailability/%{facts.os.architecture}/os/RPM-GPG-KEY-AlmaLinux-%{facts.os.release.major}
|
||||
mirrorlist: absent
|
||||
crb:
|
||||
ensure: absent
|
||||
name: crb
|
||||
descr: crb repository
|
||||
target: /etc/yum.repos.d/crb.repo
|
||||
baseurl: https://artifactapi.k8s.syd1.au.unkin.net/api/v1/remote/almalinux/%{facts.os.release.full}/CRB/%{facts.os.architecture}/os/
|
||||
gpgkey: https://artifactapi.k8s.syd1.au.unkin.net/api/v1/remote/almalinux/%{facts.os.release.full}/CRB/%{facts.os.architecture}/os/RPM-GPG-KEY-AlmaLinux-%{facts.os.release.major}
|
||||
mirrorlist: absent
|
||||
powertools:
|
||||
ensure: absent
|
||||
name: powertools
|
||||
descr: powertools repository
|
||||
target: /etc/yum.repos.d/powertools.repo
|
||||
baseurl: https://artifactapi.k8s.syd1.au.unkin.net/api/v1/remote/almalinux/%{facts.os.release.full}/PowerTools/%{facts.os.architecture}/os/
|
||||
gpgkey: https://artifactapi.k8s.syd1.au.unkin.net/api/v1/remote/almalinux/%{facts.os.release.full}/PowerTools/%{facts.os.architecture}/os/RPM-GPG-KEY-AlmaLinux-%{facts.os.release.major}
|
||||
mirrorlist: absent
|
||||
epel:
|
||||
name: epel
|
||||
descr: epel repository
|
||||
target: /etc/yum.repos.d/epel.repo
|
||||
baseurl: https://packagerepo.service.consul/epel/%{facts.os.release.major}/everything-daily/%{facts.os.architecture}/os/
|
||||
gpgkey: https://packagerepo.service.consul/epel/%{facts.os.release.major}/everything-daily/%{facts.os.architecture}/os/RPM-GPG-KEY-EPEL-%{facts.os.release.major}
|
||||
mirrorlist: absent
|
||||
puppet:
|
||||
name: puppet
|
||||
descr: puppet repository
|
||||
target: /etc/yum.repos.d/puppet.repo
|
||||
baseurl: https://packagerepo.service.consul/puppet7/el/%{facts.os.release.major}-daily/%{facts.os.architecture}/os/
|
||||
gpgkey: https://packagerepo.service.consul/puppet7/el/%{facts.os.release.major}-daily/%{facts.os.architecture}/os/RPM-GPG-KEY-puppet-20250406
|
||||
baseurl: https://artifactapi.k8s.syd1.au.unkin.net/api/v1/remote/epel/%{facts.os.release.major}/Everything/%{facts.os.architecture}
|
||||
gpgkey: https://artifactapi.k8s.syd1.au.unkin.net/api/v1/remote/epel/RPM-GPG-KEY-EPEL-%{facts.os.release.major}
|
||||
mirrorlist: absent
|
||||
unkinben:
|
||||
name: unkinben
|
||||
@ -68,3 +78,193 @@ profiles::yum::global::repos:
|
||||
gpgkey: https://git.query.consul/api/packages/unkinben/rpm/repository.key
|
||||
gpgcheck: false
|
||||
mirrorlist: absent
|
||||
|
||||
# Additional repositories - default to absent, roles can override with ensure: present
|
||||
# FRRouting repositories
|
||||
frr-extras:
|
||||
ensure: absent
|
||||
name: frr-extras
|
||||
descr: frr-extras repository
|
||||
target: /etc/yum.repos.d/frr-extras.repo
|
||||
baseurl: https://artifactapi.k8s.syd1.au.unkin.net/api/v1/remote/frr/el%{facts.os.release.major}/extras
|
||||
gpgcheck: false
|
||||
mirrorlist: absent
|
||||
frr-stable:
|
||||
ensure: absent
|
||||
name: frr-stable
|
||||
descr: frr-stable repository
|
||||
target: /etc/yum.repos.d/frr-stable.repo
|
||||
baseurl: https://artifactapi.k8s.syd1.au.unkin.net/api/v1/remote/frr/el%{facts.os.release.major}/frr
|
||||
gpgcheck: false
|
||||
mirrorlist: absent
|
||||
|
||||
# PostgreSQL repositories
|
||||
postgresql-15:
|
||||
ensure: absent
|
||||
name: postgresql-15
|
||||
descr: postgresql-15 repository
|
||||
target: /etc/yum.repos.d/postgresql.repo
|
||||
baseurl: https://artifactapi.k8s.syd1.au.unkin.net/api/v1/remote/postgresql/15/redhat/rhel-%{facts.os.release.major}-%{facts.os.architecture}
|
||||
gpgkey: https://artifactapi.k8s.syd1.au.unkin.net/api/v1/remote/postgresql/keys/PGDG-RPM-GPG-KEY-RHEL
|
||||
postgresql-17:
|
||||
ensure: absent
|
||||
name: postgresql-17
|
||||
descr: postgresql-17 repository
|
||||
target: /etc/yum.repos.d/postgresql.repo
|
||||
baseurl: https://artifactapi.k8s.syd1.au.unkin.net/api/v1/remote/postgresql/17/redhat/rhel-%{facts.os.release.major}-%{facts.os.architecture}
|
||||
gpgkey: https://artifactapi.k8s.syd1.au.unkin.net/api/v1/remote/postgresql/keys/PGDG-RPM-GPG-KEY-RHEL
|
||||
postgresql-common:
|
||||
ensure: absent
|
||||
name: postgresql-common
|
||||
descr: postgresql-common repository
|
||||
target: /etc/yum.repos.d/postgresql.repo
|
||||
baseurl: https://artifactapi.k8s.syd1.au.unkin.net/api/v1/remote/postgresql/common/redhat/rhel-%{facts.os.release.major}-%{facts.os.architecture}
|
||||
gpgkey: https://artifactapi.k8s.syd1.au.unkin.net/api/v1/remote/postgresql/keys/PGDG-RPM-GPG-KEY-RHEL
|
||||
|
||||
# Ceph repositories
|
||||
ceph:
|
||||
ensure: absent
|
||||
name: ceph
|
||||
descr: ceph repository
|
||||
target: /etc/yum.repos.d/ceph.repo
|
||||
baseurl: https://artifactapi.k8s.syd1.au.unkin.net/api/v1/remote/ceph-reef/el%{facts.os.release.major}/%{facts.os.architecture}
|
||||
gpgcheck: false
|
||||
mirrorlist: absent
|
||||
ceph-noarch:
|
||||
ensure: absent
|
||||
name: ceph-noarch
|
||||
descr: ceph noarch repository
|
||||
target: /etc/yum.repos.d/ceph.repo
|
||||
baseurl: https://artifactapi.k8s.syd1.au.unkin.net/api/v1/remote/ceph-reef/el%{facts.os.release.major}/noarch
|
||||
gpgcheck: false
|
||||
mirrorlist: absent
|
||||
|
||||
# Rancher RKE2 repositories
|
||||
rancher-rke2-common-latest:
|
||||
ensure: absent
|
||||
name: rancher-rke2-common
|
||||
descr: rancher-rke2-common repository
|
||||
target: /etc/yum.repos.d/rancher-rke2-common.repo
|
||||
baseurl: https://artifactapi.k8s.syd1.au.unkin.net/api/v1/remote/rke2/rke2/latest/common/centos/%{facts.os.release.major}/noarch
|
||||
gpgkey: https://artifactapi.k8s.syd1.au.unkin.net/api/v1/remote/rke2/public.key
|
||||
gpgcheck: 1
|
||||
mirrorlist: absent
|
||||
rancher-rke2-1-33-latest:
|
||||
ensure: absent
|
||||
name: rancher-rke2-1.33-latest
|
||||
descr: rancher-rke2-1.33-latest repository
|
||||
target: /etc/yum.repos.d/rancher-rke2.repo
|
||||
baseurl: https://artifactapi.k8s.syd1.au.unkin.net/api/v1/remote/rke2/rke2/latest/1.33/centos/%{facts.os.release.major}/%{facts.os.architecture}
|
||||
gpgkey: https://artifactapi.k8s.syd1.au.unkin.net/api/v1/remote/rke2/public.key
|
||||
gpgcheck: 1
|
||||
mirrorlist: absent
|
||||
|
||||
# CentOS repositories for legacy systems
|
||||
centos_8_advanced_virtualization:
|
||||
ensure: absent
|
||||
name: centos_8_advanced_virtualization
|
||||
descr: centos_8_advanced_virtualization repository
|
||||
target: /etc/yum.repos.d/centos.repo
|
||||
baseurl: https://edgecache.query.consul/centos/8/virt/x86_64/advanced-virtualization
|
||||
gpgkey: http://edgecache.query.consul/centos/RPM-GPG-KEY-CentOS-SIG-Virtualization
|
||||
gpgcheck: 1
|
||||
mirrorlist: absent
|
||||
centos_8_ceph_pacific:
|
||||
ensure: absent
|
||||
name: centos_8_ceph_pacific
|
||||
descr: centos_8_ceph_pacific repository
|
||||
target: /etc/yum.repos.d/centos.repo
|
||||
baseurl: https://edgecache.query.consul/centos/8/storage/x86_64/ceph-pacific
|
||||
gpgkey: http://edgecache.query.consul/centos/RPM-GPG-KEY-CentOS-SIG-Storage
|
||||
gpgcheck: 1
|
||||
mirrorlist: absent
|
||||
centos_8_rabbitmq_38:
|
||||
ensure: absent
|
||||
name: centos_8_rabbitmq_38
|
||||
descr: centos_8_rabbitmq_38 repository
|
||||
target: /etc/yum.repos.d/centos.repo
|
||||
baseurl: https://edgecache.query.consul/centos/8/messaging/x86_64/rabbitmq-38
|
||||
gpgkey: http://edgecache.query.consul/centos/RPM-GPG-KEY-CentOS-SIG-Messaging
|
||||
gpgcheck: 1
|
||||
mirrorlist: absent
|
||||
centos_8_nfv_openvswitch:
|
||||
ensure: absent
|
||||
name: centos_8_nfv_openvswitch
|
||||
descr: centos_8_nfv_openvswitch repository
|
||||
target: /etc/yum.repos.d/centos.repo
|
||||
baseurl: https://edgecache.query.consul/centos/8/nfv/x86_64/openvswitch-2
|
||||
gpgkey: http://edgecache.query.consul/centos/RPM-GPG-KEY-CentOS-SIG-NFV
|
||||
gpgcheck: 1
|
||||
mirrorlist: absent
|
||||
centos_8_openstack_xena:
|
||||
ensure: absent
|
||||
name: centos_8_openstack_xena
|
||||
descr: centos_8_openstack_xena repository
|
||||
target: /etc/yum.repos.d/centos.repo
|
||||
baseurl: https://edgecache.query.consul/centos/8/cloud/x86_64/openstack-xena
|
||||
gpgkey: http://edgecache.query.consul/centos/RPM-GPG-KEY-CentOS-SIG-Cloud
|
||||
gpgcheck: 1
|
||||
mirrorlist: absent
|
||||
centos_8_opstools:
|
||||
ensure: absent
|
||||
name: centos_8_opstools
|
||||
descr: centos_8_opstools repository
|
||||
target: /etc/yum.repos.d/centos.repo
|
||||
baseurl: https://edgecache.query.consul/centos/8/opstools/x86_64/collectd-5
|
||||
gpgkey: http://edgecache.query.consul/centos/RPM-GPG-KEY-CentOS-SIG-OpsTools
|
||||
gpgcheck: 1
|
||||
mirrorlist: absent
|
||||
centos_8_ovirt45:
|
||||
ensure: absent
|
||||
name: centos_8_ovirt45
|
||||
descr: centos_8_ovirt45 repository
|
||||
target: /etc/yum.repos.d/centos.repo
|
||||
baseurl: https://edgecache.query.consul/centos/8/virt/x86_64/ovirt-45
|
||||
gpgkey: http://edgecache.query.consul/centos/RPM-GPG-KEY-CentOS-SIG-Virtualization
|
||||
gpgcheck: 1
|
||||
mirrorlist: absent
|
||||
centos_8_stream_gluster10:
|
||||
ensure: absent
|
||||
name: centos_8_stream_gluster10
|
||||
descr: centos_8_stream_gluster10 repository
|
||||
target: /etc/yum.repos.d/centos.repo
|
||||
baseurl: https://edgecache.query.consul/centos/8-stream/storage/x86_64/gluster-10
|
||||
gpgkey: http://edgecache.query.consul/centos/RPM-GPG-KEY-CentOS-SIG-Storage
|
||||
gpgcheck: 1
|
||||
mirrorlist: absent
|
||||
|
||||
# Additional repositories
|
||||
zfs-kmod:
|
||||
ensure: absent
|
||||
name: zfs-kmod
|
||||
descr: zfs-kmod repository
|
||||
target: /etc/yum.repos.d/zfs.repo
|
||||
baseurl: https://artifactapi.k8s.syd1.au.unkin.net/api/v1/remote/zfs/epel/%{facts.os.release.major}/kmod/%{facts.os.architecture}/
|
||||
gpgcheck: false
|
||||
mirrorlist: absent
|
||||
rpmfusion-free:
|
||||
ensure: absent
|
||||
name: rpmfusion-free
|
||||
descr: rpmfusion-free repository
|
||||
target: /etc/yum.repos.d/rpmfusion-free.repo
|
||||
baseurl: https://packagerepo.service.consul/rpmfusion-free-el%{facts.os.release.major}-%{facts.os.architecture}/
|
||||
gpgkey: https://packagerepo.service.consul/rpmfusion-free-el%{facts.os.release.major}-%{facts.os.architecture}/repodata/repomd.xml.key
|
||||
gpgcheck: 1
|
||||
mirrorlist: absent
|
||||
rpmfusion-nonfree:
|
||||
ensure: absent
|
||||
name: rpmfusion-nonfree
|
||||
descr: rpmfusion-nonfree repository
|
||||
target: /etc/yum.repos.d/rpmfusion-nonfree.repo
|
||||
baseurl: https://packagerepo.service.consul/rpmfusion-nonfree-el%{facts.os.release.major}-%{facts.os.architecture}/
|
||||
gpgkey: https://packagerepo.service.consul/rpmfusion-nonfree-el%{facts.os.release.major}-%{facts.os.architecture}/repodata/repomd.xml.key
|
||||
gpgcheck: 1
|
||||
mirrorlist: absent
|
||||
unkin:
|
||||
name: unkin
|
||||
descr: unkin repository
|
||||
target: /etc/yum.repos.d/unkin.repo
|
||||
baseurl: https://git.query.consul/api/packages/unkin/rpm/almalinux/el%{facts.os.release.major}
|
||||
gpgkey: https://git.query.consul/api/packages/unkin/rpm/repository.key
|
||||
gpgcheck: false
|
||||
mirrorlist: absent
|
||||
|
||||
@ -11,4 +11,4 @@ profiles::apt::components:
|
||||
- main
|
||||
- non-free
|
||||
|
||||
profiles::puppet::agent::puppet_version: '7.25.0-1bullseye'
|
||||
profiles::puppet::agent::version: '7.25.0-1bullseye'
|
||||
|
||||
@ -12,4 +12,4 @@ profiles::apt::components:
|
||||
- non-free
|
||||
- non-free-firmware
|
||||
|
||||
profiles::puppet::agent::puppet_version: 'latest'
|
||||
profiles::puppet::agent::version: 'latest'
|
||||
|
||||
@ -3,13 +3,8 @@ hiera_include:
|
||||
- profiles::nginx::simpleproxy
|
||||
|
||||
profiles::yum::global::repos:
|
||||
ceph-reef:
|
||||
name: ceph-reef
|
||||
descr: ceph reef repository
|
||||
target: /etc/yum.repos.d/ceph-reef.repo
|
||||
baseurl: https://edgecache.query.consul/ceph/yum/el%{facts.os.release.major}/%{facts.os.architecture}
|
||||
gpgcheck: 0,
|
||||
mirrorlist: absent
|
||||
ceph:
|
||||
ensure: present
|
||||
|
||||
profiles::ceph::client::keyrings:
|
||||
media:
|
||||
|
||||
@ -54,24 +54,8 @@ profiles::consul::client::node_rules:
|
||||
|
||||
profiles::yum::global::repos:
|
||||
rpmfusion-free:
|
||||
name: rpmfusion-free
|
||||
descr: rpmfusion-free repository
|
||||
target: /etc/yum.repos.d/rpmfusion.repo
|
||||
baseurl: https://download1.rpmfusion.org/free/el/updates/%{facts.os.release.major}/%{facts.os.architecture}
|
||||
gpgkey: https://download1.rpmfusion.org/free/el/RPM-GPG-KEY-rpmfusion-free-el-%{facts.os.release.major}
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
rpmfusion-nonfree:
|
||||
name: rpmfusion-nonfree
|
||||
descr: rpmfusion-nonfree repository
|
||||
target: /etc/yum.repos.d/rpmfusion.repo
|
||||
baseurl: https://download1.rpmfusion.org/nonfree/el/updates/%{facts.os.release.major}/%{facts.os.architecture}
|
||||
gpgkey: https://download1.rpmfusion.org/nonfree/el/RPM-GPG-KEY-rpmfusion-nonfree-el-%{facts.os.release.major}
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
unkinben:
|
||||
name: unkinben
|
||||
descr: unkinben repository
|
||||
target: /etc/yum.repos.d/unkin.repo
|
||||
baseurl: https://git.query.consul/api/packages/unkinben/rpm/el8
|
||||
gpgkey: https://git.query.consul/api/packages/unkinben/rpm/repository.key
|
||||
gpgcheck: false
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
|
||||
@ -72,16 +72,6 @@ profiles::consul::client::node_rules:
|
||||
|
||||
profiles::yum::global::repos:
|
||||
rpmfusion-free:
|
||||
name: rpmfusion-free
|
||||
descr: rpmfusion-free repository
|
||||
target: /etc/yum.repos.d/rpmfusion.repo
|
||||
baseurl: https://download1.rpmfusion.org/free/el/updates/%{facts.os.release.major}/%{facts.os.architecture}
|
||||
gpgkey: https://download1.rpmfusion.org/free/el/RPM-GPG-KEY-rpmfusion-free-el-%{facts.os.release.major}
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
rpmfusion-nonfree:
|
||||
name: rpmfusion-nonfree
|
||||
descr: rpmfusion-nonfree repository
|
||||
target: /etc/yum.repos.d/rpmfusion.repo
|
||||
baseurl: https://download1.rpmfusion.org/nonfree/el/updates/%{facts.os.release.major}/%{facts.os.architecture}
|
||||
gpgkey: https://download1.rpmfusion.org/nonfree/el/RPM-GPG-KEY-rpmfusion-nonfree-el-%{facts.os.release.major}
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
|
||||
@ -70,7 +70,7 @@ profiles::nginx::simpleproxy::locations:
|
||||
arrstack_web_external:
|
||||
location_satisfy: any
|
||||
location_allow:
|
||||
- 198.18.13.47
|
||||
- 198.18.13.50
|
||||
- 198.18.13.51
|
||||
- 198.18.13.52
|
||||
- 198.18.26.161
|
||||
- 198.18.27.131
|
||||
- 198.18.28.165
|
||||
- 198.18.29.32
|
||||
|
||||
@ -31,30 +31,10 @@ frrouting::daemons:
|
||||
# additional repos
|
||||
profiles::yum::global::repos:
|
||||
ceph:
|
||||
name: ceph
|
||||
descr: ceph repository
|
||||
target: /etc/yum.repos.d/ceph.repo
|
||||
baseurl: https://edgecache.query.consul/ceph/yum/el%{facts.os.release.major}/%{facts.os.architecture}
|
||||
gpgkey: https://download.ceph.com/keys/release.asc
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
ceph-noarch:
|
||||
name: ceph-noarch
|
||||
descr: ceph-noarch repository
|
||||
target: /etc/yum.repos.d/ceph-noarch.repo
|
||||
baseurl: https://edgecache.query.consul/ceph/yum/el%{facts.os.release.major}/noarch
|
||||
gpgkey: https://download.ceph.com/keys/release.asc
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
frr-extras:
|
||||
name: frr-extras
|
||||
descr: frr-extras repository
|
||||
target: /etc/yum.repos.d/frr-extras.repo
|
||||
baseurl: https://packagerepo.service.consul/frr/el9/extras-daily/%{facts.os.architecture}/os
|
||||
gpgkey: https://packagerepo.service.consul/frr/el9/extras-daily/%{facts.os.architecture}/os/RPM-GPG-KEY-FRR
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
frr-stable:
|
||||
name: frr-stable
|
||||
descr: frr-stable repository
|
||||
target: /etc/yum.repos.d/frr-stable.repo
|
||||
baseurl: https://packagerepo.service.consul/frr/el9/stable-daily/%{facts.os.architecture}/os
|
||||
gpgkey: https://packagerepo.service.consul/frr/el9/stable-daily/%{facts.os.architecture}/os/RPM-GPG-KEY-FRR
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
|
||||
@ -66,6 +66,9 @@ glauth::users:
|
||||
- 20025 # jupyterhub_admin
|
||||
- 20026 # jupyterhub_user
|
||||
- 20027 # grafana_user
|
||||
- 20028 # k8s/au/syd1 operator
|
||||
- 20029 # k8s/au/syd1 admin
|
||||
- 20030 # k8s/au/syd1 root
|
||||
loginshell: '/bin/bash'
|
||||
homedir: '/home/benvin'
|
||||
passsha256: 'd2434f6b4764ef75d5b7b96a876a32deedbd6aa726a109c3f32e823ca66f604a'
|
||||
@ -211,6 +214,36 @@ glauth::users:
|
||||
loginshell: '/bin/bash'
|
||||
homedir: '/home/waewak'
|
||||
passsha256: 'd9bb99634215fe031c3bdca94149a165192fe8384ecaa238a19354c2f760a811'
|
||||
debvin:
|
||||
user_name: 'debvin'
|
||||
givenname: 'Debbie'
|
||||
sn: 'Vincent'
|
||||
mail: 'debvin@users.main.unkin.net'
|
||||
uidnumber: 20009
|
||||
primarygroup: 20000
|
||||
othergroups:
|
||||
- 20010 # jelly
|
||||
loginshell: '/bin/bash'
|
||||
homedir: '/home/debvin'
|
||||
passsha256: 'cdac05ddb02e665d4ea65a974995f38a10236bc158731d92d78f6cde89b294a1'
|
||||
jassol:
|
||||
user_name: 'jassol'
|
||||
givenname: 'Jason'
|
||||
sn: 'Solomon'
|
||||
mail: 'jassol@users.main.unkin.net'
|
||||
uidnumber: 20010
|
||||
primarygroup: 20000
|
||||
othergroups:
|
||||
- 20010 # jelly
|
||||
- 20011 # sonarr
|
||||
- 20012 # radarr
|
||||
- 20013 # lidarr
|
||||
- 20014 # readarr
|
||||
- 20016 # nzbget
|
||||
- 20027 # grafana user
|
||||
loginshell: '/bin/bash'
|
||||
homedir: '/home/jassol'
|
||||
passsha256: 'd8e215d3c94b954e1318c9c7243ce72713f2fb1d006037724fe857c1fb7e88e9'
|
||||
|
||||
glauth::services:
|
||||
svc_jellyfin:
|
||||
@ -355,3 +388,12 @@ glauth::groups:
|
||||
grafana_user:
|
||||
group_name: 'grafana_user'
|
||||
gidnumber: 20027
|
||||
kubernetes_au_syd1_cluster_operator:
|
||||
group_name: 'kubernetes_au_syd1_cluster_operator'
|
||||
gidnumber: 20028
|
||||
kubernetes_au_syd1_cluster_admin:
|
||||
group_name: 'kubernetes_au_syd1_cluster_admin'
|
||||
gidnumber: 20029
|
||||
kubernetes_au_syd1_cluster_root:
|
||||
group_name: 'kubernetes_au_syd1_cluster_root'
|
||||
gidnumber: 20030
|
||||
|
||||
8
hieradata/roles/infra/ceph/rgw.eyaml
Normal file
8
hieradata/roles/infra/ceph/rgw.eyaml
Normal file
@ -0,0 +1,8 @@
|
||||
---
|
||||
|
||||
profiles::ceph::rgw::ceph_client_keys:
|
||||
ausyd1nxvm2115: ENC[PKCS7,MIIBmQYJKoZIhvcNAQcDoIIBijCCAYYCAQAxggEhMIIBHQIBADAFMAACAQEwDQYJKoZIhvcNAQEBBQAEggEAovOUUtTm/9nXWj6D+kVhmbZedVVkA5N80ULNYqISrv0A+KlXCx/V2sB56SpJ9eL6aPHdC+CKkrtrrdo1h2j2em7wTA5ghhbFVF0NIS7bbn0jzpl5YqbUrNxOtDSpjzX1aCGnyMvw69Lp+NcHwxIj+1XFgK6335138s9wbX3HmYF3jcsQkMqTzynELe1OQPWFXVKTjUfFFLdCQOFryp8UY8L9j/PpV6wd4w6p7R6eXhX21rjSaN4aqN1zjsnF2OVhL8Ge0QxMhePWKGOqsUfi72kh3II028DjqU0DcZQvoxnoqRPyUUjysH0nTKoLeXOGNgJdphY1dHBJ+SJnw2gqQDBcBgkqhkiG9w0BBwEwHQYJYIZIAWUDBAEqBBB9wA3ZJjmU95W+9r8va5uagDAMAv2APfdzrPnAU7NJPL+IW08osGuQWWamqF+XGVeHRoBmoFKwZ7grYRV2e3aabyc=]
|
||||
ausyd1nxvm2116: ENC[PKCS7,MIIBmQYJKoZIhvcNAQcDoIIBijCCAYYCAQAxggEhMIIBHQIBADAFMAACAQEwDQYJKoZIhvcNAQEBBQAEggEAf5ksy/pyUZSwTh+HiKw+1Uhj16A0DVZEKAbkKUQzXVmc+QpL4Dn7YBoXlEwrY8CcsTrxTQjADvtu9FC3o34QIdh06noSgYYA+7fna2A+9+oYoNtgwC3b8LeglxO/SQ9dKoJ90jRtmlw5P/CtrxA2RelMK6FNRekp1CaWMM4q20fJGgr/E33vgx38UJyp4/q0bTu2lLehCuDUP80j3XGbSNZ2snfYdIo91Cl+nSxLSU2TdnFpWaabsH19HwDnkWGiILlLBVvvhY7copCxs5DS1ueoOTCsqnWSrTrBMJjnu7WZd/s4NLw/0q/UP5xcFA51caY3Kv+sI6bfIYkNoLazwDBcBgkqhkiG9w0BBwEwHQYJYIZIAWUDBAEqBBDrCF16nrtukRPamx1VbGY+gDAK5dw0kV8MpATpwxTkJG6JtlFlwdpU9THs1bNwqSRD1ZhEWxQeWwsyyTtjUXi4bP8=]
|
||||
ausyd1nxvm2117: ENC[PKCS7,MIIBmQYJKoZIhvcNAQcDoIIBijCCAYYCAQAxggEhMIIBHQIBADAFMAACAQEwDQYJKoZIhvcNAQEBBQAEggEAKtvsgDk2QTmL8flBBE4nA43sRSnroq4I6T2CUAYv/lRdzCCWvE961o/51zyQEz8L5QyDo7L3gcGYIBqthYjRe9Gp5a5d4qds0qskupgQKnb0KR2RLFVUtH5vxHqyJZHjXaP+PQreyRoSIfRWXAdoZu544FeJ9DKKbmEzZaH5B2OdDMrf4Ufuud0maEAw0PJthS//ghCfGi74F1xlJnIWVvMhp66b0iMxC+ACClEHunG3oKx7M/w05HllG0wcxPTg4PFrbnFXjRuIxsykF9aVHJkRnCdgbMXRM4o6FrYyZRR74F1HKRujFCUA7kYWDKLxHxJpYCvCHp4HMhfzjs824zBcBgkqhkiG9w0BBwEwHQYJYIZIAWUDBAEqBBBbaSig6kgVVNfmSI53sNimgDDQ5O10Dzfa7S7RdJVLLUFBaZ5TG2g3Bwmy0k3wKZvABYMuYyOxQdfk6eMsKC+sC5w=]
|
||||
ausyd1nxvm2118: ENC[PKCS7,MIIBmQYJKoZIhvcNAQcDoIIBijCCAYYCAQAxggEhMIIBHQIBADAFMAACAQEwDQYJKoZIhvcNAQEBBQAEggEAdKl0ude3ZxW0ihyA758mpQp43qZvzTI2Bp83WzCp2ifJCgAIjRWdd01P64rSHaa4lus/wqB9jxg38g6DrN4ejX56Y/CJ6GQKxz6b1BO5nDfsLx6QEzCt+cfg5d/PPoTtEpz2VSvDfxFUrHiABA6++Sqzb9Og+nQCFMYJD3NHCk67QpkjPGQ/ejZk4MNXZQVCfKOlFqay/fF0jEmQixFOlX/Fdm9UoKttbrKluUmzpaVUzfGRaTTFVgzc3x2t/z1q1k0P7ClI9Uu02kUXpFVs9LPX99Zc2GtrnP06mYqqARhWF1NMK0zlmxtKpfObahRP/HmtI3fgnQsU1Cpwah0emTBcBgkqhkiG9w0BBwEwHQYJYIZIAWUDBAEqBBAILqpYx3FKY3xXLJRu2oDlgDCIOXeX6hxpu0qpj5c/9jMUSeV2DIydnxO+MiT3mceS50ip8B+zGQy5UedPmLt36Zs=]
|
||||
ausyd1nxvm2119: ENC[PKCS7,MIIBmQYJKoZIhvcNAQcDoIIBijCCAYYCAQAxggEhMIIBHQIBADAFMAACAQEwDQYJKoZIhvcNAQEBBQAEggEASlZglUxazp+9azfV3QkgRv+ACo+MH0RO5b18blbelgdmr38iwK7MwwEFpEfVJEyc/ph9RunWwrMmofDQHj5bBribfzZ2pH2CGiOrR0i5lZMtN0yQXPBA/+jm1Pi1AWGJLtoquuhMbibuHOTiXwBCBVrHHHaFTR5Xt34ABN/p/mCaG+N9nWux93msHCCextCalKBMmPhmI2q6HodfjanEVgYAe3/5hRPnpsi6IGSDNGygsTC3MG+hjGMpNF8izbwk9Lpzn6kY51aeNxI2ed9Jm8UZ/k+8b+o7ZQyWIBbf7DTFpEzk4G46puaDbXIorBWQ4azCjN3gt8VB91hwihtzcDBcBgkqhkiG9w0BBwEwHQYJYIZIAWUDBAEqBBB6U+9z4cSzMTA1z9bmoX82gDBfy5zbRPK8GxImJo6evecMOTtaY2c4aEnESXtGBCS02enmxljv9dv1UYQD0/a6S3A=]
|
||||
50
hieradata/roles/infra/ceph/rgw.yaml
Normal file
50
hieradata/roles/infra/ceph/rgw.yaml
Normal file
@ -0,0 +1,50 @@
|
||||
---
|
||||
hiera_include:
|
||||
- profiles::ceph::rgw
|
||||
- profiles::nginx::simpleproxy
|
||||
|
||||
profiles::ceph::rgw::enable: true
|
||||
|
||||
# FIXME: puppet-python wants to try manage python-dev, which is required by the ceph package
|
||||
python::manage_dev_package: false
|
||||
|
||||
# additional altnames
|
||||
profiles::pki::vault::alt_names:
|
||||
- radosgw.main.unkin.net
|
||||
- radosgw.service.consul
|
||||
- radosgw.query.consul
|
||||
- "radosgw.service.%{facts.country}-%{facts.region}.consul"
|
||||
|
||||
# additional repos
|
||||
profiles::yum::global::repos:
|
||||
ceph:
|
||||
ensure: present
|
||||
ceph-noarch:
|
||||
ensure: present
|
||||
|
||||
# manage a simple nginx reverse proxy
|
||||
profiles::nginx::simpleproxy::nginx_vhost: 'radosgw.service.consul'
|
||||
profiles::nginx::simpleproxy::nginx_aliases:
|
||||
- radosgw.service.au-syd1.consul
|
||||
profiles::nginx::simpleproxy::proxy_port: 7480
|
||||
profiles::nginx::simpleproxy::proxy_path: '/'
|
||||
nginx::client_max_body_size: 5000M
|
||||
|
||||
# manage consul service
|
||||
consul::services:
|
||||
radosgw:
|
||||
service_name: 'radosgw'
|
||||
address: "%{facts.networking.ip}"
|
||||
port: 443
|
||||
checks:
|
||||
- id: 'radosgw_https_check'
|
||||
name: 'RADOSGW HTTPS Check'
|
||||
http: "https://%{facts.networking.fqdn}:443"
|
||||
method: 'GET'
|
||||
tls_skip_verify: true
|
||||
interval: '10s'
|
||||
timeout: '1s'
|
||||
profiles::consul::client::node_rules:
|
||||
- resource: service
|
||||
segment: radosgw
|
||||
disposition: write
|
||||
@ -57,19 +57,9 @@ profiles::consul::client::node_rules:
|
||||
# additional repos
|
||||
profiles::yum::global::repos:
|
||||
frr-extras:
|
||||
name: frr-extras
|
||||
descr: frr-extras repository
|
||||
target: /etc/yum.repos.d/frr-extras.repo
|
||||
baseurl: https://packagerepo.service.consul/frr/el%{facts.os.release.major}/extras-daily/%{facts.os.architecture}/os
|
||||
gpgkey: https://packagerepo.service.consul/frr/el%{facts.os.release.major}/extras-daily/%{facts.os.architecture}/os/RPM-GPG-KEY-FRR
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
frr-stable:
|
||||
name: frr-stable
|
||||
descr: frr-stable repository
|
||||
target: /etc/yum.repos.d/frr-stable.repo
|
||||
baseurl: https://packagerepo.service.consul/frr/el%{facts.os.release.major}/stable-daily/%{facts.os.architecture}/os
|
||||
gpgkey: https://packagerepo.service.consul/frr/el%{facts.os.release.major}/stable-daily/%{facts.os.architecture}/os/RPM-GPG-KEY-FRR
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
|
||||
# cobbler settings
|
||||
profiles::cobbler::params::service_cname: 'cobbler.main.unkin.net'
|
||||
|
||||
@ -41,19 +41,9 @@ profiles::consul::client::node_rules:
|
||||
# additional repos
|
||||
profiles::yum::global::repos:
|
||||
frr-extras:
|
||||
name: frr-extras
|
||||
descr: frr-extras repository
|
||||
target: /etc/yum.repos.d/frr-extras.repo
|
||||
baseurl: https://packagerepo.service.consul/frr/el9/extras-daily/%{facts.os.architecture}/os
|
||||
gpgkey: https://packagerepo.service.consul/frr/el9/extras-daily/%{facts.os.architecture}/os/RPM-GPG-KEY-FRR
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
frr-stable:
|
||||
name: frr-stable
|
||||
descr: frr-stable repository
|
||||
target: /etc/yum.repos.d/frr-stable.repo
|
||||
baseurl: https://packagerepo.service.consul/frr/el9/stable-daily/%{facts.os.architecture}/os
|
||||
gpgkey: https://packagerepo.service.consul/frr/el9/stable-daily/%{facts.os.architecture}/os/RPM-GPG-KEY-FRR
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
|
||||
profiles::dhcp::server::ntpservers:
|
||||
- 0.au.pool.ntp.org
|
||||
|
||||
2
hieradata/roles/infra/dns/externaldns.eyaml
Normal file
2
hieradata/roles/infra/dns/externaldns.eyaml
Normal file
@ -0,0 +1,2 @@
|
||||
---
|
||||
externaldns::externaldns_key_secret: ENC[PKCS7,MIIBmQYJKoZIhvcNAQcDoIIBijCCAYYCAQAxggEhMIIBHQIBADAFMAACAQEwDQYJKoZIhvcNAQEBBQAEggEABqbZiK1NDTU+w2k7orz2HrB0EXwun7hn4pR6TeCHMp2IfrkPxlQT+f1J9c0PqJaAKvnyz+Cx0xNCrlnONqk+J57f48kYKYV+Vw+L0AYHYFj8/TizY5CwLpJS2XKyfRd4iEsWMonvfIYn71t3+YuXm4dkoEqGekW93qCr/KFtjAu0K3e+ypyl4EJqWokiUs7IbcSBNvrjUkP4yR8F/wHVKM1E5yfr+D1+nmMmt7Ob/J+am14492TppE2C7Xadg4us+kdYtuBsv9kTSi1GwwqUDjbeJVmfK3pKHjXdF+PI07AFLzo5bBZTJOzQfQ4SywpH8R5BDQoUCyHiaskB5wrmSDBcBgkqhkiG9w0BBwEwHQYJYIZIAWUDBAEqBBB2LU9ZhefSg9PqqkwnfV65gDBvXuXco0moKCGjHqm5KcojWCK1BoS/+mltlr8kw9grZjN9jxHRLn1FjgBlq418c8w=]
|
||||
55
hieradata/roles/infra/dns/externaldns.yaml
Normal file
55
hieradata/roles/infra/dns/externaldns.yaml
Normal file
@ -0,0 +1,55 @@
|
||||
---
|
||||
hiera_include:
|
||||
- externaldns
|
||||
- frrouting
|
||||
- exporters::frr_exporter
|
||||
|
||||
externaldns::bind_master_hostname: 'ausyd1nxvm2127.main.unkin.net'
|
||||
externaldns::k8s_zones:
|
||||
- 'k8s.syd1.au.unkin.net'
|
||||
- '200.18.198.in-addr.arpa'
|
||||
externaldns::slave_servers:
|
||||
- 'ausyd1nxvm2128.main.unkin.net'
|
||||
- 'ausyd1nxvm2129.main.unkin.net'
|
||||
externaldns::externaldns_key_algorithm: 'hmac-sha256'
|
||||
|
||||
# networking
|
||||
anycast_ip: 198.18.19.20
|
||||
systemd::manage_networkd: true
|
||||
systemd::manage_all_network_files: true
|
||||
networking::interfaces:
|
||||
eth0:
|
||||
type: physical
|
||||
forwarding: true
|
||||
dhcp: true
|
||||
anycast0:
|
||||
type: dummy
|
||||
ipaddress: "%{hiera('anycast_ip')}"
|
||||
netmask: 255.255.255.255
|
||||
mtu: 1500
|
||||
|
||||
# frrouting
|
||||
exporters::frr_exporter::enable: true
|
||||
frrouting::ospfd_router_id: "%{facts.networking.ip}"
|
||||
frrouting::ospfd_redistribute:
|
||||
- connected
|
||||
frrouting::ospfd_interfaces:
|
||||
eth0:
|
||||
area: 0.0.0.0
|
||||
anycast0:
|
||||
area: 0.0.0.0
|
||||
frrouting::daemons:
|
||||
ospfd: true
|
||||
|
||||
# consul
|
||||
profiles::consul::client::node_rules:
|
||||
- resource: service
|
||||
segment: frr_exporter
|
||||
disposition: write
|
||||
|
||||
# additional repos
|
||||
profiles::yum::global::repos:
|
||||
frr-extras:
|
||||
ensure: present
|
||||
frr-stable:
|
||||
ensure: present
|
||||
@ -200,16 +200,6 @@ profiles::consul::client::node_rules:
|
||||
# additional repos
|
||||
profiles::yum::global::repos:
|
||||
frr-extras:
|
||||
name: frr-extras
|
||||
descr: frr-extras repository
|
||||
target: /etc/yum.repos.d/frr-extras.repo
|
||||
baseurl: https://packagerepo.service.consul/frr/el9/extras-daily/%{facts.os.architecture}/os
|
||||
gpgkey: https://packagerepo.service.consul/frr/el9/extras-daily/%{facts.os.architecture}/os/RPM-GPG-KEY-FRR
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
frr-stable:
|
||||
name: frr-stable
|
||||
descr: frr-stable repository
|
||||
target: /etc/yum.repos.d/frr-stable.repo
|
||||
baseurl: https://packagerepo.service.consul/frr/el9/stable-daily/%{facts.os.architecture}/os
|
||||
gpgkey: https://packagerepo.service.consul/frr/el9/stable-daily/%{facts.os.architecture}/os/RPM-GPG-KEY-FRR
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
|
||||
@ -82,6 +82,11 @@ profiles::dns::resolver::zones:
|
||||
- 10.10.16.32
|
||||
- 10.10.16.33
|
||||
forward: 'only'
|
||||
k8s.syd1.au.unkin.net-forward:
|
||||
domain: 'k8s.syd1.au.unkin.net'
|
||||
zone_type: 'forward'
|
||||
forwarders: "%{alias('profiles_dns_upstream_forwarder_k8s')}"
|
||||
forward: 'only'
|
||||
unkin.net-forward:
|
||||
domain: 'unkin.net'
|
||||
zone_type: 'forward'
|
||||
@ -172,6 +177,11 @@ profiles::dns::resolver::zones:
|
||||
zone_type: 'forward'
|
||||
forwarders: "%{alias('profiles_dns_upstream_forwarder_unkin')}"
|
||||
forward: 'only'
|
||||
200.18.198.in-addr.arpa-forward:
|
||||
domain: '200.18.198.in-addr.arpa'
|
||||
zone_type: 'forward'
|
||||
forwarders: "%{alias('profiles_dns_upstream_forwarder_k8s')}"
|
||||
forward: 'only'
|
||||
consul-forward:
|
||||
domain: 'consul'
|
||||
zone_type: 'forward'
|
||||
@ -188,6 +198,7 @@ profiles::dns::resolver::views:
|
||||
- network.unkin.net-forward
|
||||
- prod.unkin.net-forward
|
||||
- consul-forward
|
||||
- k8s.syd1.au.unkin.net-forward
|
||||
- 13.18.198.in-addr.arpa-forward
|
||||
- 14.18.198.in-addr.arpa-forward
|
||||
- 15.18.198.in-addr.arpa-forward
|
||||
@ -250,16 +261,6 @@ profiles::consul::client::node_rules:
|
||||
# additional repos
|
||||
profiles::yum::global::repos:
|
||||
frr-extras:
|
||||
name: frr-extras
|
||||
descr: frr-extras repository
|
||||
target: /etc/yum.repos.d/frr-extras.repo
|
||||
baseurl: https://packagerepo.service.consul/frr/el9/extras-daily/%{facts.os.architecture}/os
|
||||
gpgkey: https://packagerepo.service.consul/frr/el9/extras-daily/%{facts.os.architecture}/os/RPM-GPG-KEY-FRR
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
frr-stable:
|
||||
name: frr-stable
|
||||
descr: frr-stable repository
|
||||
target: /etc/yum.repos.d/frr-stable.repo
|
||||
baseurl: https://packagerepo.service.consul/frr/el9/stable-daily/%{facts.os.architecture}/os
|
||||
gpgkey: https://packagerepo.service.consul/frr/el9/stable-daily/%{facts.os.architecture}/os/RPM-GPG-KEY-FRR
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
|
||||
@ -2,14 +2,15 @@
|
||||
hiera_include:
|
||||
- docker
|
||||
- profiles::gitea::runner
|
||||
- incus::client
|
||||
|
||||
docker::version: latest
|
||||
docker::curl_ensure: false
|
||||
docker::root_dir: /data/docker
|
||||
|
||||
profiles::gitea::runner::instance: https://git.unkin.net
|
||||
profiles::gitea::runner::home: /data/runner
|
||||
profiles::gitea::runner::version: '0.2.10'
|
||||
profiles::gitea::runner::source: "https://gitea.com/gitea/act_runner/releases/download/v%{hiera('profiles::gitea::runner::version')}/act_runner-%{hiera('profiles::gitea::runner::version')}-linux-amd64"
|
||||
profiles::gitea::runner::version: '0.2.12'
|
||||
profiles::gitea::runner::config:
|
||||
log:
|
||||
level: info
|
||||
@ -39,7 +40,8 @@ profiles::gitea::runner::config:
|
||||
privileged: false
|
||||
options:
|
||||
workdir_parent: /workspace
|
||||
valid_volumes: []
|
||||
valid_volumes:
|
||||
- /etc/pki/tls/vault
|
||||
docker_host: ""
|
||||
force_pull: true
|
||||
force_rebuild: false
|
||||
|
||||
@ -71,7 +71,7 @@ profiles::nginx::simpleproxy::nginx_aliases:
|
||||
|
||||
profiles::nginx::simpleproxy::proxy_port: 3000
|
||||
profiles::nginx::simpleproxy::proxy_path: '/'
|
||||
nginx::client_max_body_size: 1024M
|
||||
nginx::client_max_body_size: 5144M
|
||||
|
||||
# enable external access via haproxy
|
||||
profiles::gitea::haproxy::enable: true
|
||||
|
||||
@ -35,19 +35,9 @@ frrouting::daemons:
|
||||
# additional repos
|
||||
profiles::yum::global::repos:
|
||||
frr-extras:
|
||||
name: frr-extras
|
||||
descr: frr-extras repository
|
||||
target: /etc/yum.repos.d/frr-extras.repo
|
||||
baseurl: https://packagerepo.service.consul/frr/el9/extras-daily/%{facts.os.architecture}/os
|
||||
gpgkey: https://packagerepo.service.consul/frr/el9/extras-daily/%{facts.os.architecture}/os/RPM-GPG-KEY-FRR
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
frr-stable:
|
||||
name: frr-stable
|
||||
descr: frr-stable repository
|
||||
target: /etc/yum.repos.d/frr-stable.repo
|
||||
baseurl: https://packagerepo.service.consul/frr/el9/stable-daily/%{facts.os.architecture}/os
|
||||
gpgkey: https://packagerepo.service.consul/frr/el9/stable-daily/%{facts.os.architecture}/os/RPM-GPG-KEY-FRR
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
|
||||
# haproxy metrics
|
||||
consul::services:
|
||||
@ -163,6 +153,50 @@ profiles::haproxy::frontends:
|
||||
- 'set-header X-Forwarded-Proto https'
|
||||
- 'set-header X-Real-IP %[src]'
|
||||
- 'use-service prometheus-exporter if { path /metrics }'
|
||||
fe_imap:
|
||||
description: 'Frontend for Stalwart IMAP (STARTTLS)'
|
||||
bind:
|
||||
0.0.0.0:143: []
|
||||
mode: 'tcp'
|
||||
options:
|
||||
log: global
|
||||
default_backend: be_stalwart_imap
|
||||
tcp-request:
|
||||
- inspect-delay 5s
|
||||
- content accept if { req_len 0 }
|
||||
fe_imaps:
|
||||
description: 'Frontend for Stalwart IMAPS (implicit TLS)'
|
||||
bind:
|
||||
0.0.0.0:993: []
|
||||
mode: 'tcp'
|
||||
options:
|
||||
log: global
|
||||
default_backend: be_stalwart_imaps
|
||||
tcp-request:
|
||||
- inspect-delay 5s
|
||||
- content accept if { req_len 0 }
|
||||
fe_smtp:
|
||||
description: 'Frontend for Stalwart SMTP'
|
||||
bind:
|
||||
0.0.0.0:25: []
|
||||
mode: 'tcp'
|
||||
options:
|
||||
log: global
|
||||
default_backend: be_stalwart_smtp
|
||||
tcp-request:
|
||||
- inspect-delay 5s
|
||||
- content accept if { req_len 0 }
|
||||
fe_submission:
|
||||
description: 'Frontend for Stalwart SMTP Submission'
|
||||
bind:
|
||||
0.0.0.0:587: []
|
||||
mode: 'tcp'
|
||||
options:
|
||||
log: global
|
||||
default_backend: be_stalwart_submission
|
||||
tcp-request:
|
||||
- inspect-delay 5s
|
||||
- content accept if { req_len 0 }
|
||||
|
||||
profiles::haproxy::backends:
|
||||
be_letsencrypt:
|
||||
|
||||
@ -6,6 +6,7 @@ hiera_include:
|
||||
- zfs
|
||||
- profiles::ceph::node
|
||||
- profiles::ceph::client
|
||||
- profiles::ceph::dashboard
|
||||
- profiles::storage::cephfsvols
|
||||
- exporters::frr_exporter
|
||||
|
||||
@ -36,6 +37,7 @@ profiles::ssh::sign::principals:
|
||||
- "%{facts.networking.interfaces.enp3s0.ip}"
|
||||
|
||||
# configure consul service
|
||||
profiles::consul::client::host_addr: "%{hiera('networking_loopback0_ip')}"
|
||||
consul::services:
|
||||
incus:
|
||||
service_name: 'incus'
|
||||
@ -83,44 +85,22 @@ profiles::consul::client::node_rules:
|
||||
# additional repos
|
||||
profiles::yum::global::repos:
|
||||
ceph:
|
||||
name: ceph
|
||||
descr: ceph repository
|
||||
target: /etc/yum.repos.d/ceph.repo
|
||||
baseurl: https://edgecache.query.consul/ceph/yum/el%{facts.os.release.major}/%{facts.os.architecture}
|
||||
gpgkey: https://download.ceph.com/keys/release.asc
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
ceph-noarch:
|
||||
name: ceph-noarch
|
||||
descr: ceph-noarch repository
|
||||
target: /etc/yum.repos.d/ceph-noarch.repo
|
||||
baseurl: https://edgecache.query.consul/ceph/yum/el%{facts.os.release.major}/noarch
|
||||
gpgkey: https://download.ceph.com/keys/release.asc
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
frr-extras:
|
||||
name: frr-extras
|
||||
descr: frr-extras repository
|
||||
target: /etc/yum.repos.d/frr-extras.repo
|
||||
baseurl: https://packagerepo.service.consul/frr/el9/extras-daily/%{facts.os.architecture}/os
|
||||
gpgkey: https://packagerepo.service.consul/frr/el9/extras-daily/%{facts.os.architecture}/os/RPM-GPG-KEY-FRR
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
frr-stable:
|
||||
name: frr-stable
|
||||
descr: frr-stable repository
|
||||
target: /etc/yum.repos.d/frr-stable.repo
|
||||
baseurl: https://packagerepo.service.consul/frr/el9/stable-daily/%{facts.os.architecture}/os
|
||||
gpgkey: https://packagerepo.service.consul/frr/el9/stable-daily/%{facts.os.architecture}/os/RPM-GPG-KEY-FRR
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
zfs-kmod:
|
||||
name: zfs-kmod
|
||||
descr: zfs-kmod repository
|
||||
target: /etc/yum.repos.d/zfs-kmod.repo
|
||||
baseurl: https://packagerepo.service.consul/zfs/rhel9/kmod-daily/%{facts.os.architecture}/os
|
||||
gpgkey: https://packagerepo.service.consul/zfs/rhel9/kmod-daily/%{facts.os.architecture}/os/RPM-GPG-KEY-openzfs-2022
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
|
||||
# dns
|
||||
profiles::dns::base::primary_interface: loopback0
|
||||
|
||||
# dashboard/haproxy
|
||||
profiles::ceph::dashboard::ipaddress: "%{hiera('networking_loopback0_ip')}"
|
||||
|
||||
# networking
|
||||
systemd::manage_networkd: true
|
||||
systemd::manage_all_network_files: true
|
||||
@ -150,12 +130,11 @@ networking::interfaces:
|
||||
netmask: 255.255.255.255
|
||||
mtu: 1500
|
||||
|
||||
exporters::node_exporter::addr: "%{hiera('networking_loopback0_ip')}"
|
||||
|
||||
# frrouting
|
||||
exporters::frr_exporter::enable: true
|
||||
exporters::frr_exporter::addr: "%{hiera('networking_loopback0_ip')}"
|
||||
frrouting::ospfd_router_id: "%{hiera('networking_loopback0_ip')}"
|
||||
frrouting::ospf_preferred_source_enable: true
|
||||
frrouting::ospf_preferred_source: "%{hiera('networking_loopback0_ip')}"
|
||||
frrouting::ospfd_redistribute:
|
||||
- connected
|
||||
frrouting::ospfd_interfaces:
|
||||
|
||||
1
hieradata/roles/infra/k8s.eyaml
Normal file
1
hieradata/roles/infra/k8s.eyaml
Normal file
@ -0,0 +1 @@
|
||||
rke2::node_token: ENC[PKCS7,MIIB2gYJKoZIhvcNAQcDoIIByzCCAccCAQAxggEhMIIBHQIBADAFMAACAQEwDQYJKoZIhvcNAQEBBQAEggEAOD+w5nJFqEYWFj+tZQ65Oi19eDhaWtpLQ0gwEdBtMmY9sPJ63l1q2qH933NH6TOd1UlMDvGfoDLCae+yt/MAW//cJ15X3QbiVQ23DdfCOlUEZN6fjVrveEt/yIeFQmWvnkMS4pRfPbgQu2OHm37PpuPE7s6dUyGItAYjchrRhtQ7ibhXDnN7miG+oVRXP2T8b/V5WPdmA222DSV6r/AnqaWkna9W/oh/I2sNKeEm5q3f8bh8Gxt1dDy3VwaZ3lAh3uR+SUm7P/6PTYw8opxiumFBvos0mRiXIdOwUuqrAS8hafWBhxnDLlTBfz62Nc4wQmQ8gz0bHJZSipH9G6mIEDCBnAYJKoZIhvcNAQcBMB0GCWCGSAFlAwQBKgQQ2WG6ROWFlQdXx0TuO5oABoBwTDtAXIj7y6I1B3zCnFoMHETf5d7ulPGdgwZsENf0UIHpg2l0w503MUHHbu6YDFiDiTE0oDNJPVHid7TO+XwWFgh5v1MWi/XeEBgCs6nMCW8qkX0Z3UXaZdSBUll1M4sRtuqscBnoD/LLs2kKfxrqQg==]
|
||||
138
hieradata/roles/infra/k8s.yaml
Normal file
138
hieradata/roles/infra/k8s.yaml
Normal file
@ -0,0 +1,138 @@
|
||||
---
|
||||
hiera_include:
|
||||
- profiles::selinux::setenforce
|
||||
- profiles::ceph::node
|
||||
- profiles::ceph::client
|
||||
- exporters::frr_exporter
|
||||
- frrouting
|
||||
- rke2
|
||||
|
||||
# manage rke2
|
||||
rke2::bootstrap_node: prodnxsr0001.main.unkin.net
|
||||
rke2::join_url: https://join-k8s.service.consul:9345
|
||||
rke2::config_hash:
|
||||
bind-address: "%{hiera('networking_loopback0_ip')}"
|
||||
node-ip: "%{hiera('networking_loopback0_ip')}"
|
||||
node-external-ip: "%{hiera('networking_loopback0_ip')}"
|
||||
write-kubeconfig-mode: 644
|
||||
kubelet-arg:
|
||||
- '--node-status-update-frequency=4s'
|
||||
- '--max-pods=100'
|
||||
node-label:
|
||||
- "region=%{facts.region}"
|
||||
- "country=%{facts.country}"
|
||||
- "asset=%{facts.dmi.product.serial_number}"
|
||||
- "zone=%{zone}"
|
||||
|
||||
# FIXME: puppet-python wants to try manage python-dev, which is required by the ceph package
|
||||
python::manage_dev_package: false
|
||||
|
||||
profiles::packages::include:
|
||||
bridge-utils: {}
|
||||
cephadm: {}
|
||||
|
||||
profiles::selinux::setenforce::mode: disabled
|
||||
|
||||
profiles::ceph::client::manage_ceph_conf: false
|
||||
profiles::ceph::client::manage_ceph_package: false
|
||||
profiles::ceph::client::manage_ceph_paths: false
|
||||
profiles::ceph::client::fsid: 'de96a98f-3d23-465a-a899-86d3d67edab8'
|
||||
profiles::ceph::client::mons:
|
||||
- 198.18.23.9
|
||||
- 198.18.23.10
|
||||
- 198.18.23.11
|
||||
- 198.18.23.12
|
||||
- 198.18.23.13
|
||||
|
||||
# additional repos
|
||||
profiles::yum::global::repos:
|
||||
ceph:
|
||||
ensure: present
|
||||
ceph-noarch:
|
||||
ensure: present
|
||||
frr-extras:
|
||||
ensure: present
|
||||
frr-stable:
|
||||
ensure: present
|
||||
rancher-rke2-common-latest:
|
||||
ensure: present
|
||||
rancher-rke2-1-33-latest:
|
||||
ensure: present
|
||||
|
||||
# dns
|
||||
profiles::dns::base::primary_interface: loopback0
|
||||
|
||||
# networking
|
||||
systemd::manage_networkd: true
|
||||
systemd::manage_all_network_files: true
|
||||
networking::interfaces:
|
||||
"%{hiera('networking_1000_iface')}":
|
||||
type: physical
|
||||
ipaddress: "%{hiera('networking_1000_ip')}"
|
||||
gateway: 198.18.15.254
|
||||
txqueuelen: 10000
|
||||
forwarding: true
|
||||
"%{hiera('networking_2500_iface')}":
|
||||
type: physical
|
||||
ipaddress: "%{hiera('networking_2500_ip')}"
|
||||
mtu: 1500
|
||||
txqueuelen: 10000
|
||||
forwarding: true
|
||||
loopback0:
|
||||
type: dummy
|
||||
ipaddress: "%{hiera('networking_loopback0_ip')}"
|
||||
netmask: 255.255.255.255
|
||||
mtu: 1500
|
||||
loopback1:
|
||||
type: dummy
|
||||
ipaddress: "%{hiera('networking_loopback1_ip')}"
|
||||
netmask: 255.255.255.255
|
||||
mtu: 1500
|
||||
loopback2:
|
||||
type: dummy
|
||||
ipaddress: "%{hiera('networking_loopback2_ip')}"
|
||||
netmask: 255.255.255.255
|
||||
mtu: 1500
|
||||
|
||||
# configure consul service
|
||||
profiles::consul::client::host_addr: "%{hiera('networking_loopback0_ip')}"
|
||||
profiles::consul::client::node_rules:
|
||||
- resource: service
|
||||
segment: frr_exporter
|
||||
disposition: write
|
||||
|
||||
# frrouting
|
||||
exporters::frr_exporter::enable: true
|
||||
frrouting::ospfd_router_id: "%{hiera('networking_loopback0_ip')}"
|
||||
frrouting::ospf_preferred_source_enable: true
|
||||
frrouting::ospf_preferred_source: "%{hiera('networking_loopback0_ip')}"
|
||||
frrouting::ospfd_redistribute:
|
||||
- connected
|
||||
frrouting::ospfd_interfaces:
|
||||
"%{hiera('networking_1000_iface')}":
|
||||
area: 0.0.0.0
|
||||
"%{hiera('networking_2500_iface')}":
|
||||
area: 0.0.0.0
|
||||
loopback0:
|
||||
area: 0.0.0.0
|
||||
loopback1:
|
||||
area: 0.0.0.0
|
||||
loopback2:
|
||||
area: 0.0.0.0
|
||||
frrouting::daemons:
|
||||
ospfd: true
|
||||
frrouting::ospf_exclude_k8s_enable: true
|
||||
frrouting::k8s_cluster_cidr: '10.42.0.0/16' # RKE2 cluster-cidr (pods)
|
||||
frrouting::k8s_service_cidr: '10.43.0.0/16' # RKE2 service-cidr
|
||||
|
||||
# add loopback interfaces to ssh list
|
||||
ssh::server::options:
|
||||
ListenAddress:
|
||||
- "%{hiera('networking_loopback0_ip')}"
|
||||
- "%{hiera('networking_1000_ip')}"
|
||||
- "%{hiera('networking_2500_ip')}"
|
||||
|
||||
profiles::ssh::sign::principals:
|
||||
- "%{hiera('networking_loopback0_ip')}"
|
||||
- "%{hiera('networking_1000_ip')}"
|
||||
- "%{hiera('networking_2500_ip')}"
|
||||
@ -1,10 +1,3 @@
|
||||
---
|
||||
# networking
|
||||
systemd::manage_networkd: true
|
||||
systemd::manage_all_network_files: true
|
||||
networking::interfaces:
|
||||
eth0:
|
||||
type: physical
|
||||
forwarding: true
|
||||
dhcp: true
|
||||
mtu: 1500
|
||||
# manage rke2
|
||||
rke2::node_type: agent
|
||||
|
||||
1
hieradata/roles/infra/k8s/control.eyaml
Normal file
1
hieradata/roles/infra/k8s/control.eyaml
Normal file
@ -0,0 +1 @@
|
||||
---
|
||||
@ -1,42 +1,81 @@
|
||||
---
|
||||
profiles::pki::vault::alt_names:
|
||||
- k8s-control.service.consul
|
||||
- k8s-control.query.consul
|
||||
- "k8s-control.service.%{facts.country}-%{facts.region}.consul"
|
||||
|
||||
profiles::ssh::sign::principals:
|
||||
- k8s-control.service.consul
|
||||
- k8s-control.query.consul
|
||||
- "k8s-control.service.%{facts.country}-%{facts.region}.consul"
|
||||
# manage rke2
|
||||
rke2::node_type: server
|
||||
rke2::helm_install: true
|
||||
rke2::helm_repos:
|
||||
harbor: https://helm.goharbor.io
|
||||
traefik: https://traefik.github.io/charts
|
||||
hashicorp: https://helm.releases.hashicorp.com
|
||||
rke2::extra_config_files:
|
||||
- rke2-canal-config
|
||||
- rke2-nginx-ingress-config
|
||||
rke2::config_hash:
|
||||
advertise-address: "%{hiera('networking_loopback0_ip')}"
|
||||
tls-san:
|
||||
- "join-k8s.service.consul"
|
||||
- "api-k8s.service.consul"
|
||||
- "api.k8s.unkin.net"
|
||||
- "join.k8s.unkin.net"
|
||||
cni: canal
|
||||
cluster-cidr: 10.42.0.0/16
|
||||
service-cidr: 10.43.0.0/16
|
||||
cluster-dns: 10.43.0.10
|
||||
etcd-arg: "--quota-backend-bytes 2048000000"
|
||||
etcd-snapshot-schedule-cron: "0 3 * * *"
|
||||
etcd-snapshot-retention: 10
|
||||
kube-apiserver-arg:
|
||||
- '--default-not-ready-toleration-seconds=30'
|
||||
- '--default-unreachable-toleration-seconds=30'
|
||||
kube-controller-manager-arg:
|
||||
- '--node-monitor-period=4s'
|
||||
protect-kernel-defaults: true
|
||||
disable-kube-proxy: false
|
||||
|
||||
# configure consul service
|
||||
consul::services:
|
||||
k8s-control:
|
||||
service_name: 'k8s-control'
|
||||
tags:
|
||||
- 'k8s'
|
||||
- 'container'
|
||||
api-k8s:
|
||||
service_name: 'api-k8s'
|
||||
address: "%{facts.networking.fqdn}"
|
||||
port: 6443
|
||||
checks:
|
||||
- id: 'k8s-control_https_check'
|
||||
name: 'k8s-control HTTPS Check'
|
||||
http: "https://%{facts.networking.fqdn}:6443"
|
||||
method: 'GET'
|
||||
tls_skip_verify: true
|
||||
- id: 'api-k8s_livez_check'
|
||||
name: 'api-k8s livez Check'
|
||||
args:
|
||||
- sudo
|
||||
- /usr/local/bin/check_k8s_api.sh
|
||||
interval: '10s'
|
||||
timeout: '1s'
|
||||
join-k8s:
|
||||
service_name: 'join-k8s'
|
||||
address: "%{facts.networking.fqdn}"
|
||||
port: 9345
|
||||
checks:
|
||||
- id: 'rke2_tcp_check_9345'
|
||||
name: 'rke2 TCP Check 9345'
|
||||
tcp: "%{hiera('networking_loopback0_ip')}:9345"
|
||||
interval: '10s'
|
||||
timeout: '1s'
|
||||
- id: 'rke2_server_ping_check'
|
||||
name: 'rke2 Server Ping Check'
|
||||
http: "https://%{hiera('networking_loopback0_ip')}:9345/ping"
|
||||
interval: '10s'
|
||||
timeout: '3s'
|
||||
tls_skip_verify: true
|
||||
profiles::consul::client::node_rules:
|
||||
- resource: service
|
||||
segment: k8s-control
|
||||
segment: api-k8s
|
||||
disposition: write
|
||||
- resource: service
|
||||
segment: join-k8s
|
||||
disposition: write
|
||||
|
||||
# networking
|
||||
systemd::manage_networkd: true
|
||||
systemd::manage_all_network_files: true
|
||||
networking::interfaces:
|
||||
eth0:
|
||||
type: physical
|
||||
forwarding: true
|
||||
dhcp: true
|
||||
mtu: 1500
|
||||
profiles::pki::vault::alt_names:
|
||||
- api-k8s.service.consul
|
||||
- api-k8s.query.consul
|
||||
- "api-k8s.service.%{facts.country}-%{facts.region}.consul"
|
||||
|
||||
sudo::configs:
|
||||
consul-checks:
|
||||
priority: 20
|
||||
content: |
|
||||
consul ALL=(ALL) NOPASSWD: /usr/local/bin/check_k8s_api.sh
|
||||
|
||||
@ -5,6 +5,24 @@ hiera_include:
|
||||
- profiles::ceph::node
|
||||
- profiles::ceph::client
|
||||
- exporters::frr_exporter
|
||||
- profiles::rke2::node
|
||||
|
||||
# manage rke2
|
||||
profiles::rke2::node::servers:
|
||||
- prodnxsr0001.main.unkin.net
|
||||
- prodnxsr0002.main.unkin.net
|
||||
- prodnxsr0003.main.unkin.net
|
||||
|
||||
rke2::config_hash:
|
||||
bind-address: "%{hiera('networking_loopback0_ip')}"
|
||||
advertise-address: "%{hiera('networking_loopback0_ip')}"
|
||||
node-ip: "%{hiera('networking_loopback0_ip')}"
|
||||
node-external-ip: "%{hiera('networking_loopback0_ip')}"
|
||||
cluster-domain: "svc.k8s.unkin.net"
|
||||
tls-san:
|
||||
- "api.k8s.unkin.net"
|
||||
- "join.k8s.unkin.net"
|
||||
cni: cilium
|
||||
|
||||
# FIXME: puppet-python wants to try manage python-dev, which is required by the ceph package
|
||||
python::manage_dev_package: false
|
||||
@ -25,36 +43,21 @@ profiles::ceph::client::mons:
|
||||
- 198.18.23.11
|
||||
- 198.18.23.12
|
||||
- 198.18.23.13
|
||||
|
||||
# additional repos
|
||||
profiles::yum::global::repos:
|
||||
ceph:
|
||||
name: ceph
|
||||
descr: ceph repository
|
||||
target: /etc/yum.repos.d/ceph.repo
|
||||
baseurl: https://edgecache.query.consul/ceph/yum/el%{facts.os.release.major}/%{facts.os.architecture}
|
||||
gpgkey: https://download.ceph.com/keys/release.asc
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
ceph-noarch:
|
||||
name: ceph-noarch
|
||||
descr: ceph-noarch repository
|
||||
target: /etc/yum.repos.d/ceph-noarch.repo
|
||||
baseurl: https://edgecache.query.consul/ceph/yum/el%{facts.os.release.major}/noarch
|
||||
gpgkey: https://download.ceph.com/keys/release.asc
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
frr-extras:
|
||||
name: frr-extras
|
||||
descr: frr-extras repository
|
||||
target: /etc/yum.repos.d/frr-extras.repo
|
||||
baseurl: https://packagerepo.service.consul/frr/el9/extras-daily/%{facts.os.architecture}/os
|
||||
gpgkey: https://packagerepo.service.consul/frr/el9/extras-daily/%{facts.os.architecture}/os/RPM-GPG-KEY-FRR
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
frr-stable:
|
||||
name: frr-stable
|
||||
descr: frr-stable repository
|
||||
target: /etc/yum.repos.d/frr-stable.repo
|
||||
baseurl: https://packagerepo.service.consul/frr/el9/stable-daily/%{facts.os.architecture}/os
|
||||
gpgkey: https://packagerepo.service.consul/frr/el9/stable-daily/%{facts.os.architecture}/os/RPM-GPG-KEY-FRR
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
rancher-rke2-common-latest:
|
||||
ensure: present
|
||||
rancher-rke2-1-33-latest:
|
||||
ensure: present
|
||||
|
||||
# dns
|
||||
profiles::dns::base::primary_interface: loopback0
|
||||
@ -91,18 +94,47 @@ networking::interfaces:
|
||||
netmask: 255.255.255.255
|
||||
mtu: 1500
|
||||
|
||||
# consul
|
||||
# configure consul service
|
||||
profiles::consul::client::host_addr: "%{hiera('networking_loopback0_ip')}"
|
||||
consul::services:
|
||||
api-k8s:
|
||||
service_name: 'api-k8s'
|
||||
address: "%{facts.networking.fqdn}"
|
||||
port: 6443
|
||||
checks:
|
||||
- id: 'api-k8s_https_check'
|
||||
name: 'api-k8s HTTPS Check'
|
||||
http: "https://%{facts.networking.fqdn}:6443"
|
||||
method: 'GET'
|
||||
tls_skip_verify: true
|
||||
interval: '10s'
|
||||
timeout: '1s'
|
||||
join-k8s:
|
||||
service_name: 'join-k8s'
|
||||
address: "%{facts.networking.fqdn}"
|
||||
port: 9345
|
||||
checks:
|
||||
- id: 'etcd_tcp_check_9345'
|
||||
name: 'ETCD TCP Check 9345'
|
||||
tcp: "%{facts.networking.fqdn}:9345"
|
||||
interval: '10s'
|
||||
timeout: '1s'
|
||||
profiles::consul::client::node_rules:
|
||||
- resource: service
|
||||
segment: api-k8s
|
||||
disposition: write
|
||||
- resource: service
|
||||
segment: join-k8s
|
||||
disposition: write
|
||||
- resource: service
|
||||
segment: frr_exporter
|
||||
disposition: write
|
||||
|
||||
exporters::node_exporter::addr: "%{hiera('networking_loopback0_ip')}"
|
||||
|
||||
# frrouting
|
||||
exporters::frr_exporter::enable: true
|
||||
exporters::frr_exporter::addr: "%{hiera('networking_loopback0_ip')}"
|
||||
frrouting::ospfd_router_id: "%{hiera('networking_loopback0_ip')}"
|
||||
frrouting::ospf_preferred_source_enable: true
|
||||
frrouting::ospf_preferred_source: "%{hiera('networking_loopback0_ip')}"
|
||||
frrouting::ospfd_redistribute:
|
||||
- connected
|
||||
frrouting::ospfd_interfaces:
|
||||
@ -130,3 +162,8 @@ profiles::ssh::sign::principals:
|
||||
- "%{hiera('networking_loopback0_ip')}"
|
||||
- "%{hiera('networking_1000_ip')}"
|
||||
- "%{hiera('networking_2500_ip')}"
|
||||
|
||||
profiles::pki::vault::alt_names:
|
||||
- api-k8s.service.consul
|
||||
- api-k8s.query.consul
|
||||
- "api-k8s.service.%{facts.country}-%{facts.region}.consul"
|
||||
|
||||
@ -14,6 +14,8 @@ victorialogs::node::options:
|
||||
envflag.enable: 'true'
|
||||
select.disable: 'undef'
|
||||
storageNode.tls: 'undef'
|
||||
syslog.listenAddr.tcp: ':21514'
|
||||
syslog.timezone: 'Australia/Sydney'
|
||||
storageNode:
|
||||
- ausyd1nxvm2108.main.unkin.net:9428
|
||||
- ausyd1nxvm2109.main.unkin.net:9428
|
||||
@ -45,7 +47,20 @@ consul::services:
|
||||
tls_skip_verify: true
|
||||
interval: '10s'
|
||||
timeout: '1s'
|
||||
syslog:
|
||||
service_name: 'syslog'
|
||||
address: "%{facts.networking.ip}"
|
||||
port: 21514
|
||||
checks:
|
||||
- id: 'vlinsert_syslog_tcp_check'
|
||||
name: 'VictoriaLogs Syslog TCP Check'
|
||||
tcp: "%{facts.networking.fqdn}:21514"
|
||||
interval: '30s'
|
||||
timeout: '5s'
|
||||
profiles::consul::client::node_rules:
|
||||
- resource: service
|
||||
segment: vlinsert
|
||||
disposition: write
|
||||
- resource: service
|
||||
segment: syslog
|
||||
disposition: write
|
||||
|
||||
5
hieradata/roles/infra/mail/backend.eyaml
Normal file
5
hieradata/roles/infra/mail/backend.eyaml
Normal file
@ -0,0 +1,5 @@
|
||||
---
|
||||
profiles::sql::postgresdb::dbpass: ENC[PKCS7,MIIBmQYJKoZIhvcNAQcDoIIBijCCAYYCAQAxggEhMIIBHQIBADAFMAACAQEwDQYJKoZIhvcNAQEBBQAEggEAEZkKX2ThGom2PffofEuRBHbyiq68PCsq0+19eSa02fpVPKgZ/5BEjzBhwvrt0BWZWsjYGhccFQ69DR+lTuqS50GcRSAiNQ2LDX2a3J1pu39oIKsNVmcJTza0f5T0VeI3A7sZkn7jL+NVz5ANp8V0EMfAjaduGQ7Jac+8dBsvTrLbJ+1AZVrjaKPxOI1+5tpE7qx35mM0oDVy0NwmlaVf8vbK6jyzyUJRs4Sb+mpioPi5sDxHgClzsQnu93HqqAIqR5UzsUv7MDMljOGYUF5ITyPU836I1LEZ9UfiVO7AikQ3A31LaSUWvwsxRHKxiQXJ7v6W/O+Nt3jdIR0eoqC5xTBcBgkqhkiG9w0BBwEwHQYJYIZIAWUDBAEqBBC11+SUDLmz6bBrGyfYT9DPgDB3UhuhJ3kUruMTdRCW0Y6hoSBBQYCO+ZRFJToGTkz/BcxVw2Xtwjc7UmKmLodsDAo=]
|
||||
stalwart::s3_access_key: ENC[PKCS7,MIIBiQYJKoZIhvcNAQcDoIIBejCCAXYCAQAxggEhMIIBHQIBADAFMAACAQEwDQYJKoZIhvcNAQEBBQAEggEAm1fkuVcAt4UKFnpljH0o3eb3x1r2ieEU4NUzuSx8DDdOv4rh8rkqghPGn1xs6LGd9zLprc9ZWFowXCP4I2L0gZ4PekI04rV4lL02xYechIah1JAntdwECEks7rmI4BbPabIUIkaHW4i/WtntRNvv38g9JjiWrvOoJSABEsiIVZL0ct6NykLgQk27r4rcP8j7ukNQqPCAA02d4y9CB/5g6RKtYkv6FwZfLA/rFTUIXhKJsNtTa9Gm1yhvb/Y859X4qvsSFymCm/B8+2Bz5H57lE0r6xOBBBS5PfpeQ2YRxmedlkj0HxRFpl2e1W94OazMbbI6dlXa5ceHqHXDZL78EzBMBgkqhkiG9w0BBwEwHQYJYIZIAWUDBAEqBBDhiBSQPgE+fwD1Y9Gc1wQ3gCCGeGdA4BKCvIhYaQmzv9wUFuNaO75qHrub0vY8Od1ilQ==]
|
||||
stalwart::s3_secret_key: ENC[PKCS7,MIIBmQYJKoZIhvcNAQcDoIIBijCCAYYCAQAxggEhMIIBHQIBADAFMAACAQEwDQYJKoZIhvcNAQEBBQAEggEAdFOGR2axfPIksXExnW8vAZBSBm8V/QqPtIMtuzleDLOGsOuyLgJAbHPZJoj/w5bBV9bBKMGfWs94GGbKZ2PafiJeCtqfFJnUAIpVBTZ5cfUhQXERAHcketMvKUi1J8jKLSFc9I3uTjIVcIasjk0oe7WmCPnoikjl1qJZ/lVDH0cXHevjMuohxEyyka5jzC0ixCOkxyqOV2LOqSc6J5d0WSsSWqw0lDmY7vJAtqNtH6y6ghKZo5zdLQOsF2Bseg3oVejRNqYwfUsaDnfiRwJS/Rm/TmtXb4i6Jn8coYplDaJHtxQSXJ+KvOy6LS7M8X5sQ/UZSRqcwM647Yg4pwVWqDBcBgkqhkiG9w0BBwEwHQYJYIZIAWUDBAEqBBBKBF7hTs0e0FKdRpFYbx4kgDB0xkNe78n6VLbUTHX7vTw6J+K/mxf+XCV95/EIkvBbWBYuU8ZMHNQKMEExmci4C4o=]
|
||||
stalwart::fallback_admin_password: ENC[PKCS7,MIIBmQYJKoZIhvcNAQcDoIIBijCCAYYCAQAxggEhMIIBHQIBADAFMAACAQEwDQYJKoZIhvcNAQEBBQAEggEAMp9wmIhRwj5kxfUcvc+/q/oUs/vBhSqP19ZfErM4vLDK20VOBTnPhSP2lfVh9pqO0c2hpWFeuqBWMynghO+HUBJfAn29Vrc8a9iSBxQ3XuF/uiRq1inOKCQpdsU18TyCrYV9AJFNf9U20JuUoav79m7EKLHS07PHAZ0osqIYy93eXdCFhwXAGHijp4wMMQz/5z1F1mZoSrc1cXe3y8iBeAvvjnRfpw14gOKZBjmEGUbo7AIyc3wax5hbOQYf/v+Hd90JarvAufxGytg9WKO20cChWYbmYDnIkytVt3vHdHf4RT8M635l6qwLr/70O1MdE7bkrVRKP8M3KLyH072pJTBcBgkqhkiG9w0BBwEwHQYJYIZIAWUDBAEqBBDSJwptBDvPd0WpxiIovZsjgDBBwesNW+UNo4b0idhyqsyWL2rtO7wLStWHgUIvRFJACCrTKKqlu7sta6mhu/ZsnF0=]
|
||||
46
hieradata/roles/infra/mail/backend.yaml
Normal file
46
hieradata/roles/infra/mail/backend.yaml
Normal file
@ -0,0 +1,46 @@
|
||||
---
|
||||
hiera_include:
|
||||
- stalwart
|
||||
- profiles::sql::postgresdb
|
||||
- profiles::stalwart::haproxy
|
||||
|
||||
# additional altnames
|
||||
profiles::pki::vault::alt_names:
|
||||
- mail.main.unkin.net
|
||||
- mail-webadmin.main.unkin.net
|
||||
- main-in.main.unkin.net
|
||||
- autoconfig.main.unkin.net
|
||||
- autodiscovery.main.unkin.net
|
||||
|
||||
# manage a pgsql database + user
|
||||
profiles::sql::postgresdb::cluster_name: "patroni-shared-%{facts.environment}"
|
||||
profiles::sql::postgresdb::dbname: stalwart
|
||||
profiles::sql::postgresdb::dbuser: stalwart
|
||||
|
||||
# export backends to haproxy
|
||||
profiles::stalwart::haproxy::enable: true
|
||||
|
||||
# Cluster role for node discovery
|
||||
stalwart::cluster_role: "%{facts.enc_role}"
|
||||
|
||||
# PostgreSQL connection
|
||||
stalwart::postgresql_host: "master.%{hiera('profiles::sql::postgresdb::cluster_name')}.service.%{facts.country}-%{facts.region}.consul"
|
||||
stalwart::postgresql_database: "%{hiera('profiles::sql::postgresdb::dbname')}"
|
||||
stalwart::postgresql_user: "%{hiera('profiles::sql::postgresdb::dbuser')}"
|
||||
stalwart::postgresql_password: "%{hiera('profiles::sql::postgresdb::dbpass')}"
|
||||
|
||||
# S3/Ceph-RGW connection
|
||||
stalwart::s3_endpoint: 'https://radosgw.service.consul'
|
||||
stalwart::s3_bucket: 'stalwart-maildata'
|
||||
stalwart::s3_region: "%{facts.region}"
|
||||
|
||||
# Domains and relay
|
||||
stalwart::domains:
|
||||
- 'mail.unkin.net'
|
||||
stalwart::postfix_relay_host: 'out-mta.main.unkin.net'
|
||||
stalwart::service_hostname: 'mail.main.unkin.net'
|
||||
stalwart::manage_dns_records: false
|
||||
|
||||
## With load balancer:
|
||||
#stalwart::manage_dns_records: true
|
||||
#stalwart::loadbalancer_host: 'mail-lb.example.com'
|
||||
52
hieradata/roles/infra/mail/gateway.yaml
Normal file
52
hieradata/roles/infra/mail/gateway.yaml
Normal file
@ -0,0 +1,52 @@
|
||||
---
|
||||
|
||||
# additional altnames
|
||||
profiles::pki::vault::alt_names:
|
||||
- in-mta.main.unkin.net
|
||||
|
||||
# base postfix configuration (passed to postfix class)
|
||||
postfix::relayhost: 'direct'
|
||||
postfix::myorigin: 'main.unkin.net'
|
||||
postfix::mydestination: 'blank'
|
||||
postfix::mynetworks: '127.0.0.0/8 [::1]/128'
|
||||
postfix::mta: true
|
||||
postfix::manage_aliases: true
|
||||
|
||||
# profile parameters for customization
|
||||
profiles::postfix::gateway::myhostname: 'in-mta.main.unkin.net'
|
||||
|
||||
# postfix map content (templates)
|
||||
profiles::postfix::gateway::relay_recipients_maps:
|
||||
'@main.unkin.net': 'OK'
|
||||
|
||||
profiles::postfix::gateway::relay_domains_maps:
|
||||
'main.unkin.net': 'OK'
|
||||
|
||||
profiles::postfix::gateway::postscreen_access_maps:
|
||||
'127.0.0.1/32': 'permit'
|
||||
'10.10.12.200/32': 'permit'
|
||||
|
||||
profiles::postfix::gateway::helo_access_maps:
|
||||
'.dynamic.': 'REJECT'
|
||||
'.dialup.': 'REJECT'
|
||||
'unknown': 'REJECT'
|
||||
'localhost': 'REJECT You are not localhost'
|
||||
|
||||
# postfix transports
|
||||
postfix::transports:
|
||||
'main.unkin.net':
|
||||
ensure: present
|
||||
destination: 'relay'
|
||||
nexthop: 'mail-in.main.unkin.net:25'
|
||||
|
||||
# postfix virtuals
|
||||
postfix::virtuals:
|
||||
'root':
|
||||
ensure: present
|
||||
destination: 'ben@main.unkin.net'
|
||||
'postmaster':
|
||||
ensure: present
|
||||
destination: 'ben@main.unkin.net'
|
||||
'abuse':
|
||||
ensure: present
|
||||
destination: 'ben@main.unkin.net'
|
||||
@ -11,6 +11,9 @@ profiles::metrics::grafana::db_name: "%{hiera('profiles::sql::postgresdb::dbname
|
||||
profiles::metrics::grafana::db_user: "%{hiera('profiles::sql::postgresdb::dbuser')}"
|
||||
profiles::metrics::grafana::db_pass: "%{hiera('profiles::sql::postgresdb::dbpass')}"
|
||||
profiles::metrics::grafana::pgsql_backend: true
|
||||
profiles::metrics::grafana::plugins:
|
||||
victoriametrics-logs-datasource:
|
||||
ensure: present
|
||||
|
||||
# additional altnames
|
||||
profiles::pki::vault::alt_names:
|
||||
|
||||
@ -3,6 +3,16 @@ hiera_include:
|
||||
- vmcluster::vmagent
|
||||
|
||||
vmcluster::vmagent::enable: true
|
||||
vmcluster::vmagent::static_targets:
|
||||
vyos_node:
|
||||
targets:
|
||||
- '198.18.21.160:9100'
|
||||
scrape_interval: '15s'
|
||||
metrics_path: '/metrics'
|
||||
scheme: 'http'
|
||||
labels:
|
||||
instance: 'syrtvm0001.main.unkin.net'
|
||||
job: 'vyos_node'
|
||||
vmcluster::vmagent::options:
|
||||
tls: 'true'
|
||||
tlsCertFile: '/etc/pki/tls/vault/certificate.crt'
|
||||
|
||||
@ -24,13 +24,8 @@ frrouting::ospfd_interfaces:
|
||||
area: 0.0.0.1
|
||||
|
||||
profiles::yum::global::repos:
|
||||
ceph-reef:
|
||||
name: ceph-reef
|
||||
descr: ceph reef repository
|
||||
target: /etc/yum.repos.d/ceph-reef.repo
|
||||
baseurl: https://edgecache.query.consul/ceph/yum/el%{facts.os.release.major}/%{facts.os.architecture}
|
||||
gpgcheck: 0,
|
||||
mirrorlist: absent
|
||||
ceph:
|
||||
ensure: present
|
||||
|
||||
profiles::ceph::client::keyrings:
|
||||
nomad:
|
||||
|
||||
@ -1,50 +1,18 @@
|
||||
---
|
||||
profiles::yum::global::repos:
|
||||
centos_8_advanced_virtualization:
|
||||
name: 'virt-advanced-virtualization'
|
||||
descr: 'CentOS Advanced Virtualization'
|
||||
target: /etc/yum.repos.d/ovirt.repo
|
||||
baseurl: https://edgecache.query.consul/centos/8-stream/virt/x86_64/advancedvirt-common
|
||||
gpgkey: 'https://www.centos.org/keys/RPM-GPG-KEY-CentOS-SIG-Virtualization'
|
||||
ensure: present
|
||||
centos_8_ceph_pacific:
|
||||
name: 'storage-ceph-pacific'
|
||||
descr: 'CentOS Ceph Pacific'
|
||||
target: /etc/yum.repos.d/ovirt.repo
|
||||
baseurl: https://edgecache.query.consul/centos/8-stream/storage/x86_64/ceph-pacific
|
||||
gpgkey: 'https://www.centos.org/keys/RPM-GPG-KEY-CentOS-SIG-Storage'
|
||||
ensure: present
|
||||
centos_8_rabbitmq_38:
|
||||
name: 'messaging-rabbitmq-38'
|
||||
descr: 'CentOS RabbitMQ 38'
|
||||
target: /etc/yum.repos.d/ovirt.repo
|
||||
baseurl: https://edgecache.query.consul/centos/8-stream/messaging/x86_64/rabbitmq-38
|
||||
gpgkey: 'https://www.centos.org/keys/RPM-GPG-KEY-CentOS-SIG-Messaging'
|
||||
ensure: present
|
||||
centos_8_nfv_openvswitch:
|
||||
name: 'nfv-openvswitch-2'
|
||||
descr: 'CentOS NFV OpenvSwitch'
|
||||
target: /etc/yum.repos.d/ovirt.repo
|
||||
baseurl: https://edgecache.query.consul/centos/8-stream/nfv/x86_64/openvswitch-2
|
||||
gpgkey: 'https://www.centos.org/keys/RPM-GPG-KEY-CentOS-SIG-NFV'
|
||||
ensure: present
|
||||
centos_8_openstack_xena:
|
||||
name: 'cloud-openstack-xena'
|
||||
descr: 'CentOS OpenStack Xena'
|
||||
target: /etc/yum.repos.d/ovirt.repo
|
||||
baseurl: https://edgecache.query.consul/centos/8-stream/cloud/x86_64/openstack-xena
|
||||
gpgkey: 'https://www.centos.org/keys/RPM-GPG-KEY-CentOS-SIG-Cloud'
|
||||
ensure: present
|
||||
centos_8_opstools:
|
||||
name: 'opstools-collectd-5'
|
||||
descr: 'CentOS OpsTools - collectd'
|
||||
target: /etc/yum.repos.d/ovirt.repo
|
||||
baseurl: https://edgecache.query.consul/centos/8-stream/opstools/x86_64/collectd-5
|
||||
gpgkey: 'https://www.centos.org/keys/RPM-GPG-KEY-CentOS-SIG-OpsTools'
|
||||
ensure: present
|
||||
centos_8_ovirt45:
|
||||
name: 'virt-ovirt-45'
|
||||
descr: 'CentOS oVirt 4.5'
|
||||
target: /etc/yum.repos.d/ovirt.repo
|
||||
baseurl: https://edgecache.query.consul/centos/8-stream/virt/x86_64/ovirt-45
|
||||
gpgkey: 'https://www.centos.org/keys/RPM-GPG-KEY-CentOS-SIG-Virtualization'
|
||||
ensure: present
|
||||
centos_8_stream_gluster10:
|
||||
name: 'storage-gluster-10'
|
||||
descr: 'CentOS oVirt 4.5 - Glusterfs 10'
|
||||
target: /etc/yum.repos.d/ovirt.repo
|
||||
baseurl: https://edgecache.query.consul/centos/8-stream/storage/x86_64/gluster-10
|
||||
gpgkey: 'https://www.centos.org/keys/RPM-GPG-KEY-CentOS-SIG-Storage'
|
||||
ensure: present
|
||||
|
||||
@ -9,50 +9,18 @@ sudo::purge_ignore:
|
||||
|
||||
profiles::yum::global::repos:
|
||||
centos_8_advanced_virtualization:
|
||||
name: 'virt-advanced-virtualization'
|
||||
descr: 'CentOS Advanced Virtualization'
|
||||
target: /etc/yum.repos.d/ovirt.repo
|
||||
baseurl: https://edgecache.query.consul/centos/8-stream/virt/x86_64/advancedvirt-common
|
||||
gpgkey: 'https://www.centos.org/keys/RPM-GPG-KEY-CentOS-SIG-Virtualization'
|
||||
ensure: present
|
||||
centos_8_ceph_pacific:
|
||||
name: 'storage-ceph-pacific'
|
||||
descr: 'CentOS Ceph Pacific'
|
||||
target: /etc/yum.repos.d/ovirt.repo
|
||||
baseurl: https://edgecache.query.consul/centos/8-stream/storage/x86_64/ceph-pacific
|
||||
gpgkey: 'https://www.centos.org/keys/RPM-GPG-KEY-CentOS-SIG-Storage'
|
||||
ensure: present
|
||||
centos_8_rabbitmq_38:
|
||||
name: 'messaging-rabbitmq-38'
|
||||
descr: 'CentOS RabbitMQ 38'
|
||||
target: /etc/yum.repos.d/ovirt.repo
|
||||
baseurl: https://edgecache.query.consul/centos/8-stream/messaging/x86_64/rabbitmq-38
|
||||
gpgkey: 'https://www.centos.org/keys/RPM-GPG-KEY-CentOS-SIG-Messaging'
|
||||
ensure: present
|
||||
centos_8_nfv_openvswitch:
|
||||
name: 'nfv-openvswitch-2'
|
||||
descr: 'CentOS NFV OpenvSwitch'
|
||||
target: /etc/yum.repos.d/ovirt.repo
|
||||
baseurl: https://edgecache.query.consul/centos/8-stream/nfv/x86_64/openvswitch-2
|
||||
gpgkey: 'https://www.centos.org/keys/RPM-GPG-KEY-CentOS-SIG-NFV'
|
||||
ensure: present
|
||||
centos_8_openstack_xena:
|
||||
name: 'cloud-openstack-xena'
|
||||
descr: 'CentOS OpenStack Xena'
|
||||
target: /etc/yum.repos.d/ovirt.repo
|
||||
baseurl: https://edgecache.query.consul/centos/8-stream/cloud/x86_64/openstack-xena
|
||||
gpgkey: 'https://www.centos.org/keys/RPM-GPG-KEY-CentOS-SIG-Cloud'
|
||||
ensure: present
|
||||
centos_8_opstools:
|
||||
name: 'opstools-collectd-5'
|
||||
descr: 'CentOS OpsTools - collectd'
|
||||
target: /etc/yum.repos.d/ovirt.repo
|
||||
baseurl: https://edgecache.query.consul/centos/8-stream/opstools/x86_64/collectd-5
|
||||
gpgkey: 'https://www.centos.org/keys/RPM-GPG-KEY-CentOS-SIG-OpsTools'
|
||||
ensure: present
|
||||
centos_8_ovirt45:
|
||||
name: 'virt-ovirt-45'
|
||||
descr: 'CentOS oVirt 4.5'
|
||||
target: /etc/yum.repos.d/ovirt.repo
|
||||
baseurl: https://edgecache.query.consul/centos/8-stream/virt/x86_64/ovirt-45
|
||||
gpgkey: 'https://www.centos.org/keys/RPM-GPG-KEY-CentOS-SIG-Virtualization'
|
||||
ensure: present
|
||||
centos_8_stream_gluster10:
|
||||
name: 'storage-gluster-10'
|
||||
descr: 'CentOS oVirt 4.5 - Glusterfs 10'
|
||||
target: /etc/yum.repos.d/ovirt.repo
|
||||
baseurl: https://edgecache.query.consul/centos/8-stream/storage/x86_64/gluster-10
|
||||
gpgkey: 'https://www.centos.org/keys/RPM-GPG-KEY-CentOS-SIG-Storage'
|
||||
ensure: present
|
||||
|
||||
@ -15,3 +15,4 @@ certbot::domains:
|
||||
- fafflix.unkin.net
|
||||
- git.unkin.net
|
||||
- grafana.unkin.net
|
||||
- dashboard.ceph.unkin.net
|
||||
|
||||
@ -1,3 +1,3 @@
|
||||
---
|
||||
profiles::packages::include:
|
||||
puppetserver: {}
|
||||
openvox-server: {}
|
||||
|
||||
@ -62,14 +62,6 @@ profiles::consul::client::node_rules:
|
||||
|
||||
profiles::yum::global::repos:
|
||||
postgresql-17:
|
||||
name: postgresql-17
|
||||
descr: postgresql-17 repository
|
||||
target: /etc/yum.repos.d/postgresql.repo
|
||||
baseurl: https://packagerepo.service.consul/postgresql/rhel%{facts.os.release.major}/17-daily/%{facts.os.architecture}/os
|
||||
gpgkey: https://packagerepo.service.consul/postgresql/rhel%{facts.os.release.major}/17-daily/%{facts.os.architecture}/os/PGDG-RPM-GPG-KEY-RHEL
|
||||
ensure: present
|
||||
postgresql-common:
|
||||
name: postgresql-common
|
||||
descr: postgresql-common repository
|
||||
target: /etc/yum.repos.d/postgresql.repo
|
||||
baseurl: https://packagerepo.service.consul/postgresql/rhel%{facts.os.release.major}/common-daily/%{facts.os.architecture}/os
|
||||
gpgkey: https://packagerepo.service.consul/postgresql/rhel%{facts.os.release.major}/common-daily/%{facts.os.architecture}/os/PGDG-RPM-GPG-KEY-RHEL
|
||||
ensure: present
|
||||
|
||||
@ -3,90 +3,6 @@ profiles::packages::include:
|
||||
createrepo: {}
|
||||
|
||||
profiles::reposync::repos_list:
|
||||
almalinux_9.6_baseos:
|
||||
repository: 'baseos'
|
||||
description: 'AlmaLinux 9.6 BaseOS'
|
||||
osname: 'almalinux'
|
||||
release: '9.6'
|
||||
mirrorlist: 'https://mirrors.almalinux.org/mirrorlist/9.6/baseos'
|
||||
gpgkey: 'http://mirror.aarnet.edu.au/pub/almalinux/RPM-GPG-KEY-AlmaLinux-9'
|
||||
almalinux_9.6_appstream:
|
||||
repository: 'appstream'
|
||||
description: 'AlmaLinux 9.6 AppStream'
|
||||
osname: 'almalinux'
|
||||
release: '9.6'
|
||||
mirrorlist: 'https://mirrors.almalinux.org/mirrorlist/9.6/appstream'
|
||||
gpgkey: 'http://mirror.aarnet.edu.au/pub/almalinux/RPM-GPG-KEY-AlmaLinux-9'
|
||||
almalinux_9.6_crb:
|
||||
repository: 'crb'
|
||||
description: 'AlmaLinux 9.6 CRB'
|
||||
osname: 'almalinux'
|
||||
release: '9.6'
|
||||
mirrorlist: 'https://mirrors.almalinux.org/mirrorlist/9.6/crb'
|
||||
gpgkey: 'http://mirror.aarnet.edu.au/pub/almalinux/RPM-GPG-KEY-AlmaLinux-9'
|
||||
almalinux_9.6_ha:
|
||||
repository: 'ha'
|
||||
description: 'AlmaLinux 9.6 HighAvailability'
|
||||
osname: 'almalinux'
|
||||
release: '9.6'
|
||||
mirrorlist: 'https://mirrors.almalinux.org/mirrorlist/9.6/highavailability'
|
||||
gpgkey: 'http://mirror.aarnet.edu.au/pub/almalinux/RPM-GPG-KEY-AlmaLinux-9'
|
||||
almalinux_9.6_extras:
|
||||
repository: 'extras'
|
||||
description: 'AlmaLinux 9.6 extras'
|
||||
osname: 'almalinux'
|
||||
release: '9.6'
|
||||
mirrorlist: 'https://mirrors.almalinux.org/mirrorlist/9.6/extras'
|
||||
gpgkey: 'http://mirror.aarnet.edu.au/pub/almalinux/RPM-GPG-KEY-AlmaLinux-9'
|
||||
almalinux_9_5_baseos:
|
||||
repository: 'baseos'
|
||||
description: 'AlmaLinux 9.5 BaseOS'
|
||||
osname: 'almalinux'
|
||||
release: '9.5'
|
||||
mirrorlist: 'https://mirrors.almalinux.org/mirrorlist/9.5/baseos'
|
||||
gpgkey: 'http://mirror.aarnet.edu.au/pub/almalinux/RPM-GPG-KEY-AlmaLinux-9'
|
||||
almalinux_9_5_appstream:
|
||||
repository: 'appstream'
|
||||
description: 'AlmaLinux 9.5 AppStream'
|
||||
osname: 'almalinux'
|
||||
release: '9.5'
|
||||
mirrorlist: 'https://mirrors.almalinux.org/mirrorlist/9.5/appstream'
|
||||
gpgkey: 'http://mirror.aarnet.edu.au/pub/almalinux/RPM-GPG-KEY-AlmaLinux-9'
|
||||
almalinux_9_5_crb:
|
||||
repository: 'crb'
|
||||
description: 'AlmaLinux 9.5 CRB'
|
||||
osname: 'almalinux'
|
||||
release: '9.5'
|
||||
mirrorlist: 'https://mirrors.almalinux.org/mirrorlist/9.5/crb'
|
||||
gpgkey: 'http://mirror.aarnet.edu.au/pub/almalinux/RPM-GPG-KEY-AlmaLinux-9'
|
||||
almalinux_9_5_ha:
|
||||
repository: 'ha'
|
||||
description: 'AlmaLinux 9.5 HighAvailability'
|
||||
osname: 'almalinux'
|
||||
release: '9.5'
|
||||
mirrorlist: 'https://mirrors.almalinux.org/mirrorlist/9.5/highavailability'
|
||||
gpgkey: 'http://mirror.aarnet.edu.au/pub/almalinux/RPM-GPG-KEY-AlmaLinux-9'
|
||||
almalinux_9_5_extras:
|
||||
repository: 'extras'
|
||||
description: 'AlmaLinux 9.5 extras'
|
||||
osname: 'almalinux'
|
||||
release: '9.5'
|
||||
mirrorlist: 'https://mirrors.almalinux.org/mirrorlist/9.5/extras'
|
||||
gpgkey: 'http://mirror.aarnet.edu.au/pub/almalinux/RPM-GPG-KEY-AlmaLinux-9'
|
||||
epel_8:
|
||||
repository: 'everything'
|
||||
description: 'EPEL8'
|
||||
osname: 'epel'
|
||||
release: '8'
|
||||
mirrorlist: 'https://mirrors.fedoraproject.org/mirrorlist?repo=epel-8&arch=x86_64'
|
||||
gpgkey: 'https://epel.mirror.digitalpacific.com.au/RPM-GPG-KEY-EPEL-8'
|
||||
epel_9:
|
||||
repository: 'everything'
|
||||
description: 'EPEL9'
|
||||
osname: 'epel'
|
||||
release: '9'
|
||||
mirrorlist: 'https://mirrors.fedoraproject.org/mirrorlist?repo=epel-9&arch=x86_64'
|
||||
gpgkey: 'https://epel.mirror.digitalpacific.com.au/RPM-GPG-KEY-EPEL-9'
|
||||
docker_stable_el8:
|
||||
repository: 'stable'
|
||||
description: 'Docker CE Stable EL8'
|
||||
@ -101,34 +17,6 @@ profiles::reposync::repos_list:
|
||||
release: 'el9'
|
||||
baseurl: 'https://download.docker.com/linux/centos/9/x86_64/stable/'
|
||||
gpgkey: 'https://download.docker.com/linux/centos/gpg'
|
||||
frr_stable_el8:
|
||||
repository: 'stable'
|
||||
description: 'FRR Stable EL8'
|
||||
osname: 'frr'
|
||||
release: 'el8'
|
||||
baseurl: 'https://rpm.frrouting.org/repo/el8/frr/'
|
||||
gpgkey: 'https://packagerepo.service.consul/frr/gpg/RPM-GPG-KEY-FRR'
|
||||
frr_extras_el8:
|
||||
repository: 'extras'
|
||||
description: 'FRR Extras EL8'
|
||||
osname: 'frr'
|
||||
release: 'el8'
|
||||
baseurl: 'https://rpm.frrouting.org/repo/el8/extras/'
|
||||
gpgkey: 'https://packagerepo.service.consul/frr/gpg/RPM-GPG-KEY-FRR'
|
||||
frr_stable_el9:
|
||||
repository: 'stable'
|
||||
description: 'FRR Stable EL9'
|
||||
osname: 'frr'
|
||||
release: 'el9'
|
||||
baseurl: 'https://rpm.frrouting.org/repo/el9/frr/'
|
||||
gpgkey: 'https://packagerepo.service.consul/frr/gpg/RPM-GPG-KEY-FRR'
|
||||
frr_extras_el9:
|
||||
repository: 'extras'
|
||||
description: 'FRR Extras el9'
|
||||
osname: 'frr'
|
||||
release: 'el9'
|
||||
baseurl: 'https://rpm.frrouting.org/repo/el9/extras/'
|
||||
gpgkey: 'https://packagerepo.service.consul/frr/gpg/RPM-GPG-KEY-FRR'
|
||||
k8s_1.32:
|
||||
repository: '1.32'
|
||||
description: 'Kubernetes 1.32'
|
||||
@ -143,62 +31,6 @@ profiles::reposync::repos_list:
|
||||
release: '1.33'
|
||||
baseurl: 'https://pkgs.k8s.io/core:/stable:/v1.33/rpm/'
|
||||
gpgkey: 'https://pkgs.k8s.io/core:/stable:/v1.33/rpm/repodata/repomd.xml.key'
|
||||
mariadb_11_8_el8:
|
||||
repository: 'el8'
|
||||
description: 'MariaDB 11.8'
|
||||
osname: 'mariadb'
|
||||
release: '11.8'
|
||||
baseurl: 'http://mariadb.mirror.digitalpacific.com.au/yum/11.8/rhel8-amd64/'
|
||||
gpgkey: 'https://mariadb.mirror.digitalpacific.com.au/yum/RPM-GPG-KEY-MariaDB'
|
||||
mariadb_11_8_el9:
|
||||
repository: 'el9'
|
||||
description: 'MariaDB 11.8'
|
||||
osname: 'mariadb'
|
||||
release: '11.8'
|
||||
baseurl: 'http://mariadb.mirror.digitalpacific.com.au/yum/11.8/rhel9-amd64/'
|
||||
gpgkey: 'https://mariadb.mirror.digitalpacific.com.au/yum/RPM-GPG-KEY-MariaDB'
|
||||
openvox7_el8:
|
||||
repository: '8'
|
||||
description: 'openvox 7 EL8'
|
||||
osname: 'openvox7'
|
||||
release: 'el'
|
||||
baseurl: 'https://yum.voxpupuli.org/openvox7/el/8/x86_64/'
|
||||
gpgkey: 'https://yum.voxpupuli.org/GPG-KEY-openvox.pub'
|
||||
openvox7_el9:
|
||||
repository: '9'
|
||||
description: 'openvox 7 EL9'
|
||||
osname: 'openvox7'
|
||||
release: 'el'
|
||||
baseurl: 'https://yum.voxpupuli.org/openvox7/el/9/x86_64/'
|
||||
gpgkey: 'https://yum.voxpupuli.org/GPG-KEY-openvox.pub'
|
||||
openvox7_el10:
|
||||
repository: '10'
|
||||
description: 'openvox 7 EL10'
|
||||
osname: 'openvox7'
|
||||
release: 'el'
|
||||
baseurl: 'https://yum.voxpupuli.org/openvox7/el/10/x86_64/'
|
||||
gpgkey: 'https://yum.voxpupuli.org/GPG-KEY-openvox.pub'
|
||||
openvox8_el8:
|
||||
repository: '8'
|
||||
description: 'openvox 8 EL8'
|
||||
osname: 'openvox8'
|
||||
release: 'el'
|
||||
baseurl: 'https://yum.voxpupuli.org/openvox8/el/8/x86_64/'
|
||||
gpgkey: 'https://yum.voxpupuli.org/GPG-KEY-openvox.pub'
|
||||
openvox8_el9:
|
||||
repository: '9'
|
||||
description: 'openvox 8 EL9'
|
||||
osname: 'openvox8'
|
||||
release: 'el'
|
||||
baseurl: 'https://yum.voxpupuli.org/openvox8/el/9/x86_64/'
|
||||
gpgkey: 'https://yum.voxpupuli.org/GPG-KEY-openvox.pub'
|
||||
openvox8_el10:
|
||||
repository: '10'
|
||||
description: 'openvox 8 EL10'
|
||||
osname: 'openvox8'
|
||||
release: 'el'
|
||||
baseurl: 'https://yum.voxpupuli.org/openvox8/el/10/x86_64/'
|
||||
gpgkey: 'https://yum.voxpupuli.org/GPG-KEY-openvox.pub'
|
||||
puppet7_el8:
|
||||
repository: '8'
|
||||
description: 'Puppet 7 EL8'
|
||||
@ -227,62 +59,6 @@ profiles::reposync::repos_list:
|
||||
release: 'el'
|
||||
baseurl: 'https://yum.puppet.com/puppet8/el/9/x86_64/'
|
||||
gpgkey: 'https://yum.puppet.com/RPM-GPG-KEY-puppet-20250406'
|
||||
postgresql_rhel8_common:
|
||||
repository: 'common'
|
||||
description: 'PostgreSQL Common RHEL 8'
|
||||
osname: 'postgresql'
|
||||
release: 'rhel8'
|
||||
baseurl: 'https://download.postgresql.org/pub/repos/yum/common/redhat/rhel-8-x86_64/'
|
||||
gpgkey: 'https://download.postgresql.org/pub/repos/yum/keys/PGDG-RPM-GPG-KEY-RHEL'
|
||||
postgresql_rhel9_common:
|
||||
repository: 'common'
|
||||
description: 'PostgreSQL Common RHEL 9'
|
||||
osname: 'postgresql'
|
||||
release: 'rhel9'
|
||||
baseurl: 'https://download.postgresql.org/pub/repos/yum/common/redhat/rhel-9-x86_64/'
|
||||
gpgkey: 'https://download.postgresql.org/pub/repos/yum/keys/PGDG-RPM-GPG-KEY-RHEL'
|
||||
postgresql_rhel8_15:
|
||||
repository: '15'
|
||||
description: 'PostgreSQL 15 RHEL 8'
|
||||
osname: 'postgresql'
|
||||
release: 'rhel8'
|
||||
baseurl: 'https://download.postgresql.org/pub/repos/yum/15/redhat/rhel-8-x86_64/'
|
||||
gpgkey: 'https://download.postgresql.org/pub/repos/yum/keys/PGDG-RPM-GPG-KEY-RHEL'
|
||||
postgresql_rhel9_15:
|
||||
repository: '15'
|
||||
description: 'PostgreSQL 15 RHEL 9'
|
||||
osname: 'postgresql'
|
||||
release: 'rhel9'
|
||||
baseurl: 'https://download.postgresql.org/pub/repos/yum/15/redhat/rhel-9-x86_64/'
|
||||
gpgkey: 'https://download.postgresql.org/pub/repos/yum/keys/PGDG-RPM-GPG-KEY-RHEL'
|
||||
postgresql_rhel8_16:
|
||||
repository: '16'
|
||||
description: 'PostgreSQL 16 RHEL 8'
|
||||
osname: 'postgresql'
|
||||
release: 'rhel8'
|
||||
baseurl: 'https://download.postgresql.org/pub/repos/yum/16/redhat/rhel-8-x86_64/'
|
||||
gpgkey: 'https://download.postgresql.org/pub/repos/yum/keys/PGDG-RPM-GPG-KEY-RHEL'
|
||||
postgresql_rhel9_16:
|
||||
repository: '16'
|
||||
description: 'PostgreSQL 16 RHEL 9'
|
||||
osname: 'postgresql'
|
||||
release: 'rhel9'
|
||||
baseurl: 'https://download.postgresql.org/pub/repos/yum/16/redhat/rhel-9-x86_64/'
|
||||
gpgkey: 'https://download.postgresql.org/pub/repos/yum/keys/PGDG-RPM-GPG-KEY-RHEL'
|
||||
postgresql_rhel8_17:
|
||||
repository: '17'
|
||||
description: 'PostgreSQL 17 RHEL 8'
|
||||
osname: 'postgresql'
|
||||
release: 'rhel8'
|
||||
baseurl: 'https://download.postgresql.org/pub/repos/yum/17/redhat/rhel-8-x86_64/'
|
||||
gpgkey: 'https://download.postgresql.org/pub/repos/yum/keys/PGDG-RPM-GPG-KEY-RHEL'
|
||||
postgresql_rhel9_17:
|
||||
repository: '17'
|
||||
description: 'PostgreSQL 17 RHEL 9'
|
||||
osname: 'postgresql'
|
||||
release: 'rhel9'
|
||||
baseurl: 'https://download.postgresql.org/pub/repos/yum/17/redhat/rhel-9-x86_64/'
|
||||
gpgkey: 'https://download.postgresql.org/pub/repos/yum/keys/PGDG-RPM-GPG-KEY-RHEL'
|
||||
zfs_dkms_rhel8:
|
||||
repository: 'dkms'
|
||||
description: 'ZFS DKMS RHEL 8'
|
||||
|
||||
@ -1,17 +1,9 @@
|
||||
---
|
||||
profiles::yum::global::repos:
|
||||
postgresql-15:
|
||||
name: postgresql-15
|
||||
descr: postgresql-15 repository
|
||||
target: /etc/yum.repos.d/postgresql.repo
|
||||
baseurl: https://packagerepo.service.consul/postgresql/rhel%{facts.os.release.major}/15-daily/%{facts.os.architecture}/os
|
||||
gpgkey: https://packagerepo.service.consul/postgresql/rhel%{facts.os.release.major}/15-daily/%{facts.os.architecture}/os/PGDG-RPM-GPG-KEY-RHEL
|
||||
ensure: present
|
||||
postgresql-common:
|
||||
name: postgresql-common
|
||||
descr: postgresql-common repository
|
||||
target: /etc/yum.repos.d/postgresql.repo
|
||||
baseurl: https://packagerepo.service.consul/postgresql/rhel%{facts.os.release.major}/common-daily/%{facts.os.architecture}/os
|
||||
gpgkey: https://packagerepo.service.consul/postgresql/rhel%{facts.os.release.major}/common-daily/%{facts.os.architecture}/os/PGDG-RPM-GPG-KEY-RHEL
|
||||
ensure: present
|
||||
|
||||
profiles::sql::patroni::cluster_name: "patroni-%{facts.environment}"
|
||||
profiles::sql::patroni::postgres_exporter_enabled: true
|
||||
|
||||
@ -47,14 +47,6 @@ profiles::consul::client::node_rules:
|
||||
|
||||
profiles::yum::global::repos:
|
||||
postgresql-17:
|
||||
name: postgresql-17
|
||||
descr: postgresql-17 repository
|
||||
target: /etc/yum.repos.d/postgresql.repo
|
||||
baseurl: https://packagerepo.service.consul/postgresql/rhel%{facts.os.release.major}/17-daily/%{facts.os.architecture}/os
|
||||
gpgkey: https://packagerepo.service.consul/postgresql/rhel%{facts.os.release.major}/17-daily/%{facts.os.architecture}/os/PGDG-RPM-GPG-KEY-RHEL
|
||||
ensure: present
|
||||
postgresql-common:
|
||||
name: postgresql-common
|
||||
descr: postgresql-common repository
|
||||
target: /etc/yum.repos.d/postgresql.repo
|
||||
baseurl: https://packagerepo.service.consul/postgresql/rhel%{facts.os.release.major}/common-daily/%{facts.os.architecture}/os
|
||||
gpgkey: https://packagerepo.service.consul/postgresql/rhel%{facts.os.release.major}/common-daily/%{facts.os.architecture}/os/PGDG-RPM-GPG-KEY-RHEL
|
||||
ensure: present
|
||||
|
||||
@ -29,6 +29,7 @@ profiles::consul::server::acl:
|
||||
profiles::pki::vault::alt_names:
|
||||
- consul.main.unkin.net
|
||||
- consul.service.consul
|
||||
- "consul.service.%{facts.country}-%{facts.region}.consul"
|
||||
- consul
|
||||
|
||||
# manage a simple nginx reverse proxy
|
||||
@ -38,6 +39,7 @@ profiles::nginx::simpleproxy::nginx_aliases:
|
||||
- consul.main.unkin.net
|
||||
profiles::nginx::simpleproxy::proxy_port: 8500
|
||||
profiles::nginx::simpleproxy::proxy_path: '/'
|
||||
nginx::client_max_body_size: 512M
|
||||
|
||||
# consul
|
||||
profiles::consul::client::node_rules:
|
||||
@ -134,19 +136,9 @@ frrouting::ospfd_interfaces:
|
||||
frrouting::daemons:
|
||||
ospfd: true
|
||||
|
||||
# additional repos
|
||||
# additional repos - enable needed repositories
|
||||
profiles::yum::global::repos:
|
||||
frr-extras:
|
||||
name: frr-extras
|
||||
descr: frr-extras repository
|
||||
target: /etc/yum.repos.d/frr-extras.repo
|
||||
baseurl: https://packagerepo.service.consul/frr/el9/extras-daily/%{facts.os.architecture}/os
|
||||
gpgkey: https://packagerepo.service.consul/frr/el9/extras-daily/%{facts.os.architecture}/os/RPM-GPG-KEY-FRR
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
frr-stable:
|
||||
name: frr-stable
|
||||
descr: frr-stable repository
|
||||
target: /etc/yum.repos.d/frr-stable.repo
|
||||
baseurl: https://packagerepo.service.consul/frr/el9/stable-daily/%{facts.os.architecture}/os
|
||||
gpgkey: https://packagerepo.service.consul/frr/el9/stable-daily/%{facts.os.architecture}/os/RPM-GPG-KEY-FRR
|
||||
mirrorlist: absent
|
||||
ensure: present
|
||||
|
||||
@ -2,9 +2,12 @@
|
||||
profiles::vault::server::members_role: roles::infra::storage::vault
|
||||
profiles::vault::server::members_lookup: true
|
||||
profiles::vault::server::data_dir: /data/vault
|
||||
profiles::vault::server::plugin_dir: /opt/openbao-plugins
|
||||
profiles::vault::server::manage_storage_dir: true
|
||||
profiles::vault::server::tls_disable: false
|
||||
vault::download_url: http://repos.main.unkin.net/unkin/8/x86_64/os/Archives/vault_1.15.5_linux_amd64.zip
|
||||
profiles::vault::server::package_name: openbao
|
||||
profiles::vault::server::package_ensure: 2.4.4
|
||||
profiles::vault::server::disable_openbao: false
|
||||
|
||||
# additional altnames
|
||||
profiles::pki::vault::alt_names:
|
||||
@ -22,3 +25,6 @@ profiles::nginx::simpleproxy::proxy_scheme: 'http'
|
||||
profiles::nginx::simpleproxy::proxy_host: '127.0.0.1'
|
||||
profiles::nginx::simpleproxy::proxy_port: 8200
|
||||
profiles::nginx::simpleproxy::proxy_path: '/'
|
||||
|
||||
profiles::packages::include:
|
||||
openbao-plugins: {}
|
||||
|
||||
@ -4,7 +4,6 @@ class exporters::frr_exporter (
|
||||
String $group = 'frr_exporter',
|
||||
Boolean $manage_user = true,
|
||||
Boolean $manage_service = true,
|
||||
Stdlib::Host $addr = $facts['networking']['ip'],
|
||||
Stdlib::Port $port = 9342,
|
||||
Stdlib::Absolutepath $exec_path = '/usr/bin/frr_exporter',
|
||||
Stdlib::Absolutepath $socket_dir = '/var/run/frr',
|
||||
@ -17,6 +16,14 @@ class exporters::frr_exporter (
|
||||
ensure => installed,
|
||||
}
|
||||
|
||||
# ensure the frr_exporter user can read the directory
|
||||
file { $socket_dir:
|
||||
ensure => directory,
|
||||
owner => 'frr',
|
||||
group => 'frr',
|
||||
mode => '0751',
|
||||
}
|
||||
|
||||
# manage the user/group
|
||||
if $manage_user {
|
||||
group { $group:
|
||||
@ -46,7 +53,7 @@ class exporters::frr_exporter (
|
||||
# manage consul service
|
||||
consul::service { 'frr_exporter':
|
||||
service_name => 'frr_exporter',
|
||||
address => $addr,
|
||||
address => $facts['networking']['ip'],
|
||||
port => $port,
|
||||
tags => [
|
||||
'metrics',
|
||||
|
||||
@ -4,7 +4,6 @@ class exporters::node_exporter (
|
||||
String $group = 'node_exporter',
|
||||
Boolean $manage_user = true,
|
||||
Boolean $manage_service = true,
|
||||
Stdlib::Host $addr = $facts['networking']['ip'],
|
||||
Stdlib::Port $port = 9100,
|
||||
Stdlib::Absolutepath $exec_path = '/usr/bin/node_exporter',
|
||||
Boolean $cleanup_old_node_exporter = false,
|
||||
@ -59,7 +58,7 @@ class exporters::node_exporter (
|
||||
# manage consul service
|
||||
consul::service { 'node_exporter':
|
||||
service_name => 'node_exporter',
|
||||
address => $addr,
|
||||
address => $facts['networking']['ip'],
|
||||
port => $port,
|
||||
tags => [
|
||||
'metrics',
|
||||
|
||||
15
modules/externaldns/manifests/init.pp
Normal file
15
modules/externaldns/manifests/init.pp
Normal file
@ -0,0 +1,15 @@
|
||||
# ExternalDNS BIND module - automatically configures master or slave
|
||||
class externaldns (
|
||||
Stdlib::Fqdn $bind_master_hostname,
|
||||
Array[Stdlib::Fqdn] $k8s_zones = [],
|
||||
Array[Stdlib::Fqdn] $slave_servers = [],
|
||||
String $externaldns_key_secret = '',
|
||||
String $externaldns_key_algorithm = 'hmac-sha256',
|
||||
) {
|
||||
|
||||
if $trusted['certname'] == $bind_master_hostname {
|
||||
include externaldns::master
|
||||
} else {
|
||||
include externaldns::slave
|
||||
}
|
||||
}
|
||||
45
modules/externaldns/manifests/master.pp
Normal file
45
modules/externaldns/manifests/master.pp
Normal file
@ -0,0 +1,45 @@
|
||||
# ExternalDNS BIND master server class
|
||||
class externaldns::master inherits externaldns {
|
||||
|
||||
include bind
|
||||
|
||||
# Query PuppetDB for slave server IP addresses
|
||||
$slave_ips = $externaldns::slave_servers.map |$fqdn| {
|
||||
puppetdb_query("inventory[facts.networking.ip] { certname = '${fqdn}' }")[0]['facts.networking.ip']
|
||||
}.filter |$ip| { $ip != undef }
|
||||
|
||||
# Create TSIG key for ExternalDNS authentication
|
||||
bind::key { 'externaldns-key':
|
||||
algorithm => $externaldns::externaldns_key_algorithm,
|
||||
secret => $externaldns::externaldns_key_secret,
|
||||
}
|
||||
|
||||
# Create ACL for slave servers
|
||||
if !empty($slave_ips) {
|
||||
bind::acl { 'dns-slaves':
|
||||
addresses => $slave_ips,
|
||||
}
|
||||
}
|
||||
|
||||
# Create master zones for each Kubernetes domain
|
||||
$externaldns::k8s_zones.each |$zone| {
|
||||
bind::zone { $zone:
|
||||
zone_type => 'master',
|
||||
dynamic => true,
|
||||
allow_updates => ['key externaldns-key'],
|
||||
allow_transfers => empty($slave_ips) ? {
|
||||
true => [],
|
||||
false => ['dns-slaves'],
|
||||
},
|
||||
ns_notify => !empty($slave_ips),
|
||||
also_notify => $slave_ips,
|
||||
dnssec => false,
|
||||
}
|
||||
}
|
||||
|
||||
# Create default view to include the zones
|
||||
bind::view { 'externaldns':
|
||||
recursion => false,
|
||||
zones => $externaldns::k8s_zones,
|
||||
}
|
||||
}
|
||||
36
modules/externaldns/manifests/slave.pp
Normal file
36
modules/externaldns/manifests/slave.pp
Normal file
@ -0,0 +1,36 @@
|
||||
# ExternalDNS BIND slave server class
|
||||
class externaldns::slave inherits externaldns {
|
||||
|
||||
include bind
|
||||
|
||||
# Query PuppetDB for master server IP address
|
||||
$query = "inventory[facts.networking.ip] { certname = '${externaldns::bind_master_hostname}' }"
|
||||
$master_ip = puppetdb_query($query)[0]['facts.networking.ip']
|
||||
|
||||
# Create TSIG key for zone transfers (same as master)
|
||||
bind::key { 'externaldns-key':
|
||||
algorithm => $externaldns::externaldns_key_algorithm,
|
||||
secret => $externaldns::externaldns_key_secret,
|
||||
}
|
||||
|
||||
# Create ACL for master server
|
||||
bind::acl { 'dns-master':
|
||||
addresses => [$master_ip],
|
||||
}
|
||||
|
||||
# Create slave zones for each Kubernetes domain
|
||||
$externaldns::k8s_zones.each |$zone| {
|
||||
bind::zone { $zone:
|
||||
zone_type => 'slave',
|
||||
masters => [$master_ip],
|
||||
allow_notify => ['dns-master'],
|
||||
ns_notify => false,
|
||||
}
|
||||
}
|
||||
|
||||
# Create default view to include the zones
|
||||
bind::view { 'externaldns':
|
||||
recursion => false,
|
||||
zones => $externaldns::k8s_zones,
|
||||
}
|
||||
}
|
||||
@ -14,8 +14,18 @@ class frrouting (
|
||||
Optional[String] $mpls_ldp_router_id = undef,
|
||||
Optional[String] $mpls_ldp_transport_addr = undef,
|
||||
Array[String] $mpls_ldp_interfaces = [],
|
||||
Boolean $ospf_preferred_source_enable = false,
|
||||
Optional[Stdlib::IP::Address] $ospf_preferred_source = undef,
|
||||
Boolean $ospf_exclude_k8s_enable = false,
|
||||
Optional[Stdlib::IP::Address::V4::CIDR] $k8s_cluster_cidr = undef, # pod/cluster CIDR (e.g. 10.42.0.0/16)
|
||||
Optional[Stdlib::IP::Address::V4::CIDR] $k8s_service_cidr = undef, # service CIDR (e.g. 10.43.0.0/16)
|
||||
) {
|
||||
|
||||
# sanity check
|
||||
if $ospf_exclude_k8s_enable and $k8s_cluster_cidr == undef and $k8s_service_cidr == undef {
|
||||
warning('frrouting: ospf_exclude_k8s_enable is true but no k8s_*_cidr provided; nothing will be filtered.')
|
||||
}
|
||||
|
||||
$daemons_defaults = {
|
||||
'bgpd' => false,
|
||||
'ospfd' => true,
|
||||
|
||||
@ -2,6 +2,7 @@
|
||||
frr defaults traditional
|
||||
hostname <%= @hostname %>
|
||||
no ipv6 forwarding
|
||||
|
||||
<% @ospfd_interfaces.each do |iface, params| -%>
|
||||
interface <%= iface %>
|
||||
<% if params['area'] -%>
|
||||
@ -12,12 +13,31 @@ interface <%= iface %>
|
||||
<% end -%>
|
||||
exit
|
||||
<% end -%>
|
||||
|
||||
<%# ---- K8s exclude (prefix-list + route-map) ---- -%>
|
||||
<% if @ospf_exclude_k8s_enable && (@k8s_cluster_cidr || @k8s_service_cidr) -%>
|
||||
! Do not redistribute Kubernetes Pod/Service CIDRs
|
||||
ip prefix-list K8S-BLOCK seq 5 permit <%= @k8s_cluster_cidr %> le 32<% if !@k8s_cluster_cidr %> ! (unset)<% end %>
|
||||
<% if @k8s_service_cidr -%>
|
||||
ip prefix-list K8S-BLOCK seq 10 permit <%= @k8s_service_cidr %> le 32
|
||||
<% end -%>
|
||||
route-map CONNECTED-NON-K8S deny 5
|
||||
match ip address prefix-list K8S-BLOCK
|
||||
exit
|
||||
route-map CONNECTED-NON-K8S permit 100
|
||||
exit
|
||||
<% end -%>
|
||||
|
||||
router ospf
|
||||
ospf router-id <%= @ospfd_router_id %>
|
||||
log-adjacency-changes detail
|
||||
<% @ospfd_redistribute.each do |type| -%>
|
||||
<% if @ospf_exclude_k8s_enable && type == 'connected' && (@k8s_cluster_cidr || @k8s_service_cidr) -%>
|
||||
redistribute connected route-map CONNECTED-NON-K8S
|
||||
<% else -%>
|
||||
redistribute <%= type %>
|
||||
<% end -%>
|
||||
<% end -%>
|
||||
<% @ospfd_networks.each do |network| -%>
|
||||
network <%= network %>
|
||||
<% end -%>
|
||||
@ -31,6 +51,8 @@ router ospf
|
||||
mpls-te inter-as area 0.0.0.0
|
||||
<% end -%>
|
||||
exit
|
||||
|
||||
<%# ---- MPLS/LDP config ---- -%>
|
||||
<% if @mpls_ldp_router_id and @mpls_ldp_transport_addr and @mpls_ldp_interfaces.any? -%>
|
||||
mpls ldp
|
||||
router-id <%= @mpls_ldp_router_id %>
|
||||
@ -43,3 +65,13 @@ mpls ldp
|
||||
exit-address-family
|
||||
exit
|
||||
<% end -%>
|
||||
|
||||
<%# ---- Preferred OSPF source ---- -%>
|
||||
<% if @ospf_preferred_source_enable && @ospf_preferred_source -%>
|
||||
ip prefix-list ANY seq 5 permit 0.0.0.0/0 le 32
|
||||
route-map OSPF-SRC permit 10
|
||||
match ip address prefix-list ANY
|
||||
set src <%= @ospf_preferred_source %>
|
||||
exit
|
||||
ip protocol ospf route-map OSPF-SRC
|
||||
<% end -%>
|
||||
|
||||
28
modules/incus/lib/facter/incus_trust_list.rb
Normal file
28
modules/incus/lib/facter/incus_trust_list.rb
Normal file
@ -0,0 +1,28 @@
|
||||
# frozen_string_literal: true
|
||||
|
||||
# lib/facter/incus_trust_list.rb
|
||||
require 'json'
|
||||
|
||||
Facter.add(:incus_trust_list) do
|
||||
confine do
|
||||
# Only run on systems that have incus installed and running
|
||||
incus_path = Facter::Util::Resolution.which('incus')
|
||||
incus_path && File.exist?('/var/lib/incus/server.key')
|
||||
end
|
||||
|
||||
setcode do
|
||||
incus_path = Facter::Util::Resolution.which('incus')
|
||||
next {} unless incus_path
|
||||
|
||||
begin
|
||||
# Run incus config trust list --format=json
|
||||
trust_output = Facter::Core::Execution.execute("#{incus_path} config trust list --format=json")
|
||||
next {} if trust_output.empty?
|
||||
|
||||
# Parse the JSON output
|
||||
JSON.parse(trust_output)
|
||||
rescue StandardError
|
||||
{}
|
||||
end
|
||||
end
|
||||
end
|
||||
16
modules/incus/manifests/client.pp
Normal file
16
modules/incus/manifests/client.pp
Normal file
@ -0,0 +1,16 @@
|
||||
# incus::client
|
||||
#
|
||||
# This class configures a host as an incus client and exports its certificate
|
||||
# for automatic trust management on incus servers.
|
||||
#
|
||||
class incus::client {
|
||||
|
||||
# Export this client's certificate for collection by incus servers
|
||||
@@incus::client_cert { $facts['networking']['fqdn']:
|
||||
hostname => $facts['networking']['fqdn'],
|
||||
certificate => $facts['vault_cert_content'],
|
||||
fingerprint => $facts['vault_cert_fingerprint'],
|
||||
tag => 'incus_client',
|
||||
}
|
||||
|
||||
}
|
||||
41
modules/incus/manifests/client_cert.pp
Normal file
41
modules/incus/manifests/client_cert.pp
Normal file
@ -0,0 +1,41 @@
|
||||
# Define the exported resource type for incus client certificates
|
||||
define incus::client_cert (
|
||||
String $hostname,
|
||||
Optional[String] $certificate = undef,
|
||||
Optional[String] $fingerprint = undef,
|
||||
) {
|
||||
|
||||
# Only proceed if we have both certificate and fingerprint
|
||||
if $certificate and $fingerprint {
|
||||
|
||||
$trust_list = $facts['incus_trust_list']
|
||||
$existing_client = $trust_list.filter |$client| { $client['name'] == $hostname }
|
||||
|
||||
if $existing_client.empty {
|
||||
# Add new certificate
|
||||
exec { "incus_trust_add_${hostname}":
|
||||
path => ['/bin', '/usr/bin'],
|
||||
command => "echo '${certificate}' > /tmp/${hostname}.crt && \
|
||||
incus config trust add-certificate /tmp/${hostname}.crt --name ${hostname} && \
|
||||
rm -f /tmp/${hostname}.crt",
|
||||
unless => "incus config trust list --format=json | grep '\"name\":\"${hostname}\"'",
|
||||
}
|
||||
} else {
|
||||
# Check if fingerprints are different
|
||||
$existing_fingerprint = $existing_client[0]['fingerprint']
|
||||
|
||||
if $existing_fingerprint != $fingerprint {
|
||||
# Remove existing and add new certificate only if fingerprints differ
|
||||
exec { "incus_trust_update_${hostname}":
|
||||
path => ['/bin', '/usr/bin'],
|
||||
command => "incus config trust remove ${existing_fingerprint} && \
|
||||
echo '${certificate}' > /tmp/${hostname}.crt && \
|
||||
incus config trust add-certificate /tmp/${hostname}.crt --name ${hostname} && \
|
||||
rm -f /tmp/${hostname}.crt",
|
||||
onlyif => "incus config trust list --format=json | grep '${existing_fingerprint}'",
|
||||
}
|
||||
}
|
||||
# If fingerprints match, do nothing (certificate is already correct)
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -21,6 +21,10 @@ class incus (
|
||||
enable => true,
|
||||
hasstatus => true,
|
||||
hasrestart => true,
|
||||
subscribe => [
|
||||
File['/var/lib/incus/server.crt'],
|
||||
File['/var/lib/incus/server.key'],
|
||||
],
|
||||
}
|
||||
|
||||
file_line { 'subuid_root':
|
||||
@ -55,6 +59,22 @@ class incus (
|
||||
}
|
||||
}
|
||||
|
||||
file { '/var/lib/incus/server.crt':
|
||||
ensure => file,
|
||||
source => '/etc/pki/tls/vault/certificate.crt',
|
||||
owner => 'root',
|
||||
group => 'root',
|
||||
mode => '0644',
|
||||
}
|
||||
|
||||
file { '/var/lib/incus/server.key':
|
||||
ensure => file,
|
||||
source => '/etc/pki/tls/vault/private.key',
|
||||
owner => 'root',
|
||||
group => 'root',
|
||||
mode => '0600',
|
||||
}
|
||||
|
||||
if $facts['incus'] and $facts['incus']['config'] {
|
||||
# set core.https_address
|
||||
if $facts['incus']['config']['core.https_address'] != "${server_addr}:${server_port}" {
|
||||
@ -72,5 +92,10 @@ class incus (
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
# Collect exported client certificates and manage trust
|
||||
Incus::Client_cert <<| tag == 'incus_client' |>> {
|
||||
require => Service['incus'],
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
11
modules/libs/lib/facter/vault_cert_content.rb
Normal file
11
modules/libs/lib/facter/vault_cert_content.rb
Normal file
@ -0,0 +1,11 @@
|
||||
# frozen_string_literal: true
|
||||
|
||||
# lib/facter/vault_cert_content.rb
|
||||
|
||||
Facter.add(:vault_cert_content) do
|
||||
confine kernel: 'Linux'
|
||||
setcode do
|
||||
cert_path = '/etc/pki/tls/vault/certificate.crt'
|
||||
File.read(cert_path) if File.exist?(cert_path) && File.readable?(cert_path)
|
||||
end
|
||||
end
|
||||
23
modules/libs/lib/facter/vault_cert_fingerprint.rb
Normal file
23
modules/libs/lib/facter/vault_cert_fingerprint.rb
Normal file
@ -0,0 +1,23 @@
|
||||
# frozen_string_literal: true
|
||||
|
||||
# lib/facter/vault_cert_fingerprint.rb
|
||||
|
||||
Facter.add(:vault_cert_fingerprint) do
|
||||
confine kernel: 'Linux'
|
||||
setcode do
|
||||
require 'openssl'
|
||||
require 'digest'
|
||||
|
||||
cert_path = '/etc/pki/tls/vault/certificate.crt'
|
||||
if File.exist?(cert_path) && File.readable?(cert_path)
|
||||
begin
|
||||
cert_content = File.read(cert_path)
|
||||
cert = OpenSSL::X509::Certificate.new(cert_content)
|
||||
# Calculate SHA256 fingerprint like incus does
|
||||
Digest::SHA256.hexdigest(cert.to_der)
|
||||
rescue StandardError
|
||||
nil
|
||||
end
|
||||
end
|
||||
end
|
||||
end
|
||||
2
modules/rke2/files/check_k8s_api.sh
Normal file
2
modules/rke2/files/check_k8s_api.sh
Normal file
@ -0,0 +1,2 @@
|
||||
#!/usr/bin/bash
|
||||
/var/lib/rancher/rke2/bin/kubectl --kubeconfig=/etc/rancher/rke2/rke2.yaml get --raw /livez
|
||||
9
modules/rke2/files/rke2-canal-config.yaml
Normal file
9
modules/rke2/files/rke2-canal-config.yaml
Normal file
@ -0,0 +1,9 @@
|
||||
apiVersion: helm.cattle.io/v1
|
||||
kind: HelmChartConfig
|
||||
metadata:
|
||||
name: rke2-canal
|
||||
namespace: kube-system
|
||||
spec:
|
||||
valuesContent: |-
|
||||
flannel:
|
||||
iface: "loopback0"
|
||||
20
modules/rke2/files/rke2-nginx-ingress-config.yaml
Normal file
20
modules/rke2/files/rke2-nginx-ingress-config.yaml
Normal file
@ -0,0 +1,20 @@
|
||||
---
|
||||
apiVersion: helm.cattle.io/v1
|
||||
kind: HelmChartConfig
|
||||
metadata:
|
||||
name: rke2-ingress-nginx
|
||||
namespace: kube-system
|
||||
spec:
|
||||
valuesContent: |-
|
||||
controller:
|
||||
hostPort:
|
||||
enabled: false
|
||||
service:
|
||||
enabled: true
|
||||
type: LoadBalancer
|
||||
externalTrafficPolicy: Local
|
||||
loadBalancerClass: purelb.io/purelb
|
||||
allocateLoadBalancerNodePorts: false
|
||||
annotations:
|
||||
purelb.io/service-group: common
|
||||
purelb.io/addresses: "198.18.200.0"
|
||||
15
modules/rke2/lib/facter/helm_repos.rb
Normal file
15
modules/rke2/lib/facter/helm_repos.rb
Normal file
@ -0,0 +1,15 @@
|
||||
# frozen_string_literal: true
|
||||
|
||||
require 'facter/util/helm'
|
||||
|
||||
Facter.add(:helm_repos) do
|
||||
confine kernel: 'Linux'
|
||||
confine enc_role: [
|
||||
'roles::infra::k8s::control',
|
||||
'roles::infra::k8s::compute'
|
||||
]
|
||||
|
||||
setcode do
|
||||
Facter::Util::Helm.get_helm_repos('/usr/bin/helm')
|
||||
end
|
||||
end
|
||||
39
modules/rke2/lib/facter/k8s_masters.rb
Normal file
39
modules/rke2/lib/facter/k8s_masters.rb
Normal file
@ -0,0 +1,39 @@
|
||||
# frozen_string_literal: true
|
||||
|
||||
require 'json'
|
||||
require 'open3'
|
||||
|
||||
Facter.add(:k8s_masters) do
|
||||
confine do
|
||||
File.exist?('/etc/rancher/rke2/rke2.yaml') &&
|
||||
File.executable?('/usr/bin/kubectl')
|
||||
end
|
||||
|
||||
setcode do
|
||||
env = { 'KUBECONFIG' => '/etc/rancher/rke2/rke2.yaml' }
|
||||
cmd = ['/usr/bin/kubectl', 'get', 'nodes', '-o', 'json']
|
||||
|
||||
stdout, stderr, status = Open3.capture3(env, *cmd)
|
||||
|
||||
if status.success?
|
||||
json = JSON.parse(stdout)
|
||||
|
||||
master_count = json['items'].count do |item|
|
||||
roles = item.dig('metadata', 'labels') || {}
|
||||
|
||||
# Look for well-known labels assigned to control-plane nodes
|
||||
roles.any? do |key, _|
|
||||
key =~ %r{node-role\.kubernetes\.io/(control-plane|master|etcd)}
|
||||
end
|
||||
end
|
||||
|
||||
master_count
|
||||
else
|
||||
Facter.debug("kubectl error: #{stderr}")
|
||||
0
|
||||
end
|
||||
rescue StandardError => e
|
||||
Facter.debug("Exception in k8s_masters fact: #{e.message}")
|
||||
0
|
||||
end
|
||||
end
|
||||
29
modules/rke2/lib/facter/k8s_namespaces.rb
Normal file
29
modules/rke2/lib/facter/k8s_namespaces.rb
Normal file
@ -0,0 +1,29 @@
|
||||
# frozen_string_literal: true
|
||||
|
||||
require 'json'
|
||||
require 'open3'
|
||||
|
||||
Facter.add(:k8s_namespaces) do
|
||||
confine do
|
||||
File.exist?('/etc/rancher/rke2/rke2.yaml') &&
|
||||
File.executable?('/usr/bin/kubectl') # Adjust this path if needed
|
||||
end
|
||||
|
||||
setcode do
|
||||
env = { 'KUBECONFIG' => '/etc/rancher/rke2/rke2.yaml' }
|
||||
cmd = ['/usr/bin/kubectl', 'get', 'namespaces', '-o', 'json']
|
||||
|
||||
stdout, stderr, status = Open3.capture3(env, *cmd)
|
||||
|
||||
if status.success?
|
||||
json = JSON.parse(stdout)
|
||||
json['items'].map { |item| item['metadata']['name'] }
|
||||
else
|
||||
Facter.debug("kubectl error: #{stderr}")
|
||||
[]
|
||||
end
|
||||
rescue StandardError => e
|
||||
Facter.debug("Exception in k8s_namespaces fact: #{e.message}")
|
||||
[]
|
||||
end
|
||||
end
|
||||
31
modules/rke2/lib/facter/util/helm.rb
Normal file
31
modules/rke2/lib/facter/util/helm.rb
Normal file
@ -0,0 +1,31 @@
|
||||
# frozen_string_literal: true
|
||||
|
||||
require 'facter'
|
||||
require 'json'
|
||||
|
||||
# a simple helm module
|
||||
module Facter::Util::Helm
|
||||
def self.get_helm_repos(helm_cmd)
|
||||
return [] unless File.executable?(helm_cmd)
|
||||
|
||||
output = Facter::Core::Execution.execute(
|
||||
"#{helm_cmd} repo list --output json --repository-config /etc/helm/repositories.yaml",
|
||||
on_fail: nil
|
||||
)
|
||||
return [] if output.to_s.strip.empty?
|
||||
|
||||
parse_helm_output(output)
|
||||
rescue StandardError => e
|
||||
Facter.debug("helm_repos fact error: #{e}")
|
||||
[]
|
||||
end
|
||||
|
||||
def self.parse_helm_output(output)
|
||||
JSON.parse(output).map do |repo|
|
||||
{
|
||||
'name' => repo['name'],
|
||||
'url' => repo['url']
|
||||
}
|
||||
end
|
||||
end
|
||||
end
|
||||
85
modules/rke2/manifests/config.pp
Normal file
85
modules/rke2/manifests/config.pp
Normal file
@ -0,0 +1,85 @@
|
||||
# config rke2
|
||||
class rke2::config (
|
||||
Enum['server', 'agent'] $node_type = $rke2::node_type,
|
||||
Stdlib::Absolutepath $config_file = $rke2::config_file,
|
||||
Hash $config_hash = $rke2::config_hash,
|
||||
Stdlib::HTTPSUrl $join_url = $rke2::join_url,
|
||||
Stdlib::Fqdn $bootstrap_node = $rke2::bootstrap_node,
|
||||
String $node_token = $rke2::node_token,
|
||||
Array[String[1]] $extra_config_files = $rke2::extra_config_files,
|
||||
){
|
||||
|
||||
# if its not the bootstrap node, add join path to config
|
||||
if $node_type == 'server' {
|
||||
if $trusted['certname'] != $bootstrap_node {
|
||||
$config = merge($config_hash, {
|
||||
server => $join_url,
|
||||
token => $node_token,
|
||||
} )
|
||||
}else{
|
||||
$config = merge($config_hash, {})
|
||||
}
|
||||
} elsif $node_type == 'agent' {
|
||||
$config = merge($config_hash, {
|
||||
server => $join_url,
|
||||
token => $node_token,
|
||||
} )
|
||||
}else{
|
||||
$config = $config_hash
|
||||
}
|
||||
|
||||
# create the config file
|
||||
file { $config_file:
|
||||
ensure => file,
|
||||
content => Sensitive($config.to_yaml),
|
||||
owner => 'root',
|
||||
group => 'root',
|
||||
mode => '0644',
|
||||
require => Package["rke2-${node_type}"],
|
||||
before => Service["rke2-${node_type}"],
|
||||
}
|
||||
|
||||
# create a script to verify k8s api is up (used by consul)
|
||||
file {'/usr/local/bin/check_k8s_api.sh':
|
||||
ensure => file,
|
||||
owner => 'root',
|
||||
group => 'root',
|
||||
mode => '0755',
|
||||
source => 'puppet:///modules/rke2/check_k8s_api.sh',
|
||||
}
|
||||
|
||||
# symlink kubectl to path
|
||||
file {'/usr/bin/kubectl':
|
||||
ensure => link,
|
||||
target => '/var/lib/rancher/rke2/bin/kubectl',
|
||||
require => Package["rke2-${node_type}"],
|
||||
}
|
||||
|
||||
# when ProtectKernelDefaults=true
|
||||
sysctl { 'vm.overcommit_memory':
|
||||
value => '1',
|
||||
before => Service["rke2-${node_type}"],
|
||||
}
|
||||
sysctl { 'kernel.panic':
|
||||
value => '10',
|
||||
before => Service["rke2-${node_type}"],
|
||||
}
|
||||
|
||||
# on the controller nodes only
|
||||
if $node_type == 'server' and $facts['k8s_masters'] and $facts['k8s_masters'] > 2 {
|
||||
|
||||
# manage extra config config (these are not dependent on helm)
|
||||
$extra_config_files.each |$file| {
|
||||
|
||||
file {"/var/lib/rancher/rke2/server/manifests/${file}.yaml":
|
||||
ensure => file,
|
||||
owner => 'root',
|
||||
group => 'root',
|
||||
mode => '0644',
|
||||
source => "puppet:///modules/rke2/${file}.yaml",
|
||||
require => Service['rke2-server'],
|
||||
}
|
||||
}
|
||||
|
||||
}
|
||||
}
|
||||
43
modules/rke2/manifests/helm.pp
Normal file
43
modules/rke2/manifests/helm.pp
Normal file
@ -0,0 +1,43 @@
|
||||
# manage helm
|
||||
class rke2::helm (
|
||||
Enum['server', 'agent'] $node_type = $rke2::node_type,
|
||||
Stdlib::Fqdn $bootstrap_node = $rke2::bootstrap_node,
|
||||
Boolean $helm_install = $rke2::helm_install,
|
||||
Hash $helm_repos = $rke2::helm_repos
|
||||
){
|
||||
|
||||
# when installing helm, manage the repos
|
||||
if $helm_install {
|
||||
|
||||
package {'helm':
|
||||
ensure => installed,
|
||||
}
|
||||
|
||||
file { '/etc/helm':
|
||||
ensure => directory,
|
||||
owner => 'root',
|
||||
group => 'root',
|
||||
mode => '0755',
|
||||
}
|
||||
|
||||
# on the controller nodes only, and after 3 master nodes exist
|
||||
if $node_type == 'server' and $facts['k8s_masters'] and $facts['k8s_masters'] > 2 {
|
||||
|
||||
# check if the repo already exists
|
||||
$helm_repos.each | String $repo, Stdlib::HTTPSUrl $url | {
|
||||
|
||||
# if repo isnt in repo list from helm, install it
|
||||
if ! $facts['helm_repos'].any |$existing| { $existing['name'] == $repo } {
|
||||
|
||||
exec { "helm_add_repo_${repo}":
|
||||
command => "helm repo add ${repo} ${url} --repository-config /etc/helm/repositories.yaml",
|
||||
path => ['/usr/bin'],
|
||||
environment => [
|
||||
'KUBECONFIG=/etc/rancher/rke2/rke2.yaml',
|
||||
],
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
23
modules/rke2/manifests/init.pp
Normal file
23
modules/rke2/manifests/init.pp
Normal file
@ -0,0 +1,23 @@
|
||||
# manage rke2
|
||||
class rke2 (
|
||||
Enum['server', 'agent'] $node_type = $rke2::params::node_type,
|
||||
String $rke2_version = $rke2::params::rke2_version,
|
||||
String $rke2_release = $rke2::params::rke2_release,
|
||||
Stdlib::Absolutepath $config_file = $rke2::params::config_file,
|
||||
Hash $config_hash = $rke2::params::config_hash,
|
||||
Stdlib::HTTPSUrl $join_url = $rke2::params::join_url,
|
||||
Stdlib::Fqdn $bootstrap_node = $rke2::params::bootstrap_node,
|
||||
String $node_token = $rke2::params::node_token,
|
||||
Boolean $helm_install = $rke2::params::helm_install,
|
||||
Hash $helm_repos = $rke2::params::helm_repos,
|
||||
Array[String[1]] $extra_config_files = $rke2::params::extra_config_files,
|
||||
Stdlib::HTTPUrl $container_archive_source = $rke2::params::container_archive_source,
|
||||
) inherits rke2::params {
|
||||
|
||||
include rke2::install
|
||||
include rke2::config
|
||||
include rke2::service
|
||||
include rke2::helm
|
||||
|
||||
Class['rke2::install'] -> Class['rke2::service'] -> Class['rke2::helm']
|
||||
}
|
||||
53
modules/rke2/manifests/install.pp
Normal file
53
modules/rke2/manifests/install.pp
Normal file
@ -0,0 +1,53 @@
|
||||
# install rke2
|
||||
class rke2::install (
|
||||
Enum['server', 'agent'] $node_type = $rke2::node_type,
|
||||
String $rke2_version = $rke2::rke2_version,
|
||||
String $rke2_release = $rke2::rke2_release,
|
||||
Stdlib::HTTPUrl $container_archive_source = $rke2::container_archive_source,
|
||||
){
|
||||
|
||||
# versionlock rke2
|
||||
yum::versionlock{"rke2-${node_type}":
|
||||
ensure => present,
|
||||
version => "${rke2_version}~${rke2_release}",
|
||||
}
|
||||
|
||||
# install rke2
|
||||
package {"rke2-${node_type}":
|
||||
ensure => "${rke2_version}~${rke2_release}",
|
||||
}
|
||||
|
||||
# ensure images path exists
|
||||
file { ['/var/lib/rancher/rke2/agent', '/var/lib/rancher/rke2/agent/images']:
|
||||
ensure => 'directory',
|
||||
owner => 'root',
|
||||
group => 'root',
|
||||
mode => '0750',
|
||||
require => Package["rke2-${node_type}"],
|
||||
before => Service["rke2-${node_type}"],
|
||||
}
|
||||
|
||||
# download required archive of containers
|
||||
archive { '/var/lib/rancher/rke2/agent/images/rke2-images.linux-amd64.tar.zst':
|
||||
ensure => present,
|
||||
source => "https://artifactapi.k8s.syd1.au.unkin.net/api/v1/remote/github/rancher/rke2/releases/download/v${rke2_version}%2B${rke2_release}/rke2-images.linux-amd64.tar.zst",
|
||||
require => [
|
||||
Package["rke2-${node_type}"],
|
||||
File['/var/lib/rancher/rke2/agent/images'],
|
||||
],
|
||||
before => Service["rke2-${node_type}"],
|
||||
}
|
||||
|
||||
# ensure the images cache file exists
|
||||
file {'/var/lib/rancher/rke2/agent/images/.cache.json':
|
||||
ensure => file,
|
||||
owner => 'root',
|
||||
group => 'root',
|
||||
mode => '0644',
|
||||
require => [
|
||||
Package["rke2-${node_type}"],
|
||||
File['/var/lib/rancher/rke2/agent/images'],
|
||||
],
|
||||
before => Service["rke2-${node_type}"],
|
||||
}
|
||||
}
|
||||
15
modules/rke2/manifests/params.pp
Normal file
15
modules/rke2/manifests/params.pp
Normal file
@ -0,0 +1,15 @@
|
||||
# rke2 params
|
||||
class rke2::params (
|
||||
Enum['server', 'agent'] $node_type = 'agent',
|
||||
String $rke2_version = '1.33.4',
|
||||
String $rke2_release = 'rke2r1',
|
||||
Stdlib::Absolutepath $config_file = '/etc/rancher/rke2/config.yaml',
|
||||
Hash $config_hash = {},
|
||||
Stdlib::HTTPSUrl $join_url = 'https://127.0.0.1:9345',
|
||||
Stdlib::Fqdn $bootstrap_node = 'localhost.localdomain',
|
||||
String $node_token = '',
|
||||
Boolean $helm_install = false,
|
||||
Hash $helm_repos = {},
|
||||
Array[String[1]] $extra_config_files = [],
|
||||
Stdlib::HTTPUrl $container_archive_source = 'https://github.com/rancher/rke2/releases/download',
|
||||
) {}
|
||||
13
modules/rke2/manifests/service.pp
Normal file
13
modules/rke2/manifests/service.pp
Normal file
@ -0,0 +1,13 @@
|
||||
# manage rke2 service
|
||||
class rke2::service (
|
||||
Enum['server', 'agent'] $node_type = $rke2::node_type,
|
||||
Stdlib::Absolutepath $config_file = $rke2::config_file,
|
||||
){
|
||||
|
||||
service {"rke2-${node_type}":
|
||||
ensure => true,
|
||||
enable => true,
|
||||
subscribe => File[$config_file],
|
||||
}
|
||||
|
||||
}
|
||||
230
modules/stalwart/README.md
Normal file
230
modules/stalwart/README.md
Normal file
@ -0,0 +1,230 @@
|
||||
# Stalwart Mail Server Module
|
||||
|
||||
This Puppet module manages Stalwart Mail Server, a modern, secure, and scalable mail server implementation that supports IMAP, JMAP, WebDAV, and SMTP protocols.
|
||||
|
||||
## Overview
|
||||
|
||||
The `stalwart` module provides a comprehensive solution for deploying Stalwart Mail Server in a clustered environment with:
|
||||
|
||||
- **PostgreSQL backend** for data, full-text search, and in-memory storage
|
||||
- **S3/Ceph-RGW backend** for blob storage (emails, attachments, sieve scripts)
|
||||
- **Automatic cluster discovery** using `query_nodes()`
|
||||
- **DNS autodiscovery records** for email client configuration
|
||||
- **TLS certificate management** integration
|
||||
- **Postfix relay integration** for SMTP routing
|
||||
|
||||
## Features
|
||||
|
||||
- ✅ **Multi-node clustering** with peer-to-peer coordination
|
||||
- ✅ **PostgreSQL authentication** with SQL directory backend
|
||||
- ✅ **S3 blob storage** with compression support
|
||||
- ✅ **IMAP/IMAPS protocols** for email access
|
||||
- ✅ **HTTP/HTTPS protocols** for JMAP, WebDAV, and autodiscovery
|
||||
- ✅ **SMTP relay** for postfix integration
|
||||
- ✅ **DNS autodiscovery** record management
|
||||
- ✅ **Automatic role distribution** across cluster nodes
|
||||
- ✅ **TLS security** with Vault PKI integration
|
||||
|
||||
## Requirements
|
||||
|
||||
- **Puppet 6+** with `query_nodes()` function support
|
||||
- **Stalwart RPM package** (creates user, directories, systemd service)
|
||||
- **PostgreSQL cluster** for data storage
|
||||
- **S3-compatible storage** (Ceph-RGW, MinIO, AWS S3)
|
||||
- **DNS management** via `profiles::dns::record`
|
||||
- **PKI management** via `profiles::pki::vault::alt_names`
|
||||
|
||||
## Usage
|
||||
|
||||
### Recommended Usage with Role
|
||||
|
||||
The recommended way to use this module is via the `roles::infra::mail::backend` role with hieradata configuration:
|
||||
|
||||
```puppet
|
||||
include roles::infra::mail::backend
|
||||
```
|
||||
|
||||
Configure all parameters in `hieradata/roles/infra/mail/backend.yaml` - see `examples/role-hieradata.yaml` for a complete example.
|
||||
|
||||
### Direct Class Usage
|
||||
|
||||
```puppet
|
||||
class { 'stalwart':
|
||||
node_id => 1,
|
||||
cluster_role => 'mail-backend',
|
||||
postgresql_host => 'pgsql.example.com',
|
||||
postgresql_database => 'stalwart',
|
||||
postgresql_user => 'stalwart',
|
||||
postgresql_password => Sensitive('secretpassword'),
|
||||
s3_endpoint => 'https://ceph-rgw.example.com',
|
||||
s3_bucket => 'stalwart-blobs',
|
||||
s3_access_key => 'accesskey',
|
||||
s3_secret_key => Sensitive('secretkey'),
|
||||
domains => ['example.com'],
|
||||
postfix_relay_host => 'postfix.example.com',
|
||||
}
|
||||
```
|
||||
|
||||
## Hieradata Configuration
|
||||
|
||||
See `examples/role-hieradata.yaml` for a complete example of role-based hieradata configuration.
|
||||
|
||||
### Required Parameters
|
||||
|
||||
```yaml
|
||||
# Cluster role for node discovery
|
||||
stalwart::cluster_role: 'mail-backend'
|
||||
|
||||
# Optional: Unique node identifier (auto-calculated if not specified)
|
||||
# stalwart::node_id: 1
|
||||
|
||||
# PostgreSQL connection
|
||||
stalwart::postgresql_host: 'pgsql.example.com'
|
||||
stalwart::postgresql_database: 'stalwart'
|
||||
stalwart::postgresql_user: 'stalwart'
|
||||
stalwart::postgresql_password: >
|
||||
ENC[PKCS7,encrypted_password...]
|
||||
|
||||
# S3/Ceph-RGW connection
|
||||
stalwart::s3_endpoint: 'https://ceph-rgw.example.com'
|
||||
stalwart::s3_bucket: 'stalwart-blobs'
|
||||
stalwart::s3_access_key: 'access_key'
|
||||
stalwart::s3_secret_key: >
|
||||
ENC[PKCS7,encrypted_secret...]
|
||||
|
||||
# Domains and relay
|
||||
stalwart::domains:
|
||||
- 'example.com'
|
||||
stalwart::postfix_relay_host: 'postfix.example.com'
|
||||
```
|
||||
|
||||
## Architecture
|
||||
|
||||
### Cluster Setup
|
||||
|
||||
The module automatically discovers cluster members using `query_nodes()` based on:
|
||||
- `enc_role` matching `cluster_role` parameter
|
||||
- `country` fact matching the node's country fact
|
||||
- `region` fact matching the node's region fact
|
||||
|
||||
**Node ID Assignment:**
|
||||
- Node IDs are **automatically extracted** from the last 4 digits of the hostname
|
||||
- Example: `ausyd1nxvm1234` → node ID `1234`
|
||||
- Manual override available via `stalwart::node_id` parameter if needed
|
||||
- Hostname must end with 4 digits for automatic extraction to work
|
||||
- Ensures unique IDs when following consistent hostname patterns
|
||||
|
||||
### Storage Layout
|
||||
|
||||
- **Data Store**: PostgreSQL (metadata, folders, settings)
|
||||
- **Full-Text Search**: PostgreSQL (search indexes)
|
||||
- **In-Memory Store**: PostgreSQL (caching, sessions)
|
||||
- **Blob Store**: S3/Ceph-RGW (emails, attachments, files)
|
||||
|
||||
### Directory Structure (Created by RPM)
|
||||
|
||||
- **Config**: `/opt/stalwart/etc/config.toml`
|
||||
- **Data**: `/var/lib/stalwart/` (queue, reports)
|
||||
- **Logs**: `/var/log/stalwart/stalwart.log`
|
||||
- **Binary**: `/opt/stalwart/bin/stalwart`
|
||||
- **User**: `stalwart:stalwart` (system user)
|
||||
|
||||
### Network Ports
|
||||
|
||||
- **143**: IMAP (STARTTLS)
|
||||
- **993**: IMAPS (implicit TLS)
|
||||
- **443**: HTTPS (JMAP, WebDAV, autodiscovery)
|
||||
- **2525**: SMTP relay (postfix communication)
|
||||
- **11200**: Cluster coordination (peer-to-peer)
|
||||
- **9090**: Prometheus metrics
|
||||
|
||||
### DNS Records
|
||||
|
||||
When `manage_dns_records: true`, the module creates:
|
||||
- `autoconfig.domain.com` → server FQDN (Thunderbird)
|
||||
- `autodiscover.domain.com` → server FQDN (Outlook)
|
||||
- `_imap._tcp.domain.com` SRV record
|
||||
- `_imaps._tcp.domain.com` SRV record
|
||||
- `_caldav._tcp.domain.com` SRV record
|
||||
- `_carddav._tcp.domain.com` SRV record
|
||||
|
||||
## PostgreSQL Schema
|
||||
|
||||
The module expects these tables in the PostgreSQL database:
|
||||
|
||||
```sql
|
||||
CREATE TABLE accounts (
|
||||
name TEXT PRIMARY KEY,
|
||||
secret TEXT,
|
||||
description TEXT,
|
||||
type TEXT NOT NULL,
|
||||
quota INTEGER DEFAULT 0,
|
||||
active BOOLEAN DEFAULT true
|
||||
);
|
||||
|
||||
CREATE TABLE group_members (
|
||||
name TEXT NOT NULL,
|
||||
member_of TEXT NOT NULL,
|
||||
PRIMARY KEY (name, member_of)
|
||||
);
|
||||
|
||||
CREATE TABLE emails (
|
||||
name TEXT NOT NULL,
|
||||
address TEXT NOT NULL,
|
||||
type TEXT,
|
||||
PRIMARY KEY (name, address)
|
||||
);
|
||||
```
|
||||
|
||||
## Security
|
||||
|
||||
- **TLS required** for all connections
|
||||
- **PostgreSQL SSL** enabled by default
|
||||
- **S3 HTTPS** endpoints required
|
||||
- **Password hashing** supported (SHA512, BCRYPT, etc.)
|
||||
- **Certificate management** via Vault PKI
|
||||
|
||||
### Fallback Administrator
|
||||
|
||||
Stalwart includes a fallback administrator account for initial setup and emergency access:
|
||||
|
||||
- **Default username**: `admin` (configurable via `stalwart::fallback_admin_user`)
|
||||
- **Default password**: `admin` (configurable via `stalwart::fallback_admin_password`)
|
||||
- **Purpose**: Initial server configuration and emergency access when directory services are unavailable
|
||||
- **Security**: Password is automatically hashed using SHA-512 crypt format
|
||||
|
||||
**Important**: Change the default password in production by setting different hieradata values:
|
||||
|
||||
```yaml
|
||||
stalwart::fallback_admin_password: "your-secure-password"
|
||||
```
|
||||
|
||||
The fallback admin should only be used for initial setup and emergencies. Create regular admin accounts in PostgreSQL for day-to-day management.
|
||||
|
||||
## Monitoring
|
||||
|
||||
- **Prometheus metrics** on port 9090
|
||||
- **Log files** in `/var/log/stalwart/`
|
||||
- **Queue monitoring** in `/var/lib/stalwart/queue/`
|
||||
- **Service status** via systemd (`stalwart.service`)
|
||||
|
||||
## Troubleshooting
|
||||
|
||||
### Cluster Formation Issues
|
||||
- Verify `query_nodes()` returns expected nodes
|
||||
- Check `country` and `region` facts are consistent
|
||||
- Ensure `cluster_role` matches across all nodes
|
||||
|
||||
### Storage Connection Issues
|
||||
- Test PostgreSQL connectivity and credentials
|
||||
- Verify S3 endpoint accessibility and credentials
|
||||
- Check network connectivity between nodes
|
||||
|
||||
### TLS Certificate Issues
|
||||
- Ensure PKI alt_names include all required domains
|
||||
- Verify certificate paths exist and are readable
|
||||
- Check certificate expiration dates
|
||||
|
||||
## License
|
||||
|
||||
This module is part of the internal infrastructure management system.
|
||||
57
modules/stalwart/examples/hieradata.yaml
Normal file
57
modules/stalwart/examples/hieradata.yaml
Normal file
@ -0,0 +1,57 @@
|
||||
# Example hieradata for profiles::mail::stalwart
|
||||
# This shows the required and optional parameters for Stalwart configuration
|
||||
|
||||
# Required: Unique node ID for each server in the cluster (1, 2, 3, etc.)
|
||||
profiles::mail::stalwart::node_id: 1
|
||||
|
||||
# Required: Cluster role name for query_nodes() discovery
|
||||
profiles::mail::stalwart::cluster_role: 'mail-backend'
|
||||
|
||||
# Required: PostgreSQL connection settings
|
||||
profiles::mail::stalwart::postgresql_host: 'pgsql.example.com'
|
||||
profiles::mail::stalwart::postgresql_port: 5432
|
||||
profiles::mail::stalwart::postgresql_database: 'stalwart'
|
||||
profiles::mail::stalwart::postgresql_user: 'stalwart'
|
||||
profiles::mail::stalwart::postgresql_password: >
|
||||
ENC[PKCS7,MIIBiQYJKoZIhvcNAQcDoIIBejCCAXYCAQAxggEhMIIBHQIBADAFMAACAQEwDQYJKoZIhvcNAQEBBQAEggEAxample...]
|
||||
profiles::mail::stalwart::postgresql_ssl: true
|
||||
|
||||
# Required: S3/Ceph-RGW connection settings
|
||||
profiles::mail::stalwart::s3_endpoint: 'https://ceph-rgw.example.com'
|
||||
profiles::mail::stalwart::s3_bucket: 'stalwart-blobs'
|
||||
profiles::mail::stalwart::s3_region: 'default'
|
||||
profiles::mail::stalwart::s3_access_key: 'stalwart_access_key'
|
||||
profiles::mail::stalwart::s3_secret_key: >
|
||||
ENC[PKCS7,MIIBiQYJKoZIhvcNAQcDoIIBejCCAXYCAQAxggEhMIIBHQIBADAFMAACAQEwDQYJKoZIhvcNAQEBBQAEggEAxample...]
|
||||
profiles::mail::stalwart::s3_key_prefix: 'stalwart/'
|
||||
|
||||
# Required: Domains this mail backend serves
|
||||
profiles::mail::stalwart::domains:
|
||||
- 'example.com'
|
||||
- 'mail.example.com'
|
||||
|
||||
# Required: Postfix relay host for SMTP delivery
|
||||
profiles::mail::stalwart::postfix_relay_host: 'postfix.example.com'
|
||||
|
||||
# Optional: Protocol configuration (defaults shown)
|
||||
profiles::mail::stalwart::enable_imap: true
|
||||
profiles::mail::stalwart::enable_imap_tls: true
|
||||
profiles::mail::stalwart::enable_http: true
|
||||
profiles::mail::stalwart::enable_smtp_relay: true
|
||||
|
||||
# Optional: Management settings
|
||||
profiles::mail::stalwart::manage_dns_records: true
|
||||
profiles::mail::stalwart::log_level: 'info'
|
||||
|
||||
# Optional: TLS certificate paths (defaults shown)
|
||||
profiles::mail::stalwart::tls_cert: '/etc/pki/tls/vault/certificate.crt'
|
||||
profiles::mail::stalwart::tls_key: '/etc/pki/tls/vault/private.key'
|
||||
|
||||
# Example PKI alt_names configuration for TLS certificates
|
||||
# This should include all domains and hostnames that need certificates
|
||||
profiles::pki::vault::alt_names:
|
||||
mail-backend:
|
||||
- 'imap.example.com'
|
||||
- 'mail.example.com'
|
||||
- 'autoconfig.example.com'
|
||||
- 'autodiscover.example.com'
|
||||
58
modules/stalwart/examples/role-hieradata.yaml
Normal file
58
modules/stalwart/examples/role-hieradata.yaml
Normal file
@ -0,0 +1,58 @@
|
||||
# Example hieradata for roles::infra::mail::backend
|
||||
# Place this in: hieradata/roles/infra/mail/backend.yaml
|
||||
|
||||
# Stalwart module configuration - all parameters passed directly to the module
|
||||
# stalwart::node_id: 1234 # Optional - automatically extracted from last 4 digits of hostname
|
||||
stalwart::cluster_role: 'mail-backend'
|
||||
|
||||
# PostgreSQL connection settings
|
||||
stalwart::postgresql_host: 'pgsql.example.com'
|
||||
stalwart::postgresql_port: 5432
|
||||
stalwart::postgresql_database: 'stalwart'
|
||||
stalwart::postgresql_user: 'stalwart'
|
||||
stalwart::postgresql_password: >
|
||||
ENC[PKCS7,MIIBiQYJKoZIhvcNAQcDoIIBejCCAXYCAQAxggEhMIIBHQIBADAFMAACAQEwDQYJKoZIhvcNAQEBBQAEggEAxample...]
|
||||
stalwart::postgresql_ssl: true
|
||||
|
||||
# S3/Ceph-RGW connection settings
|
||||
stalwart::s3_endpoint: 'https://ceph-rgw.example.com'
|
||||
stalwart::s3_bucket: 'stalwart-blobs'
|
||||
stalwart::s3_region: 'default'
|
||||
stalwart::s3_access_key: 'stalwart_access_key'
|
||||
stalwart::s3_secret_key: >
|
||||
ENC[PKCS7,MIIBiQYJKoZIhvcNAQcDoIIBejCCAXYCAQAxggEhMIIBHQIBADAFMAACAQEwDQYJKoZIhvcNAQEBBQAEggEAxample...]
|
||||
stalwart::s3_key_prefix: 'stalwart/'
|
||||
|
||||
# Domains this mail backend serves
|
||||
stalwart::domains:
|
||||
- 'example.com'
|
||||
- 'mail.example.com'
|
||||
|
||||
# Postfix relay host for SMTP delivery
|
||||
stalwart::postfix_relay_host: 'postfix.example.com'
|
||||
|
||||
# Optional protocol configuration (defaults shown)
|
||||
stalwart::enable_imap: true
|
||||
stalwart::enable_imap_tls: true
|
||||
stalwart::enable_http: true
|
||||
stalwart::enable_smtp_relay: true
|
||||
|
||||
# Optional management settings
|
||||
stalwart::manage_dns_records: true
|
||||
stalwart::log_level: 'info'
|
||||
|
||||
# Optional TLS certificate paths (defaults work with profiles::pki::vault)
|
||||
# stalwart::tls_cert: '/etc/pki/tls/vault/certificate.crt'
|
||||
# stalwart::tls_key: '/etc/pki/tls/vault/private.key'
|
||||
|
||||
# Optional path overrides (RPM package sets up these defaults)
|
||||
# stalwart::config_dir: '/opt/stalwart/etc'
|
||||
# stalwart::data_dir: '/var/lib/stalwart'
|
||||
|
||||
# PKI alt_names configuration for TLS certificates
|
||||
# This should include all domains and hostnames that need certificates
|
||||
profiles::pki::vault::alt_names:
|
||||
- 'imap.example.com'
|
||||
- 'mail.example.com'
|
||||
- 'autoconfig.example.com'
|
||||
- 'autodiscover.example.com'
|
||||
84
modules/stalwart/manifests/config.pp
Normal file
84
modules/stalwart/manifests/config.pp
Normal file
@ -0,0 +1,84 @@
|
||||
# @summary Manages Stalwart Mail Server configuration
|
||||
#
|
||||
# @api private
|
||||
class stalwart::config {
|
||||
assert_private()
|
||||
|
||||
# Create base directories (package creates user/group and base dirs)
|
||||
file { [$stalwart::config_dir, $stalwart::data_dir, $stalwart::webadmin_unpack_path]:
|
||||
ensure => directory,
|
||||
owner => 'stalwart',
|
||||
group => 'stalwart',
|
||||
mode => '0750',
|
||||
}
|
||||
|
||||
# Ensure log directory exists
|
||||
file { '/var/log/stalwart':
|
||||
ensure => directory,
|
||||
owner => 'stalwart',
|
||||
group => 'stalwart',
|
||||
mode => '0755',
|
||||
}
|
||||
|
||||
# Main configuration file
|
||||
file { "${stalwart::config_dir}/config.toml":
|
||||
ensure => file,
|
||||
owner => 'stalwart',
|
||||
group => 'stalwart',
|
||||
mode => '0640',
|
||||
content => epp('stalwart/config.toml.epp', {
|
||||
'cluster_size' => $stalwart::cluster_size,
|
||||
'other_cluster_members' => $stalwart::other_cluster_members,
|
||||
'haproxy_ips' => $stalwart::haproxy_ips,
|
||||
'effective_node_id' => $stalwart::effective_node_id,
|
||||
'bind_address' => $stalwart::bind_address,
|
||||
'advertise_address' => $stalwart::advertise_address,
|
||||
'postgresql_host' => $stalwart::postgresql_host,
|
||||
'postgresql_port' => $stalwart::postgresql_port,
|
||||
'postgresql_database' => $stalwart::postgresql_database,
|
||||
'postgresql_user' => $stalwart::postgresql_user,
|
||||
'postgresql_password' => $stalwart::postgresql_password.unwrap,
|
||||
'postgresql_ssl' => $stalwart::postgresql_ssl,
|
||||
's3_endpoint' => $stalwart::s3_endpoint,
|
||||
's3_bucket' => $stalwart::s3_bucket,
|
||||
's3_region' => $stalwart::s3_region,
|
||||
's3_access_key' => $stalwart::s3_access_key,
|
||||
's3_secret_key' => $stalwart::s3_secret_key.unwrap,
|
||||
's3_key_prefix' => $stalwart::s3_key_prefix,
|
||||
'domains' => $stalwart::domains,
|
||||
'postfix_relay_host' => $stalwart::postfix_relay_host,
|
||||
'enable_imap' => $stalwart::enable_imap,
|
||||
'enable_imap_tls' => $stalwart::enable_imap_tls,
|
||||
'enable_http' => $stalwart::enable_http,
|
||||
'enable_smtp_submission' => $stalwart::enable_smtp_submission,
|
||||
'data_dir' => $stalwart::data_dir,
|
||||
'tls_cert' => $stalwart::tls_cert,
|
||||
'tls_key' => $stalwart::tls_key,
|
||||
'log_level' => $stalwart::log_level,
|
||||
'service_hostname' => $stalwart::service_hostname,
|
||||
'fallback_admin_user' => $stalwart::fallback_admin_user,
|
||||
'fallback_admin_password' => $stalwart::fallback_admin_password,
|
||||
'webadmin_unpack_path' => $stalwart::webadmin_unpack_path,
|
||||
'webadmin_resource_url' => $stalwart::webadmin_resource_url,
|
||||
'webadmin_auto_update' => $stalwart::webadmin_auto_update,
|
||||
'node_facts' => $facts,
|
||||
}),
|
||||
notify => Service['stalwart'],
|
||||
}
|
||||
|
||||
# Create directories for storage
|
||||
file { "${stalwart::data_dir}/queue":
|
||||
ensure => directory,
|
||||
owner => 'stalwart',
|
||||
group => 'stalwart',
|
||||
mode => '0750',
|
||||
}
|
||||
|
||||
file { "${stalwart::data_dir}/reports":
|
||||
ensure => directory,
|
||||
owner => 'stalwart',
|
||||
group => 'stalwart',
|
||||
mode => '0750',
|
||||
}
|
||||
|
||||
}
|
||||
67
modules/stalwart/manifests/dns.pp
Normal file
67
modules/stalwart/manifests/dns.pp
Normal file
@ -0,0 +1,67 @@
|
||||
# @summary Manages DNS autodiscovery records for Stalwart
|
||||
#
|
||||
# @param target_host
|
||||
# FQDN to point DNS records to (defaults to current server)
|
||||
#
|
||||
# @api private
|
||||
class stalwart::dns (
|
||||
Stdlib::Fqdn $target_host = $facts['networking']['fqdn'],
|
||||
) {
|
||||
assert_private()
|
||||
|
||||
# Create autodiscovery DNS records for each domain
|
||||
$stalwart::domains.each |$domain| {
|
||||
|
||||
# Autoconfig record for Thunderbird/Mozilla clients
|
||||
profiles::dns::record { "autoconfig_${domain}":
|
||||
record => "autoconfig.${domain}",
|
||||
type => 'CNAME',
|
||||
value => "${target_host}.",
|
||||
zone => $domain,
|
||||
order => 100,
|
||||
}
|
||||
|
||||
# Autodiscover record for Outlook/Microsoft clients
|
||||
profiles::dns::record { "autodiscover_${domain}":
|
||||
record => "autodiscover.${domain}",
|
||||
type => 'CNAME',
|
||||
value => "${target_host}.",
|
||||
zone => $domain,
|
||||
order => 101,
|
||||
}
|
||||
|
||||
# IMAP SRV records
|
||||
profiles::dns::record { "imap_srv_${domain}":
|
||||
record => "_imap._tcp.${domain}",
|
||||
type => 'SRV',
|
||||
value => "10 1 143 ${target_host}.",
|
||||
zone => $domain,
|
||||
order => 102,
|
||||
}
|
||||
|
||||
profiles::dns::record { "imaps_srv_${domain}":
|
||||
record => "_imaps._tcp.${domain}",
|
||||
type => 'SRV',
|
||||
value => "10 1 993 ${target_host}.",
|
||||
zone => $domain,
|
||||
order => 103,
|
||||
}
|
||||
|
||||
# CalDAV and CardDAV SRV records
|
||||
profiles::dns::record { "caldav_srv_${domain}":
|
||||
record => "_caldav._tcp.${domain}",
|
||||
type => 'SRV',
|
||||
value => "10 1 443 ${target_host}.",
|
||||
zone => $domain,
|
||||
order => 104,
|
||||
}
|
||||
|
||||
profiles::dns::record { "carddav_srv_${domain}":
|
||||
record => "_carddav._tcp.${domain}",
|
||||
type => 'SRV',
|
||||
value => "10 1 443 ${target_host}.",
|
||||
zone => $domain,
|
||||
order => 105,
|
||||
}
|
||||
}
|
||||
}
|
||||
245
modules/stalwart/manifests/init.pp
Normal file
245
modules/stalwart/manifests/init.pp
Normal file
@ -0,0 +1,245 @@
|
||||
# @summary Main class for managing Stalwart Mail Server
|
||||
#
|
||||
# This class provides a comprehensive setup of Stalwart Mail Server with
|
||||
# clustering, authentication, storage, and protocol support.
|
||||
#
|
||||
# @example Basic Stalwart setup
|
||||
# class { 'stalwart':
|
||||
# node_id => 1,
|
||||
# postgresql_host => 'pgsql.example.com',
|
||||
# postgresql_database => 'stalwart',
|
||||
# postgresql_user => 'stalwart',
|
||||
# postgresql_password => Sensitive('secretpassword'),
|
||||
# s3_endpoint => 'https://ceph-rgw.example.com',
|
||||
# s3_bucket => 'stalwart-blobs',
|
||||
# s3_access_key => 'accesskey',
|
||||
# s3_secret_key => Sensitive('secretkey'),
|
||||
# domains => ['example.com'],
|
||||
# postfix_relay_host => 'postfix.example.com',
|
||||
# }
|
||||
#
|
||||
# @param node_id
|
||||
# Unique identifier for this node in the cluster (1-N). If not specified,
|
||||
# automatically calculated based on sorted position in cluster member list.
|
||||
#
|
||||
# @param cluster_role
|
||||
# Role name for cluster member discovery via query_nodes()
|
||||
#
|
||||
#
|
||||
# @param postgresql_host
|
||||
# PostgreSQL server hostname/IP
|
||||
#
|
||||
# @param postgresql_port
|
||||
# PostgreSQL server port
|
||||
#
|
||||
# @param postgresql_database
|
||||
# PostgreSQL database name
|
||||
#
|
||||
# @param postgresql_user
|
||||
# PostgreSQL username
|
||||
#
|
||||
# @param postgresql_password
|
||||
# PostgreSQL password (Sensitive)
|
||||
#
|
||||
# @param postgresql_ssl
|
||||
# Enable SSL/TLS for PostgreSQL connections
|
||||
#
|
||||
# @param s3_endpoint
|
||||
# S3/Ceph-RGW endpoint URL
|
||||
#
|
||||
# @param s3_bucket
|
||||
# S3 bucket name for blob storage
|
||||
#
|
||||
# @param s3_region
|
||||
# S3 region
|
||||
#
|
||||
# @param s3_access_key
|
||||
# S3 access key
|
||||
#
|
||||
# @param s3_secret_key
|
||||
# S3 secret key (Sensitive)
|
||||
#
|
||||
# @param s3_key_prefix
|
||||
# S3 key prefix for stalwart objects
|
||||
#
|
||||
# @param domains
|
||||
# Array of domains this server handles
|
||||
#
|
||||
# @param postfix_relay_host
|
||||
# Postfix relay host for SMTP delivery
|
||||
#
|
||||
# @param bind_address
|
||||
# IP address to bind services to
|
||||
#
|
||||
# @param advertise_address
|
||||
# IP address to advertise to cluster members
|
||||
#
|
||||
# @param enable_imap
|
||||
# Enable IMAP protocol listener
|
||||
#
|
||||
# @param enable_imap_tls
|
||||
# Enable IMAP over TLS listener
|
||||
#
|
||||
# @param enable_http
|
||||
# Enable HTTP listener for JMAP/WebDAV/Autodiscovery
|
||||
#
|
||||
# @param enable_smtp_relay
|
||||
# Enable SMTP for postfix relay communication
|
||||
#
|
||||
# @param enable_smtp_submission
|
||||
# Enable SMTP submission listener on port 587
|
||||
#
|
||||
# @param haproxy_role
|
||||
# Role name for HAProxy nodes to include in proxy trusted networks
|
||||
#
|
||||
# @param service_hostname
|
||||
# Service hostname used for autoconfig/autodiscover and SMTP greeting
|
||||
#
|
||||
# @param package_ensure
|
||||
# Package version to install
|
||||
#
|
||||
# @param config_dir
|
||||
# Stalwart configuration directory
|
||||
#
|
||||
# @param data_dir
|
||||
# Stalwart data directory
|
||||
#
|
||||
# @param log_level
|
||||
# Logging verbosity level
|
||||
#
|
||||
# @param manage_firewall
|
||||
# Whether to manage firewall rules
|
||||
#
|
||||
# @param tls_cert
|
||||
# Path to TLS certificate file
|
||||
#
|
||||
# @param tls_key
|
||||
# Path to TLS private key file
|
||||
#
|
||||
# @param manage_dns_records
|
||||
# Whether to create DNS autodiscovery records
|
||||
#
|
||||
class stalwart (
|
||||
String $cluster_role,
|
||||
Stdlib::Host $postgresql_host,
|
||||
String $postgresql_database,
|
||||
String $postgresql_user,
|
||||
Sensitive[String] $postgresql_password,
|
||||
Stdlib::HTTPUrl $s3_endpoint,
|
||||
String $s3_bucket,
|
||||
String $s3_access_key,
|
||||
Sensitive[String] $s3_secret_key,
|
||||
Array[Stdlib::Fqdn] $domains,
|
||||
Stdlib::Host $postfix_relay_host,
|
||||
Optional[Integer] $node_id = undef,
|
||||
Stdlib::Port $postgresql_port = 5432,
|
||||
Boolean $postgresql_ssl = true,
|
||||
String $s3_region = 'us-east-1',
|
||||
String $s3_key_prefix = 'stalwart/',
|
||||
Stdlib::IP::Address $bind_address = $facts['networking']['ip'],
|
||||
Stdlib::IP::Address $advertise_address = $facts['networking']['ip'],
|
||||
Boolean $enable_imap = true,
|
||||
Boolean $enable_imap_tls = true,
|
||||
Boolean $enable_http = true,
|
||||
Boolean $enable_smtp_relay = true,
|
||||
Boolean $enable_smtp_submission = true,
|
||||
String $haproxy_role = 'roles::infra::halb::haproxy2',
|
||||
Stdlib::Fqdn $service_hostname = $facts['networking']['fqdn'],
|
||||
String $package_ensure = 'present',
|
||||
Stdlib::Absolutepath $config_dir = '/opt/stalwart/etc',
|
||||
Stdlib::Absolutepath $data_dir = '/var/lib/stalwart',
|
||||
Enum['error','warn','info','debug','trace'] $log_level = 'info',
|
||||
Boolean $manage_firewall = false,
|
||||
Stdlib::Absolutepath $tls_cert = '/etc/pki/tls/vault/certificate.crt',
|
||||
Stdlib::Absolutepath $tls_key = '/etc/pki/tls/vault/private.key',
|
||||
Boolean $manage_dns_records = true,
|
||||
Optional[Stdlib::Fqdn] $loadbalancer_host = undef,
|
||||
String $fallback_admin_user = 'admin',
|
||||
Sensitive[String] $fallback_admin_password = Sensitive('admin'),
|
||||
Stdlib::Absolutepath $webadmin_unpack_path = "${data_dir}/webadmin",
|
||||
Stdlib::HTTPUrl $webadmin_resource_url = 'https://github.com/stalwartlabs/webadmin/releases/latest/download/webadmin.zip',
|
||||
Boolean $webadmin_auto_update = true,
|
||||
) {
|
||||
|
||||
# Calculate node_id from last 4 digits of hostname if not provided
|
||||
$my_fqdn = $facts['networking']['fqdn']
|
||||
$hostname = $facts['networking']['hostname']
|
||||
|
||||
# Query cluster members for validation
|
||||
$cluster_query = "enc_role='${cluster_role}' and country='${facts['country']}' and region='${facts['region']}'"
|
||||
$cluster_members_raw = query_nodes($cluster_query, 'networking.fqdn')
|
||||
$cluster_members = $cluster_members_raw ? {
|
||||
undef => [],
|
||||
default => $cluster_members_raw,
|
||||
}
|
||||
$sorted_cluster_members = sort($cluster_members)
|
||||
|
||||
# Calculate cluster information for templates
|
||||
$other_cluster_members = $sorted_cluster_members.filter |$member| { $member != $my_fqdn }
|
||||
$cluster_size = length($sorted_cluster_members)
|
||||
|
||||
# Query HAProxy nodes for proxy trusted networks
|
||||
$haproxy_query = "enc_role='${haproxy_role}' and country='${facts['country']}' and region='${facts['region']}'"
|
||||
$haproxy_members_raw = query_nodes($haproxy_query, 'networking.ip')
|
||||
$haproxy_ips = $haproxy_members_raw ? {
|
||||
undef => [],
|
||||
default => sort($haproxy_members_raw),
|
||||
}
|
||||
|
||||
# Extract last 4 digits from hostname (e.g., ausyd1nxvm1234 -> 1234)
|
||||
if $hostname =~ /^.*(\d{4})$/ {
|
||||
$hostname_digits = $1
|
||||
$calculated_node_id = Integer($hostname_digits)
|
||||
} else {
|
||||
fail("Unable to extract 4-digit node ID from hostname '${hostname}'. Hostname must end with 4 digits or specify node_id manually.")
|
||||
}
|
||||
|
||||
# Use provided node_id or calculated one
|
||||
$effective_node_id = $node_id ? {
|
||||
undef => $calculated_node_id,
|
||||
default => $node_id,
|
||||
}
|
||||
|
||||
# Validate parameters
|
||||
if $effective_node_id < 1 {
|
||||
fail('node_id must be a positive integer')
|
||||
}
|
||||
|
||||
if empty($domains) {
|
||||
fail('At least one domain must be specified')
|
||||
}
|
||||
|
||||
if !($my_fqdn in $sorted_cluster_members) {
|
||||
fail("This node (${my_fqdn}) is not found in cluster members for role '${cluster_role}' in ${facts['country']}-${facts['region']}")
|
||||
}
|
||||
|
||||
|
||||
# Include sub-classes in dependency order
|
||||
include stalwart::install
|
||||
include stalwart::config
|
||||
include stalwart::service
|
||||
|
||||
# Handle DNS records if requested
|
||||
if $manage_dns_records {
|
||||
if $loadbalancer_host {
|
||||
# Only first node in cluster creates DNS records pointing to load balancer
|
||||
if $my_fqdn == $sorted_cluster_members[0] {
|
||||
class { 'stalwart::dns':
|
||||
target_host => $loadbalancer_host,
|
||||
}
|
||||
}
|
||||
} else {
|
||||
# Current behavior: each server creates its own DNS records
|
||||
include stalwart::dns
|
||||
}
|
||||
}
|
||||
|
||||
# Class ordering
|
||||
Class['stalwart::install']
|
||||
-> Class['stalwart::config']
|
||||
-> Class['stalwart::service']
|
||||
|
||||
if $manage_dns_records {
|
||||
Class['stalwart::service'] -> Class['stalwart::dns']
|
||||
}
|
||||
}
|
||||
11
modules/stalwart/manifests/install.pp
Normal file
11
modules/stalwart/manifests/install.pp
Normal file
@ -0,0 +1,11 @@
|
||||
# @summary Manages Stalwart Mail Server package installation
|
||||
#
|
||||
# @api private
|
||||
class stalwart::install {
|
||||
assert_private()
|
||||
|
||||
# Install stalwart package (user/group created by package preinstall script)
|
||||
package { 'stalwart':
|
||||
ensure => $stalwart::package_ensure,
|
||||
}
|
||||
}
|
||||
26
modules/stalwart/manifests/service.pp
Normal file
26
modules/stalwart/manifests/service.pp
Normal file
@ -0,0 +1,26 @@
|
||||
# @summary Manages Stalwart Mail Server service
|
||||
#
|
||||
# @api private
|
||||
class stalwart::service {
|
||||
assert_private()
|
||||
|
||||
# Service is installed by the RPM package
|
||||
service { 'stalwart':
|
||||
ensure => running,
|
||||
enable => true,
|
||||
subscribe => [
|
||||
File[$stalwart::tls_cert],
|
||||
File[$stalwart::tls_key],
|
||||
],
|
||||
}
|
||||
|
||||
# Add capability to bind to privileged ports (143, 443, 993)
|
||||
systemd::manage_dropin { 'bind-capabilities.conf':
|
||||
ensure => present,
|
||||
unit => 'stalwart.service',
|
||||
service_entry => {
|
||||
'AmbientCapabilities' => 'CAP_NET_BIND_SERVICE',
|
||||
},
|
||||
notify => Service['stalwart'],
|
||||
}
|
||||
}
|
||||
296
modules/stalwart/templates/config.toml.epp
Normal file
296
modules/stalwart/templates/config.toml.epp
Normal file
@ -0,0 +1,296 @@
|
||||
# Stalwart Mail Server Configuration
|
||||
# Generated by Puppet - DO NOT EDIT MANUALLY
|
||||
|
||||
[server]
|
||||
hostname = "<%= $service_hostname %>"
|
||||
greeting = "Stalwart ESMTP"
|
||||
|
||||
[server.listener."smtp-relay"]
|
||||
bind = ["<%= $bind_address %>:25"]
|
||||
protocol = "smtp"
|
||||
greeting = "Stalwart SMTP Relay"
|
||||
|
||||
<% if !$haproxy_ips.empty { -%>
|
||||
[server.listener."smtp-relay".proxy]
|
||||
trusted-networks = ["127.0.0.0/8", "::1"<% $haproxy_ips.each |$ip| { %>, "<%= $ip %>"<% } %>]
|
||||
<% } -%>
|
||||
|
||||
<% if $enable_smtp_submission { -%>
|
||||
[server.listener."submission"]
|
||||
bind = ["<%= $bind_address %>:587"]
|
||||
protocol = "smtp"
|
||||
greeting = "Stalwart SMTP Submission"
|
||||
tls.require = true
|
||||
|
||||
<% if !$haproxy_ips.empty { -%>
|
||||
[server.listener."submission".proxy]
|
||||
trusted-networks = ["127.0.0.0/8", "::1"<% $haproxy_ips.each |$ip| { %>, "<%= $ip %>"<% } %>]
|
||||
<% } -%>
|
||||
<% } -%>
|
||||
|
||||
<% if $enable_imap { -%>
|
||||
[server.listener."imap"]
|
||||
bind = ["<%= $bind_address %>:143"]
|
||||
protocol = "imap"
|
||||
|
||||
<% if !$haproxy_ips.empty { -%>
|
||||
[server.listener."imap".proxy]
|
||||
trusted-networks = ["127.0.0.0/8", "::1"<% $haproxy_ips.each |$ip| { %>, "<%= $ip %>"<% } %>]
|
||||
<% } -%>
|
||||
<% } -%>
|
||||
|
||||
<% if $enable_imap_tls { -%>
|
||||
[server.listener."imaps"]
|
||||
bind = ["<%= $bind_address %>:993"]
|
||||
protocol = "imap"
|
||||
tls.implicit = true
|
||||
|
||||
<% if !$haproxy_ips.empty { -%>
|
||||
[server.listener."imaps".proxy]
|
||||
trusted-networks = ["127.0.0.0/8", "::1"<% $haproxy_ips.each |$ip| { %>, "<%= $ip %>"<% } %>]
|
||||
<% } -%>
|
||||
<% } -%>
|
||||
|
||||
<% if $enable_http { -%>
|
||||
[server.listener."https"]
|
||||
bind = ["<%= $bind_address %>:443"]
|
||||
protocol = "http"
|
||||
tls.implicit = true
|
||||
|
||||
<% if !$haproxy_ips.empty { -%>
|
||||
[server.listener."https".proxy]
|
||||
trusted-networks = ["127.0.0.0/8", "::1"<% $haproxy_ips.each |$ip| { %>, "<%= $ip %>"<% } %>]
|
||||
<% } -%>
|
||||
<% } -%>
|
||||
|
||||
[server.tls]
|
||||
enable = true
|
||||
implicit = false
|
||||
certificate = "default"
|
||||
|
||||
|
||||
[webadmin]
|
||||
path = "<%= $webadmin_unpack_path %>"
|
||||
auto-update = <%= $webadmin_auto_update %>
|
||||
resource = "<%= $webadmin_resource_url %>"
|
||||
|
||||
# Cluster Configuration
|
||||
[cluster]
|
||||
node-id = <%= $effective_node_id %>
|
||||
|
||||
<% if $cluster_size > 1 { -%>
|
||||
# Peer-to-peer coordination
|
||||
[cluster.coordinator]
|
||||
type = "peer-to-peer"
|
||||
addr = "<%= $bind_address %>:11200"
|
||||
advertise-addr = "<%= $advertise_address %>:11200"
|
||||
|
||||
<% $other_cluster_members.each |$node| { -%>
|
||||
[[cluster.coordinator.peers]]
|
||||
addr = "<%= $node %>:11200"
|
||||
<% } -%>
|
||||
|
||||
# Cluster roles for 3-node setup
|
||||
[cluster.roles.purge]
|
||||
stores = ["1", "2", "3"]
|
||||
accounts = ["1", "2"]
|
||||
|
||||
[cluster.roles.acme]
|
||||
renew = ["1"]
|
||||
|
||||
[cluster.roles.metrics]
|
||||
calculate = ["1", "2"]
|
||||
push = ["1"]
|
||||
|
||||
[cluster.roles.push-notifications]
|
||||
push-notifications = ["1", "3"]
|
||||
|
||||
[cluster.roles.fts-indexing]
|
||||
fts-indexing = ["2", "3"]
|
||||
|
||||
[cluster.roles.bayes-training]
|
||||
bayes-training = ["1"]
|
||||
|
||||
[cluster.roles.imip-processing]
|
||||
imip-processing = ["2"]
|
||||
|
||||
[cluster.roles.calendar-alerts]
|
||||
calendar-alerts = ["3"]
|
||||
<% } -%>
|
||||
|
||||
# Storage Configuration
|
||||
|
||||
# PostgreSQL store for data, FTS, and in-memory
|
||||
[store."postgresql"]
|
||||
type = "postgresql"
|
||||
host = "<%= $postgresql_host %>"
|
||||
port = <%= $postgresql_port %>
|
||||
database = "<%= $postgresql_database %>"
|
||||
user = "<%= $postgresql_user %>"
|
||||
password = "<%= $postgresql_password %>"
|
||||
timeout = "15s"
|
||||
|
||||
[store."postgresql".tls]
|
||||
enable = <%= $postgresql_ssl %>
|
||||
allow-invalid-certs = false
|
||||
|
||||
[store."postgresql".pool]
|
||||
max-connections = 10
|
||||
|
||||
[store."postgresql".purge]
|
||||
frequency = "0 3 *"
|
||||
|
||||
# S3/Ceph-RGW store for blobs
|
||||
[store."s3"]
|
||||
type = "s3"
|
||||
bucket = "<%= $s3_bucket %>"
|
||||
region = "<%= $s3_region %>"
|
||||
access-key = "<%= $s3_access_key %>"
|
||||
secret-key = "<%= $s3_secret_key %>"
|
||||
endpoint = "<%= $s3_endpoint %>"
|
||||
timeout = "30s"
|
||||
key-prefix = "<%= $s3_key_prefix %>"
|
||||
compression = "lz4"
|
||||
|
||||
[store."s3".purge]
|
||||
frequency = "30 5 *"
|
||||
|
||||
# Storage assignment
|
||||
[storage]
|
||||
data = "postgresql"
|
||||
fts = "postgresql"
|
||||
blob = "s3"
|
||||
lookup = "postgresql"
|
||||
directory = "internal"
|
||||
in-memory = "postgresql"
|
||||
|
||||
# Directory configuration
|
||||
[directory.internal]
|
||||
type = "internal"
|
||||
store = "postgresql"
|
||||
|
||||
# Authentication configuration
|
||||
[authentication.fallback-admin]
|
||||
user = "<%= $fallback_admin_user %>"
|
||||
secret = "<%= pw_hash($fallback_admin_password.unwrap, 'SHA-512', 'stalwart') %>"
|
||||
|
||||
[authentication]
|
||||
[authentication.directory]
|
||||
directories = ["internal"]
|
||||
|
||||
# Authorization configuration
|
||||
[authorization]
|
||||
directory = "internal"
|
||||
|
||||
# JMAP configuration
|
||||
[jmap]
|
||||
directory = "internal"
|
||||
|
||||
[jmap.protocol]
|
||||
request-max-size = 10485760
|
||||
get.max-objects = 500
|
||||
query.max-results = 5000
|
||||
changes.max-results = 5000
|
||||
upload.max-size = 50000000
|
||||
upload.ttl = "1h"
|
||||
|
||||
# IMAP configuration
|
||||
[imap]
|
||||
directory = "internal"
|
||||
|
||||
[imap.protocol]
|
||||
max-requests = 64
|
||||
|
||||
# Inbound rate limiting
|
||||
[[queue.limiter.inbound]]
|
||||
key = ["remote_ip"]
|
||||
rate = "500/1s"
|
||||
enable = true
|
||||
|
||||
# SMTP configuration for postfix relay
|
||||
[session.data]
|
||||
pipe.command = "sendmail"
|
||||
pipe.arguments = ["-i", "-f", "{sender}", "{recipient}"]
|
||||
|
||||
# Outbound SMTP configuration
|
||||
[queue]
|
||||
path = "<%= $data_dir %>/queue"
|
||||
|
||||
[queue.schedule]
|
||||
retry = ["2s", "5s", "1m", "5m", "15m", "30m", "1h", "2h"]
|
||||
notify = ["1d", "3d"]
|
||||
expire = "5d"
|
||||
|
||||
[session.extensions]
|
||||
future-release = "7d"
|
||||
|
||||
# Relay configuration for postfix
|
||||
[remote."postfix"]
|
||||
address = "<%= $postfix_relay_host %>"
|
||||
port = 25
|
||||
protocol = "smtp"
|
||||
|
||||
# HTTP configuration
|
||||
[server.http]
|
||||
use-x-forwarded = false
|
||||
permissive-cors = false
|
||||
|
||||
# Disable spam filtering (handled by postfix)
|
||||
[session.ehlo]
|
||||
reject-non-fqdn = false
|
||||
|
||||
[session.rcpt]
|
||||
type = "internal"
|
||||
store = "postgresql"
|
||||
max-recipients = 25
|
||||
|
||||
[session.data]
|
||||
max-messages = 10
|
||||
max-message-size = 52428800
|
||||
|
||||
# TLS configuration
|
||||
[certificate."default"]
|
||||
cert = "%{file:<%= $tls_cert %>}%"
|
||||
private-key = "%{file:<%= $tls_key %>}%"
|
||||
default = true
|
||||
|
||||
# Logging configuration
|
||||
[tracer]
|
||||
type = "log"
|
||||
level = "<%= $log_level %>"
|
||||
ansi = false
|
||||
multiline = true
|
||||
|
||||
[tracer.file]
|
||||
path = "/var/log/stalwart/stalwart.log"
|
||||
rotate = "daily"
|
||||
keep = 30
|
||||
|
||||
# Report storage
|
||||
[report]
|
||||
path = "<%= $data_dir %>/reports"
|
||||
hash = "sha256"
|
||||
encrypt = false
|
||||
|
||||
# Metrics configuration
|
||||
[metrics]
|
||||
prometheus.enable = true
|
||||
prometheus.port = 9090
|
||||
|
||||
# Queue routing configuration
|
||||
[queue.strategy]
|
||||
route = [ { if = "is_local_domain('', rcpt_domain)", then = "'local'" },
|
||||
{ else = "'relay'" } ]
|
||||
|
||||
[queue.route."local"]
|
||||
type = "local"
|
||||
|
||||
[queue.route."relay"]
|
||||
type = "relay"
|
||||
address = "<%= $postfix_relay_host %>"
|
||||
port = 25
|
||||
protocol = "smtp"
|
||||
|
||||
[queue.route."relay".tls]
|
||||
implicit = false
|
||||
allow-invalid-certs = false
|
||||
@ -10,6 +10,7 @@ class vmcluster::vmagent (
|
||||
Stdlib::Absolutepath $vars_file = '/etc/default/vmagent',
|
||||
String $consul_node_token = $facts['consul_node_token'],
|
||||
Hash[String, Variant[String, Array[String]]] $options = {},
|
||||
Hash[String, Hash] $static_targets = {},
|
||||
) {
|
||||
|
||||
# if enabled, manage this service
|
||||
|
||||
@ -35,3 +35,28 @@ scrape_configs:
|
||||
- source_labels: [__meta_consul_tag_metrics_job]
|
||||
target_label: job
|
||||
action: replace
|
||||
|
||||
<% if @static_targets -%>
|
||||
<% @static_targets.each do |job_name, config| -%>
|
||||
- job_name: '<%= job_name %>'
|
||||
static_configs:
|
||||
<% config['targets'].each do |target| -%>
|
||||
- targets: ['<%= target %>']
|
||||
<% if config['labels'] -%>
|
||||
labels:
|
||||
<% config['labels'].each do |label_name, label_value| -%>
|
||||
<%= label_name %>: '<%= label_value %>'
|
||||
<% end -%>
|
||||
<% end -%>
|
||||
<% end -%>
|
||||
<% if config['scrape_interval'] -%>
|
||||
scrape_interval: <%= config['scrape_interval'] %>
|
||||
<% end -%>
|
||||
<% if config['metrics_path'] -%>
|
||||
metrics_path: <%= config['metrics_path'] %>
|
||||
<% end -%>
|
||||
<% if config['scheme'] -%>
|
||||
scheme: <%= config['scheme'] %>
|
||||
<% end -%>
|
||||
<% end -%>
|
||||
<% end -%>
|
||||
|
||||
19
modules/zfs/lib/facter/zfs_datasets.rb
Normal file
19
modules/zfs/lib/facter/zfs_datasets.rb
Normal file
@ -0,0 +1,19 @@
|
||||
# frozen_string_literal: true
|
||||
|
||||
Facter.add(:zfs_datasets) do
|
||||
confine kernel: 'Linux'
|
||||
setcode do
|
||||
datasets = []
|
||||
|
||||
if Facter::Core::Execution.which('zfs')
|
||||
begin
|
||||
output = Facter::Core::Execution.execute('zfs list -H -o name 2>/dev/null', on_fail: nil)
|
||||
datasets = output.strip.split("\n") if output && !output.empty?
|
||||
rescue StandardError => e
|
||||
Facter.debug("Error getting zfs dataset information: #{e.message}")
|
||||
end
|
||||
end
|
||||
|
||||
datasets.empty? ? nil : datasets
|
||||
end
|
||||
end
|
||||
@ -3,7 +3,8 @@
|
||||
Facter.add('zfs_zpool_cache_present') do
|
||||
confine kernel: 'Linux'
|
||||
setcode do
|
||||
File.exist?('/etc/zfs/zpool.cache')
|
||||
cache_file = '/etc/zfs/zpool.cache'
|
||||
File.exist?(cache_file) && File.size(cache_file).positive?
|
||||
end
|
||||
end
|
||||
|
||||
|
||||
19
modules/zfs/lib/facter/zfs_zpools.rb
Normal file
19
modules/zfs/lib/facter/zfs_zpools.rb
Normal file
@ -0,0 +1,19 @@
|
||||
# frozen_string_literal: true
|
||||
|
||||
Facter.add(:zfs_zpools) do
|
||||
confine kernel: 'Linux'
|
||||
setcode do
|
||||
zpools = []
|
||||
|
||||
if Facter::Core::Execution.which('zpool')
|
||||
begin
|
||||
output = Facter::Core::Execution.execute('zpool list -H -o name 2>/dev/null', on_fail: nil)
|
||||
zpools = output.strip.split("\n") if output && !output.empty?
|
||||
rescue StandardError => e
|
||||
Facter.debug("Error getting zpool information: #{e.message}")
|
||||
end
|
||||
end
|
||||
|
||||
zpools.empty? ? nil : zpools
|
||||
end
|
||||
end
|
||||
@ -38,10 +38,13 @@ class zfs (
|
||||
|
||||
# create zpools
|
||||
$zpools.each | $zpool, $data | {
|
||||
# Only create zpool if it doesn't already exist
|
||||
if $facts['zfs_zpools'] == undef or !($zpool in $facts['zfs_zpools']) {
|
||||
zpool { $zpool:
|
||||
* => $data
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
# create datasets
|
||||
$datasets.each | $dataset, $data | {
|
||||
|
||||
54
site/profiles/files/postfix/gateway/aliases
Normal file
54
site/profiles/files/postfix/gateway/aliases
Normal file
@ -0,0 +1,54 @@
|
||||
# FILE MANAGED BY PUPPET, CHANGES WILL BE REPLACED
|
||||
|
||||
postmaster: root
|
||||
|
||||
# Many mailers use this address to represent the empty SMTP return path
|
||||
MAILER-DAEMON: postmaster
|
||||
|
||||
# Common aliases for system accounts.
|
||||
bin: root
|
||||
daemon: root
|
||||
games: root
|
||||
ingres: root
|
||||
nobody: root
|
||||
system: root
|
||||
toor: root
|
||||
foo: root
|
||||
falken: root
|
||||
|
||||
# Well-known aliases.
|
||||
admin: root
|
||||
manager: root
|
||||
dumper: root
|
||||
operator: root
|
||||
|
||||
# traps to catch security attacks
|
||||
decode: root
|
||||
moof: root
|
||||
moog: root
|
||||
|
||||
# Standard aliases also defined by RFC 2142
|
||||
abuse: postmaster
|
||||
|
||||
# reports of network infrastructure difficulties
|
||||
noc: root
|
||||
|
||||
# address to report secuirty problems
|
||||
security: root
|
||||
|
||||
# DNS administrator (DNS soa records should use this)
|
||||
hostmaster: root
|
||||
|
||||
# Usenet news service administrator
|
||||
news: usenet
|
||||
usenet: root
|
||||
|
||||
# http/web service administrator
|
||||
www: webmaster
|
||||
webmaster: root
|
||||
|
||||
# UUCP service administrator
|
||||
uucp: root
|
||||
|
||||
# FTP administrator (especially anon FTP)
|
||||
ftp: root
|
||||
@ -1,12 +1,12 @@
|
||||
# manage the root user
|
||||
class profiles::accounts::root (
|
||||
String $password,
|
||||
Optional[Array[String]] $sshkeys = undef,
|
||||
) {
|
||||
|
||||
if $sshkeys {
|
||||
accounts::user { 'root':
|
||||
sshkeys => $sshkeys,
|
||||
}
|
||||
password => $password,
|
||||
}
|
||||
|
||||
file {'/root/.config':
|
||||
|
||||
@ -1,7 +1,5 @@
|
||||
# this is the base class, which will be used by all servers
|
||||
class profiles::base (
|
||||
Array $puppet_servers,
|
||||
) {
|
||||
class profiles::base () {
|
||||
|
||||
# run a limited set of classes on the first run aimed at bootstrapping the new node
|
||||
if $facts['firstrun'] {
|
||||
@ -13,11 +11,7 @@ class profiles::base (
|
||||
|
||||
# manage the puppet agent
|
||||
include profiles::puppet::agent
|
||||
|
||||
# manage puppet clients
|
||||
if ! member($puppet_servers, $trusted['certname']) {
|
||||
include profiles::puppet::client
|
||||
}
|
||||
|
||||
# include the base profiles
|
||||
include profiles::base::repos
|
||||
|
||||
36
site/profiles/manifests/ceph/conf.pp
Normal file
36
site/profiles/manifests/ceph/conf.pp
Normal file
@ -0,0 +1,36 @@
|
||||
class profiles::ceph::conf (
|
||||
Hash $config = {}
|
||||
) {
|
||||
|
||||
package {[
|
||||
'ceph',
|
||||
'ceph-common'
|
||||
]:
|
||||
ensure => installed,
|
||||
}
|
||||
|
||||
file {'/etc/ceph':
|
||||
ensure => directory,
|
||||
owner => 'ceph',
|
||||
group => 'ceph',
|
||||
mode => '0755',
|
||||
require => Package['ceph'],
|
||||
}
|
||||
|
||||
file {'/var/log/ceph':
|
||||
ensure => directory,
|
||||
owner => 'ceph',
|
||||
group => 'ceph',
|
||||
mode => '0755',
|
||||
require => Package['ceph'],
|
||||
}
|
||||
|
||||
file { '/etc/ceph/ceph.conf':
|
||||
ensure => file,
|
||||
owner => 'ceph',
|
||||
group => 'ceph',
|
||||
mode => '0644',
|
||||
content => template('profiles/ceph/conf.erb'),
|
||||
require => Package['ceph-common'],
|
||||
}
|
||||
}
|
||||
Some files were not shown because too many files have changed in this diff Show More
Loading…
Reference in New Issue
Block a user