From 3ec5ce616edbcc8db43e9436bce75e2663cbc337 Mon Sep 17 00:00:00 2001 From: Jonas Forsberg Date: Tue, 27 Jun 2023 09:16:37 +0200 Subject: [PATCH] first commit --- common/users.sls | 38 + containers/alertmanager.sls | 26 + containers/fail2ban_exporter.sls | 3 + containers/files/99-rfxcom-serial.rules | 1 + containers/files/99-zigbee-serial.rules | 1 + .../files/alertmanager/alertmanager.yml | 16 + .../files/check_image_updates.service.jinja | 10 + containers/files/check_image_updates.sh.jinja | 35 + .../files/check_image_updates.timer.jinja | 9 + containers/files/container.sh.jinja | 144 ++++ containers/files/env_file.jinja | 3 + containers/files/loki-config.yaml | 40 + containers/files/mysql-dump.service.jinja | 12 + containers/files/mysql-dump.sh.jinja | 15 + containers/files/mysql-dump.timer.jinja | 9 + containers/files/npm-container.conf | 1 + containers/files/prometheus/alert.node.yml | 292 ++++++++ containers/files/prometheus/prometheus.yml | 59 ++ containers/files/promtail.conf.jinja | 29 + containers/files/unpoller.conf | 234 ++++++ containers/freeipa.sls | 17 + containers/gitea.sls | 10 + containers/gotify.sls | 10 + containers/grafana.sls | 10 + containers/init.sls | 46 ++ containers/loki.sls | 32 + containers/mariadb.sls | 10 + containers/mosquitto.sls | 17 + containers/nextcloud.sls | 11 + containers/node_exporter.sls | 3 + containers/nodered.sls | 25 + containers/npm.sls | 24 + containers/piwigo.sls | 24 + containers/prometheus.sls | 26 + containers/promtail.sls | 34 + containers/salt.sls | 3 + containers/unifi.sls | 10 + containers/unpoller.sls | 20 + containers/zigbee2mqtt.sls | 25 + dnsmasq/dnsmasq.conf | 696 ++++++++++++++++++ dnsmasq/dnsmasq.d/rre.conf | 23 + dnsmasq/hosts/hosts | 53 ++ lib.sls | 102 +++ pod/dnsmasq.sls | 21 + pod/files/99-rfxcom-serial.rules | 1 + pod/files/99-zigbee-serial.rules | 1 + pod/files/check_image_updates.service.jinja | 10 + pod/files/check_image_updates.sh.jinja | 35 + pod/files/check_image_updates.timer.jinja | 9 + pod/files/pod.sh.jinja | 158 ++++ pod/gotify.sls | 11 + pod/init.sls | 46 ++ pod/loki.sls | 24 + pod/mariadb.sls | 11 + pod/mosquitto.sls | 17 + pod/nextcloud.sls | 10 + pod/nodered.sls | 25 + pod/piwigo.sls | 25 + pod/prometheus.sls | 47 ++ pod/unifi.sls | 10 + restic/files/backup.sh.jinja | 99 +++ restic/files/mysql-backup.sh | 16 + restic/files/restic-backup.service.jinja | 11 + restic/files/restic-backup.timer.jinja | 11 + restic/files/restic.automount.jinja | 8 + restic/files/restic.jinja | 23 + restic/files/restic.mount.jinja | 11 + restic/init.sls | 77 ++ restic/repository.sls | 39 + top.sls | 8 + wifi-voucher/files/voucher.env.jinja | 3 + wifi-voucher/files/wifi-voucher.service | 13 + wifi-voucher/init.sls | 4 + wifi-voucher/packages.sls | 9 + wifi-voucher/remove.sls | 4 + wifi-voucher/service.sls | 31 + wifi-voucher/user.sls | 8 + 77 files changed, 3044 insertions(+) create mode 100644 common/users.sls create mode 100644 containers/alertmanager.sls create mode 100644 containers/fail2ban_exporter.sls create mode 100644 containers/files/99-rfxcom-serial.rules create mode 100644 containers/files/99-zigbee-serial.rules create mode 100644 containers/files/alertmanager/alertmanager.yml create mode 100644 containers/files/check_image_updates.service.jinja create mode 100644 containers/files/check_image_updates.sh.jinja create mode 100644 containers/files/check_image_updates.timer.jinja create mode 100644 containers/files/container.sh.jinja create mode 100644 containers/files/env_file.jinja create mode 100644 containers/files/loki-config.yaml create mode 100644 containers/files/mysql-dump.service.jinja create mode 100644 containers/files/mysql-dump.sh.jinja create mode 100644 containers/files/mysql-dump.timer.jinja create mode 100644 containers/files/npm-container.conf create mode 100644 containers/files/prometheus/alert.node.yml create mode 100644 containers/files/prometheus/prometheus.yml create mode 100644 containers/files/promtail.conf.jinja create mode 100644 containers/files/unpoller.conf create mode 100644 containers/freeipa.sls create mode 100644 containers/gitea.sls create mode 100644 containers/gotify.sls create mode 100644 containers/grafana.sls create mode 100644 containers/init.sls create mode 100644 containers/loki.sls create mode 100644 containers/mariadb.sls create mode 100644 containers/mosquitto.sls create mode 100644 containers/nextcloud.sls create mode 100644 containers/node_exporter.sls create mode 100644 containers/nodered.sls create mode 100644 containers/npm.sls create mode 100644 containers/piwigo.sls create mode 100644 containers/prometheus.sls create mode 100644 containers/promtail.sls create mode 100644 containers/salt.sls create mode 100644 containers/unifi.sls create mode 100644 containers/unpoller.sls create mode 100644 containers/zigbee2mqtt.sls create mode 100644 dnsmasq/dnsmasq.conf create mode 100644 dnsmasq/dnsmasq.d/rre.conf create mode 100644 dnsmasq/hosts/hosts create mode 100644 lib.sls create mode 100644 pod/dnsmasq.sls create mode 100644 pod/files/99-rfxcom-serial.rules create mode 100644 pod/files/99-zigbee-serial.rules create mode 100644 pod/files/check_image_updates.service.jinja create mode 100644 pod/files/check_image_updates.sh.jinja create mode 100644 pod/files/check_image_updates.timer.jinja create mode 100644 pod/files/pod.sh.jinja create mode 100644 pod/gotify.sls create mode 100644 pod/init.sls create mode 100644 pod/loki.sls create mode 100644 pod/mariadb.sls create mode 100644 pod/mosquitto.sls create mode 100644 pod/nextcloud.sls create mode 100644 pod/nodered.sls create mode 100644 pod/piwigo.sls create mode 100644 pod/prometheus.sls create mode 100644 pod/unifi.sls create mode 100644 restic/files/backup.sh.jinja create mode 100644 restic/files/mysql-backup.sh create mode 100644 restic/files/restic-backup.service.jinja create mode 100644 restic/files/restic-backup.timer.jinja create mode 100644 restic/files/restic.automount.jinja create mode 100644 restic/files/restic.jinja create mode 100644 restic/files/restic.mount.jinja create mode 100644 restic/init.sls create mode 100644 restic/repository.sls create mode 100644 top.sls create mode 100644 wifi-voucher/files/voucher.env.jinja create mode 100644 wifi-voucher/files/wifi-voucher.service create mode 100644 wifi-voucher/init.sls create mode 100644 wifi-voucher/packages.sls create mode 100644 wifi-voucher/remove.sls create mode 100644 wifi-voucher/service.sls create mode 100644 wifi-voucher/user.sls diff --git a/common/users.sls b/common/users.sls new file mode 100644 index 0000000..8e00b7c --- /dev/null +++ b/common/users.sls @@ -0,0 +1,38 @@ +{% for user, args in pillar['users'].items() %} +# remove users +{% if args['remove'] is defined %} +{{ user }}: + user.absent + {% if 'alias_target' in args %} + alias.present: + - target: {{ args['alias_target'] }} + {% endif %} +{% else %} +# add users +{{ user }}: + user.present: + - fullname: {{ args['fullname'] | default('') }} + - home: {{ args['home'] | default('/home/'+user) }} + - shell: {{ args['shell'] | default('/bin/bash') }} + {% if args['uid'] is defined %} + - uid: {{ args['uid'] }} + {% endif %} + - password: {{ args['password'] }} + - enforce_password: {{ args['enforce_password'] | default('True') }} + {% if 'groups' in args %} + - groups: {{ args['groups'] }} + {% endif %} + {% if 'alias_target' in args %} + alias.present: + - target: {{ args['alias_target'] }} + {% endif %} + +{% if 'ssh_auth' in args %} +{{ user }}_autherized_keys: + ssh_auth: + - present + - user: {{ user }} + - names: {{ args['ssh_auth'] }} +{% endif %} +{% endif %} +{% endfor %} diff --git a/containers/alertmanager.sls b/containers/alertmanager.sls new file mode 100644 index 0000000..ca57523 --- /dev/null +++ b/containers/alertmanager.sls @@ -0,0 +1,26 @@ +{% from 'lib.sls' import container_deploy with context %} + +Create alertmanager config directory: + file.recurse: + - name: /etc/alertmanager + - source: salt://containers/files/alertmanager + - user: nobody + - group: root + - dir_mode: "0755" + - file_mode: "0644" + +Create prometheus data directory: + file.directory: + - name: /srv/alertmanager + - user: nobody + - group: nobody + - mode: "0755" + +{{ container_deploy('alertmanager') }} + +Start prometheus service: + service.running: + - name: alertmanager + - enable: True + - watch: + - file: Create alertmanager config directory diff --git a/containers/fail2ban_exporter.sls b/containers/fail2ban_exporter.sls new file mode 100644 index 0000000..f5b26e3 --- /dev/null +++ b/containers/fail2ban_exporter.sls @@ -0,0 +1,3 @@ +{% from 'lib.sls' import container_deploy with context %} + +{{ container_deploy('fail2ban_exporter') }} diff --git a/containers/files/99-rfxcom-serial.rules b/containers/files/99-rfxcom-serial.rules new file mode 100644 index 0000000..8d0fb45 --- /dev/null +++ b/containers/files/99-rfxcom-serial.rules @@ -0,0 +1 @@ +SUBSYSTEM=="tty", ATTRS{idVendor}=="0403", ATTRS{idProduct}=="6001", SYMLINK+="rfxcom", MODE="0666" diff --git a/containers/files/99-zigbee-serial.rules b/containers/files/99-zigbee-serial.rules new file mode 100644 index 0000000..25cf77a --- /dev/null +++ b/containers/files/99-zigbee-serial.rules @@ -0,0 +1 @@ +SUBSYSTEM=="tty", ATTRS{idVendor}=="0451", ATTRS{idProduct}=="16a8", SYMLINK+="zigbee-serial", MODE="0666" diff --git a/containers/files/alertmanager/alertmanager.yml b/containers/files/alertmanager/alertmanager.yml new file mode 100644 index 0000000..7bd5e0c --- /dev/null +++ b/containers/files/alertmanager/alertmanager.yml @@ -0,0 +1,16 @@ +route: + group_by: ['alertname'] + group_wait: 30s + group_interval: 5m + repeat_interval: 1h + receiver: 'web.hook' +receivers: + - name: 'web.hook' + webhook_configs: + - url: 'http://127.0.0.1:5001/' +inhibit_rules: + - source_match: + severity: 'critical' + target_match: + severity: 'warning' + equal: ['alertname', 'dev', 'instance'] diff --git a/containers/files/check_image_updates.service.jinja b/containers/files/check_image_updates.service.jinja new file mode 100644 index 0000000..3a1a788 --- /dev/null +++ b/containers/files/check_image_updates.service.jinja @@ -0,0 +1,10 @@ +[Unit] +Description=Check for image updates on configured podman containers + +[Service] +Type=oneshot +User=root +ExecStart=/root/bin/check_image_updates.sh + +[Install] +WantedBy=default.target diff --git a/containers/files/check_image_updates.sh.jinja b/containers/files/check_image_updates.sh.jinja new file mode 100644 index 0000000..277cdcc --- /dev/null +++ b/containers/files/check_image_updates.sh.jinja @@ -0,0 +1,35 @@ +#!/usr/bin/env bash + +URL="{{ pillar['podman']['gotify']['url'] }}" +TOKEN="{{ pillar['podman']['gotify']['token'] }}" +TITLE="Updates on $HOSTNAME" +PRIORITY="{{ pillar['podman']['gotify']['priority'] }}" + +{% raw -%} +function check_update(){ + IFS=',' read -r -a container_info <<< "$(podman container inspect $1 --format '{{ .Name }},{{ .ImageName }},{{ .Image }}')" + + podman pull "${container_info[1]}" + if [[ "$(podman image inspect "${container_info[1]}" --format "{{.Id}}")" != "${container_info[2]}" ]];then + containers[${#containers[@]}]="${container_info[0]}" + fi +} + + +IFS=$'\n' +for line in $(podman container ls -q); do + check_update "$line" +done +if [[ "${#containers[@]}" == "0" ]]; then + exit +fi + +MESSAGE=$(cat << EOM +Following ${#containers[@]} container(s) has updates: +${containers[*]} +EOM +) + +curl "$URL/message?token=$TOKEN" -F "title=$TITLE" -F "priority=$PRIORITY" -F "message=$MESSAGE" +echo " " +{% endraw -%} diff --git a/containers/files/check_image_updates.timer.jinja b/containers/files/check_image_updates.timer.jinja new file mode 100644 index 0000000..23a5b92 --- /dev/null +++ b/containers/files/check_image_updates.timer.jinja @@ -0,0 +1,9 @@ +[Unit] +Description=Restic backup timer + +[Timer] +OnCalendar=Sun, 12:00 +Unit=check_image_updates.service + +[Install] +WantedBy=timers.target diff --git a/containers/files/container.sh.jinja b/containers/files/container.sh.jinja new file mode 100644 index 0000000..dc2e833 --- /dev/null +++ b/containers/files/container.sh.jinja @@ -0,0 +1,144 @@ +#!/usr/bin/env bash + +function pull_image(){ + if ! podman image exists {{ args['image'] }}:{{ args['tag'] }}; then + podman pull {{ args['image'] }}:{{ args['tag'] }} + fi + +} + +function create_container() { + if ! podman container exists {{ container }};then + podman container create \ + --name {{ container }} \ + {%- if args['podman_options'] is defined %} + {%- for option, value in args['podman_options'].items() %} + --{{ option }} {{ value }} \ + {%- endfor %} + {%- endif %} + {%- if args['volumes'] is defined %} + {%- for volume, mount in args['volumes'].items() %} + -v {{ volume }}:{{ mount }} \ + {%- endfor %} + {%- endif %} + {%- if args['ports'] is defined %} + {%- for ports in args['ports'] %} + -p {{ ports['host'] }}:{{ ports['container'] }}{% if ports['protocol'] is defined %}/{{ ports['protocol'] }}{% endif %} \ + {%- endfor %} + {%- endif %} + {%- if args['env'] is defined %} + {%- for key, value in args['env'].items() %} + -e {{ key }}={{ value }} \ + {%- endfor %} + {%- endif %} + {%- if args['devices'] is defined %} + {%- for key, value in args['devices'].items() %} + --device {{ key }}:{{ value}} \ + {%- endfor %} + {%- endif %} + {{ args['image'] }}:{{ args['tag'] }}{%- if args['run'] is defined %} \ + {{ args['run'] }} + {%- endif %} + fi + +} + +function generate_systemd_unit_file() { + podman generate systemd --name {{ container }} > /etc/systemd/system/{{ container }}.service +} + +function check_update() { + podman pull {{ args['image'] }}:{{ args['tag'] }} + if [[ "$(podman image inspect {{ args['image'] }}:{{ args['tag'] }} --format "{% raw %}{{.Id}}{% endraw %}")" == "$(podman inspect {{ container }} --format "{% raw %}{{ .Image }}{% endraw %}")" ]];then + echo "No image updates available" + return 0 + else + echo "Image update available" + return 1 + fi +} + +function update() { + systemctl stop {{ container }} + podman container rm {{ container }} + create_container + generate_systemd_unit_file + systemctl daemon-reload + systemctl enable --now {{ container }}.service +} + + + +function printHelp(){ + cat << EOF +Usage ${0##*/} [options..] +-h,-?, --help Show help and exit +-p, --pull pull container image ({{ container }}:{{ args['tag'] }}) +-v, --volumes create container volumes +-c, --create create {{ container }} containers +-s, --start start and enables {{ container }} container +-S, --stop stop {{ container }} container +-i, --is-running check to see if container service is running +-u, --check-update check if there are image updates avaiable + --update perform image update if it exists +-g, --generate-systemd generate user systemd service unit file +EOF +} + + +while :; do + case $1 in + -h|-\?|--help) + printHelp + exit + ;; + -p|--pull) + pull_image + shift + ;; + -v|--volumes) + create_volumes + shift + ;; + -c|--create) + create_container + shift + ;; + -s|--start) + systemctl --user enable --now {{ container }}.service + shift + ;; + -S|--stop) + systemctl --user stop {{ container }}.service + shift + ;; + -i|--is-running) + systemctl --user is-active {{ container }}.service + exit $? + shift + ;; + -g|--generate-systemd) + generate_systemd_unit_file + shift + ;; + -u|--check-update) + check_update + shift + ;; + --update) + update + shift + ;; + --) #End of all options + shift + break + ;; + -?*) + printf "'%s' is not a valid option\n" "$1" >&2 + exit 1 + ;; + *) #Break out of case, no more options + break + esac + shift +done diff --git a/containers/files/env_file.jinja b/containers/files/env_file.jinja new file mode 100644 index 0000000..1ade9b7 --- /dev/null +++ b/containers/files/env_file.jinja @@ -0,0 +1,3 @@ +{% for key, value in env_vars.items() -%} +{{ key }}={{ value }} +{% endfor -%} diff --git a/containers/files/loki-config.yaml b/containers/files/loki-config.yaml new file mode 100644 index 0000000..cdf0eb5 --- /dev/null +++ b/containers/files/loki-config.yaml @@ -0,0 +1,40 @@ +auth_enabled: false + +server: + http_listen_port: 3100 + +ingester: + lifecycler: + address: 127.0.0.1 + ring: + kvstore: + store: inmemory + replication_factor: 1 + final_sleep: 0s + chunk_idle_period: 5m + chunk_retain_period: 30s + wal: + dir: /data/wal + +schema_config: + configs: + - from: 2020-05-15 + store: boltdb + object_store: filesystem + schema: v11 + index: + prefix: index_ + period: 168h + +storage_config: + boltdb: + directory: /data/loki/index + + filesystem: + directory: /data/loki/chunks + +limits_config: + enforce_metric_name: false + reject_old_samples: true + reject_old_samples_max_age: 168h + diff --git a/containers/files/mysql-dump.service.jinja b/containers/files/mysql-dump.service.jinja new file mode 100644 index 0000000..ea2e7d1 --- /dev/null +++ b/containers/files/mysql-dump.service.jinja @@ -0,0 +1,12 @@ +{%- set user = salt['pillar.get']('podman:user', 'root') %} +{%- set home = salt['user.info'](user).home %} +[Unit] +Description=Dump all mariadb databases + +[Service] +Type=oneshot +User={{ user }} +ExecStart={{ home }}/bin/mysql-dump.sh + +[Install] +WantedBy=default.target diff --git a/containers/files/mysql-dump.sh.jinja b/containers/files/mysql-dump.sh.jinja new file mode 100644 index 0000000..17a0063 --- /dev/null +++ b/containers/files/mysql-dump.sh.jinja @@ -0,0 +1,15 @@ +#!/bin/bash +umask 0077 +BACKUP_DIR={{ pillar.containers.mariadb.backup_dir }} +databases=$(podman exec -it mariadb mysql -B -u root -p{{ pillar.containers.mariadb.env.MYSQL_ROOT_PASSWORD }} -e "SHOW DATABASES;" | tr -d "| " | grep -v Database) + +for db in ${databases[@]}; do + db=${db::-1} + if [[ "$db" != "information_schema" ]] && [[ "$db" != "performance_schema" ]] && [[ "$db" != "mysql" ]] && [[ "$db" != _* ]] && [[ "$db" != "sys" ]]; then + echo "Dumping database: $db" + podman exec -it mariadb mysqldump -u root -p{{ pillar.containers.mariadb.env.MYSQL_ROOT_PASSWORD }} --databases $db | gzip > ${BACKUP_DIR}/$(date +"%Y-%m-%d_%H-%M-%S")_$db-sql.gz + fi +done +# Delete the files older than 3 days +find $BACKUP_DIR/* -type f -name *-sql.gz -mtime +3 -exec rm {} \; + diff --git a/containers/files/mysql-dump.timer.jinja b/containers/files/mysql-dump.timer.jinja new file mode 100644 index 0000000..62b32f2 --- /dev/null +++ b/containers/files/mysql-dump.timer.jinja @@ -0,0 +1,9 @@ +[Unit] +Description=Restic backup timer + +[Timer] +OnCalendar={{ pillar.containers.mariadb.OnCalendar }} +Unit=mysql-dump.service + +[Install] +WantedBy=timers.target diff --git a/containers/files/npm-container.conf b/containers/files/npm-container.conf new file mode 100644 index 0000000..7d82332 --- /dev/null +++ b/containers/files/npm-container.conf @@ -0,0 +1 @@ +net.ipv4.ip_unprivileged_port_start=80 diff --git a/containers/files/prometheus/alert.node.yml b/containers/files/prometheus/alert.node.yml new file mode 100644 index 0000000..86c1b98 --- /dev/null +++ b/containers/files/prometheus/alert.node.yml @@ -0,0 +1,292 @@ +groups: +- name: node_exporter_alerts + rules: + - alert: Node down + expr: up{job="monitoring-pi"} == 0 + for: 2m + labels: + severity: warning + annotations: + title: Node {{ $labels.instance }} is down + description: Failed to scrape {{ $labels.job }} on {{ $labels.instance }} for more than 2 minutes. Node seems down. + + + - alert: HostOutOfMemory + expr: node_memory_MemAvailable_bytes / node_memory_MemTotal_bytes * 100 < 10 + for: 2m + labels: + severity: warning + annotations: + summary: Host out of memory (instance {{ $labels.instance }}) + description: Node memory is filling up (< 10% left)\n VALUE = {{ $value }} + + - alert: HostMemoryUnderMemoryPressure + expr: rate(node_vmstat_pgmajfault[1m]) > 1000 + for: 2m + labels: + severity: warning + annotations: + summary: Host memory under memory pressure (instance {{ $labels.instance }}) + description: The node is under heavy memory pressure. High rate of major page faults\n VALUE = {{ $value }} + + - alert: HostUnusualNetworkThroughputIn + expr: sum by (instance) (rate(node_network_receive_bytes_total[2m])) / 1024 / 1024 > 100 + for: 5m + labels: + severity: warning + annotations: + summary: Host unusual network throughput in (instance {{ $labels.instance }}) + description: Host network interfaces are probably receiving too much data (> 100 MB/s)\n VALUE = {{ $value }} + + - alert: HostUnusualNetworkThroughputOut + expr: sum by (instance) (rate(node_network_transmit_bytes_total[2m])) / 1024 / 1024 > 100 + for: 5m + labels: + severity: warning + annotations: + summary: Host unusual network throughput out (instance {{ $labels.instance }}) + description: Host network interfaces are probably sending too much data (> 100 MB/s)\n VALUE = {{ $value }} + + - alert: HostUnusualDiskReadRate + expr: sum by (instance) (rate(node_disk_read_bytes_total[2m])) / 1024 / 1024 > 50 + for: 5m + labels: + severity: warning + annotations: + summary: Host unusual disk read rate (instance {{ $labels.instance }}) + description: Disk is probably reading too much data (> 50 MB/s)\n VALUE = {{ $value }} + + - alert: HostUnusualDiskWriteRate + expr: sum by (instance) (rate(node_disk_written_bytes_total[2m])) / 1024 / 1024 > 50 + for: 2m + labels: + severity: warning + annotations: + summary: Host unusual disk write rate (instance {{ $labels.instance }}) + description: Disk is probably writing too much data (> 50 MB/s)\n VALUE = {{ $value }} + + # Please add ignored mountpoints in node_exporter parameters like + # "--collector.filesystem.ignored-mount-points=^/(sys|proc|dev|run)($|/)". + # Same rule using "node_filesystem_free_bytes" will fire when disk fills for non-root users. + - alert: HostOutOfDiskSpace + expr: (node_filesystem_avail_bytes * 100) / node_filesystem_size_bytes < 10 and ON (instance, device, mountpoint) node_filesystem_readonly == 0 + for: 2m + labels: + severity: warning + annotations: + summary: Host out of disk space (instance {{ $labels.instance }}) + description: Disk is almost full (< 10% left)\n VALUE = {{ $value }} + + # Please add ignored mountpoints in node_exporter parameters like + # "--collector.filesystem.ignored-mount-points=^/(sys|proc|dev|run)($|/)". + # Same rule using "node_filesystem_free_bytes" will fire when disk fills for non-root users. + - alert: HostDiskWillFillIn24Hours + expr: (node_filesystem_avail_bytes * 100) / node_filesystem_size_bytes < 10 and ON (instance, device, mountpoint) predict_linear(node_filesystem_avail_bytes{fstype!~"tmpfs"}[1h], 24 * 3600) < 0 and ON (instance, device, mountpoint) node_filesystem_readonly == 0 + for: 2m + labels: + severity: warning + annotations: + summary: Host disk will fill in 24 hours (instance {{ $labels.instance }}) + description: Filesystem is predicted to run out of space within the next 24 hours at current write rate\n VALUE = {{ $value }} + + - alert: HostOutOfInodes + expr: node_filesystem_files_free{mountpoint ="/rootfs"} / node_filesystem_files{mountpoint="/rootfs"} * 100 < 10 and ON (instance, device, mountpoint) node_filesystem_readonly{mountpoint="/rootfs"} == 0 + for: 2m + labels: + severity: warning + annotations: + summary: Host out of inodes (instance {{ $labels.instance }}) + description: Disk is almost running out of available inodes (< 10% left)\n VALUE = {{ $value }} + + - alert: HostInodesWillFillIn24Hours + expr: node_filesystem_files_free{mountpoint ="/rootfs"} / node_filesystem_files{mountpoint="/rootfs"} * 100 < 10 and predict_linear(node_filesystem_files_free{mountpoint="/rootfs"}[1h], 24 * 3600) < 0 and ON (instance, device, mountpoint) node_filesystem_readonly{mountpoint="/rootfs"} == 0 + for: 2m + labels: + severity: warning + annotations: + summary: Host inodes will fill in 24 hours (instance {{ $labels.instance }}) + description: Filesystem is predicted to run out of inodes within the next 24 hours at current write rate\n VALUE = {{ $value }} + + - alert: HostUnusualDiskReadLatency + expr: rate(node_disk_read_time_seconds_total[1m]) / rate(node_disk_reads_completed_total[1m]) > 0.1 and rate(node_disk_reads_completed_total[1m]) > 0 + for: 2m + labels: + severity: warning + annotations: + summary: Host unusual disk read latency (instance {{ $labels.instance }}) + description: Disk latency is growing (read operations > 100ms)\n VALUE = {{ $value }} + + - alert: HostUnusualDiskWriteLatency + expr: rate(node_disk_write_time_seconds_totali{device!~"mmcblk.+"}[1m]) / rate(node_disk_writes_completed_total{device!~"mmcblk.+"}[1m]) > 0.1 and rate(node_disk_writes_completed_total{device!~"mmcblk.+"}[1m]) > 0 + for: 2m + labels: + severity: warning + annotations: + summary: Host unusual disk write latency (instance {{ $labels.instance }}) + description: Disk latency is growing (write operations > 100ms)\n VALUE = {{ $value }} + + - alert: HostHighCpuLoad + expr: 100 - (avg by(instance) (rate(node_cpu_seconds_total{mode="idle"}[2m])) * 100) > 80 + for: 0m + labels: + severity: warning + annotations: + summary: Host high CPU load (instance {{ $labels.instance }}) + description: CPU load is > 80%\n VALUE = {{ $value }} + + - alert: HostCpuStealNoisyNeighbor + expr: avg by(instance) (rate(node_cpu_seconds_total{mode="steal"}[5m])) * 100 > 10 + for: 0m + labels: + severity: warning + annotations: + summary: Host CPU steal noisy neighbor (instance {{ $labels.instance }}) + description: CPU steal is > 10%. A noisy neighbor is killing VM performances or a spot instance may be out of credit.\n VALUE = {{ $value }} + + # 1000 context switches is an arbitrary number. + # Alert threshold depends on nature of application. + # Please read: https://github.com/samber/awesome-prometheus-alerts/issues/58 + - alert: HostContextSwitching + expr: (rate(node_context_switches_total[5m])) / (count without(cpu, mode) (node_cpu_seconds_total{mode="idle"})) > 1000 + for: 0m + labels: + severity: warning + annotations: + summary: Host context switching (instance {{ $labels.instance }}) + description: Context switching is growing on node (> 1000 / s)\n VALUE = {{ $value }} + + - alert: HostSwapIsFillingUp + expr: (1 - (node_memory_SwapFree_bytes / node_memory_SwapTotal_bytes)) * 100 > 80 + for: 2m + labels: + severity: warning + annotations: + summary: Host swap is filling up (instance {{ $labels.instance }}) + description: Swap is filling up (>80%)\n VALUE = {{ $value }} + + - alert: HostSystemdServiceCrashed + expr: node_systemd_unit_state{state="failed"} == 1 + for: 0m + labels: + severity: warning + annotations: + summary: Host SystemD service crashed (instance {{ $labels.instance }}) + description: SystemD service crashed\n VALUE = {{ $value }} + + - alert: HostPhysicalComponentTooHot + expr: node_hwmon_temp_celsius > 75 + for: 5m + labels: + severity: warning + annotations: + summary: Host physical component too hot (instance {{ $labels.instance }}) + description: Physical hardware component too hot\n VALUE = {{ $value }} + + - alert: HostNodeOvertemperatureAlarm + expr: node_hwmon_temp_crit_alarm_celsius == 1 + for: 0m + labels: + severity: critical + annotations: + summary: Host node overtemperature alarm (instance {{ $labels.instance }}) + description: Physical node temperature alarm triggered\n VALUE = {{ $value }} + + - alert: HostRaidArrayGotInactive + expr: node_md_state{state="inactive"} > 0 + for: 0m + labels: + severity: critical + annotations: + summary: Host RAID array got inactive (instance {{ $labels.instance }}) + description: RAID array {{ $labels.device }} is in degraded state due to one or more disks failures. Number of spare drives is insufficient to fix issue automatically.\n VALUE = {{ $value }} + + - alert: HostRaidDiskFailure + expr: node_md_disks{state="failed"} > 0 + for: 2m + labels: + severity: warning + annotations: + summary: Host RAID disk failure (instance {{ $labels.instance }}) + description: At least one device in RAID array on {{ $labels.instance }} failed. Array {{ $labels.md_device }} needs attention and possibly a disk swap\n VALUE = {{ $value }} + + - alert: HostOomKillDetected + expr: increase(node_vmstat_oom_kill[1m]) > 0 + for: 0m + labels: + severity: warning + annotations: + summary: Host OOM kill detected (instance {{ $labels.instance }}) + description: OOM kill detected\n VALUE = {{ $value }} + + - alert: HostEdacCorrectableErrorsDetected + expr: increase(node_edac_correctable_errors_total[1m]) > 0 + for: 0m + labels: + severity: info + annotations: + summary: Host EDAC Correctable Errors detected (instance {{ $labels.instance }}) + description: Instance has had {{ printf "%.0f" $value }} correctable memory errors reported by EDAC in the last 5 minutes.\n VALUE = {{ $value }} + + - alert: HostEdacUncorrectableErrorsDetected + expr: node_edac_uncorrectable_errors_total > 0 + for: 0m + labels: + severity: warning + annotations: + summary: Host EDAC Uncorrectable Errors detected (instance {{ $labels.instance }}) + description: Instance has had {{ printf "%.0f" $value }} uncorrectable memory errors reported by EDAC in the last 5 minutes.\n VALUE = {{ $value }} + + - alert: HostNetworkReceiveErrors + expr: rate(node_network_receive_errs_total[2m]) / rate(node_network_receive_packets_total[2m]) > 0.01 + for: 2m + labels: + severity: warning + annotations: + summary: Host Network Receive Errors (instance {{ $labels.instance }}:{{ $labels.device }}) + description: Instance interface has encountered {{ printf "%.0f" $value }} receive errors in the last five minutes.\n VALUE = {{ $value }} + + - alert: HostNetworkTransmitErrors + expr: rate(node_network_transmit_errs_total[2m]) / rate(node_network_transmit_packets_total[2m]) > 0.01 + for: 2m + labels: + severity: warning + annotations: + summary: Host Network Transmit Errors (instance {{ $labels.instance }}:{{ $labels.device }}) + description: Instance has encountered {{ printf "%.0f" $value }} transmit errors in the last five minutes.\n VALUE = {{ $value }} + + - alert: HostNetworkInterfaceSaturated + expr: (rate(node_network_receive_bytes_total{device!~"^tap.*"}[1m]) + rate(node_network_transmit_bytes_total{device!~"^tap.*"}[1m])) / node_network_speed_bytes{device!~"^tap.*"} > 0.8 + for: 1m + labels: + severity: warning + annotations: + summary: Host Network Interface Saturated (instance {{ $labels.instance }}:{{ $labels.interface }}) + description: The network interface is getting overloaded.\n VALUE = {{ $value }} + + - alert: HostConntrackLimit + expr: node_nf_conntrack_entries / node_nf_conntrack_entries_limit > 0.8 + for: 5m + labels: + severity: warning + annotations: + summary: Host conntrack limit (instance {{ $labels.instance }}) + description: The number of conntrack is approching limit\n VALUE = {{ $value }} + + - alert: HostClockSkew + expr: (node_timex_offset_seconds > 0.05 and deriv(node_timex_offset_seconds[5m]) >= 0) or (node_timex_offset_seconds < -0.05 and deriv(node_timex_offset_seconds[5m]) <= 0) + for: 2m + labels: + severity: warning + annotations: + summary: Host clock skew (instance {{ $labels.instance }}) + description: Clock skew detected. Clock is out of sync.\n VALUE = {{ $value }} + + - alert: HostClockNotSynchronising + expr: min_over_time(node_timex_sync_status[1m]) == 0 and node_timex_maxerror_seconds >= 16 + for: 2m + labels: + severity: warning + annotations: + summary: Host clock not synchronising (instance {{ $labels.instance }}) + description: Clock not synchronising.\n VALUE = {{ $value }} + diff --git a/containers/files/prometheus/prometheus.yml b/containers/files/prometheus/prometheus.yml new file mode 100644 index 0000000..3879875 --- /dev/null +++ b/containers/files/prometheus/prometheus.yml @@ -0,0 +1,59 @@ +# my global config # +global: + scrape_interval: 15s # Set the scrape interval to every 15 seconds. Default is every 1 minute. + evaluation_interval: 15s # Evaluate rules every 15 seconds. The default is every 1 minute. + # scrape_timeout is set to the global default (10s). + +# Alertmanager configuration +alerting: + alertmanagers: + - static_configs: + - targets: + - 10.2.0.22:9093 + +# Load rules once and periodically evaluate them according to the global 'evaluation_interval'. +rule_files: + - "alert.node.yml" + # - "first_rules.yml" + # - "second_rules.yml" + +# A scrape configuration containing exactly one endpoint to scrape: +# Here it's Prometheus itself. +scrape_configs: + # The job name is added as a label `job=` to any timeseries scraped from this config. + - job_name: "prometheus" + + # metrics_path defaults to '/metrics' + # scheme defaults to 'http'. + + static_configs: + - targets: ["localhost:9090"] + - job_name: "node" + static_configs: + - targets: + - "poblano.rre.nu:9100" + - "salt.rre.nu:9100" + - "pepper.rre.nu:9100" + - "woody.rre.nu:9100" + - "serrano.rre.nu:9100" + - "coronado.rre.nu:9100" + - job_name: "unpoller" + static_configs: + - targets: + - "unpoller.rre.nu:9130" + - job_name: "fail2ban" + static_configs: + - targets: + - "poblano.rre.nu:9191" + - "salt.rre.nu:9191" + - "pepper.rre.nu:9191" + - job_name: "nginx" + static_configs: + - targets: + - "10.2.0.22:9193" + + - job_name: "promtail" + static_configs: + - targets: + - "serrano.rre.nu:9080" + - "coronado.rre.nu:9080" diff --git a/containers/files/promtail.conf.jinja b/containers/files/promtail.conf.jinja new file mode 100644 index 0000000..874be59 --- /dev/null +++ b/containers/files/promtail.conf.jinja @@ -0,0 +1,29 @@ +server: + http_listen_port: {{ http_listen_port }} + grpc_listen_port: 0 + +positions: + filename: /tmp/positions.yaml + +clients: + - url: "{{ client_url }}" + +scrape_configs: + - job_name: journal + journal: + max_age: 12h + path: /var/log/journal + labels: + job: systemd-journal + relabel_configs: + - source_labels: ['__journal__systemd_unit'] + target_label: 'unit' + - source_labels: ["__journal__hostname"] + target_label: host + - source_labels: ["__journal_priority_keyword"] + target_label: level + - source_labels: ["__journal_syslog_identifier"] + target_label: syslog_identifier + - source_labels: ["__journal_container_name"] + target_label: container_name + diff --git a/containers/files/unpoller.conf b/containers/files/unpoller.conf new file mode 100644 index 0000000..38a759a --- /dev/null +++ b/containers/files/unpoller.conf @@ -0,0 +1,234 @@ +# Unpoller v2 primary configuration file. TOML FORMAT # +########################################################### + +[poller] + # Turns on line numbers, microsecond logging, and a per-device log. + # The default is false, but I personally leave this on at home (four devices). + # This may be noisy if you have a lot of devices. It adds one line per device. + debug = false + + # Turns off per-interval logs. Only startup and error logs will be emitted. + # Recommend enabling debug with this setting for better error logging. + quiet = false + + # Load dynamic plugins. Advanced use; only sample mysql plugin provided by default. + plugins = [] + +#### OUTPUTS + + # If you don't use an output, you can disable it. + +[prometheus] + disable = false + # This controls on which ip and port /metrics is exported when mode is "prometheus". + # This has no effect in other modes. Must contain a colon and port. + http_listen = "0.0.0.0:9130" + # Adding an SSL Cert and Cert Key will make Poller listen with SSL/https. + ssl_cert_path = "" + ssl_key_path = "" + # Errors are rare. Setting this to true will report them to Prometheus. + report_errors = false + ## Record data for disabled or down (unlinked) switch ports. + dead_ports = false + +[influxdb] + disable = true + # InfluxDB does not require auth by default, so the user/password are probably unimportant. + url = "http://127.0.0.1:8086" + user = "unifipoller" + # Password for InfluxDB user (above). + # If the password provided here begins with file:// then the password is read in from + # the file path that follows the file:// prefix. ex: file:///etc/influxdb/passwd.file + pass = "unifipoller" + # Be sure to create this database. See the InfluxDB Wiki page for more info. + db = "unifi" + # If your InfluxDB uses a valid SSL cert, set this to true. + verify_ssl = false + # The UniFi Controller only updates traffic stats about every 30 seconds. + # Setting this to something lower may lead to "zeros" in your data. + # If you're getting zeros now, set this to "1m" + interval = "30s" + ## Record data for disabled or down (unlinked) switch ports. + dead_ports = false + +# To enable output of UniFi Events to Loki, add a URL; it's disabled otherwise. +# User, pass and tenant_id are optional and most folks wont set them. +# Pick which logs you want per-controller in the [unifi.controller] section. +# This is a new feature. Feedback welcome! +[loki] + disable = false + url = "{{ pillar['containers']['unpoller']['loki_url'] }}" + # The rest of this is advanced & optional. See wiki. + user = "" + pass = "" + verify_ssl = false + tenant_id = "" + interval = "2m" + timeout = "10s" + +[datadog] + # How often to poll UniFi and report to Datadog. + interval = "2m" + + # To enable this output plugin + enable = false + + # Datadog Custom Options + + # address to talk to the datadog agent, by default this uses the local statsd UDP interface + # address = "localhost:8125" + + # namespace to prepend to all data, default is no additional prefix. + # namespace = "" + + # tags to append to all data + # tags = [ "customer:abc_corp" ] + + # For more advanced options for very large amount of data collected see the upstream + # github.com/unpoller/unpoller/pkg/datadogunifi repository README. + + +# Unpoller has an optional web server. To turn it on, set enable to true. If you +# wish to use SSL, provide SSL cert and key paths. This interface is currently +# read-only; it just displays information, like logs, devices and clients. +# Notice: Enabling the web server with many sites will increase memory usage. +# This is a new feature and lacks a UI, enabling only recommended for testing. +[webserver] + enable = false + port = 37288 + # The HTML path is different on Windows and BSD/macOS. + html_path = "/usr/lib/unifi-poller/web" + ssl_cert_path = "" + ssl_key_path = "" + # How many events per event group to hold. 200-2000. Use fewer with many sites. + # With 1 site, you'll have a max total of 9 event groups; 1 per plugin, 4 per site. + # Each site adds 1 event group for each of these inputs that is enabled: + # save_ids, save_events, save_anomalies, save_alarms. + max_events = 200 + +# By default the web interface does not require authentication. You can change +# that by adding a username and password hash (or multiple) below. +# To create a hash, run unifi-poller with the -e CLI argument. See Wiki for more! +[webserver.accounts] +# username = "password-hash" +# captain = "$2a$04$mxw6i0LKH6u46oaLK2cq5eCTAAFkfNiRpzNbz.EyvJZZWNa2FzIlS" + +#### INPUTS + +[unifi] + # Setting this to true and providing default credentials allows you to skip + # configuring controllers in this config file. Instead you configure them in + # your prometheus.yml config. Prometheus then sends the controller URL to + # Unpoller when it performs the scrape. This is useful if you have many, + # or changing controllers. See wiki for more. + dynamic = false + +# The following section contains the default credentials/configuration for any +# dynamic controller (see above section), or the primary controller if you do not +# provide one and dynamic is disabled. In other words, you can just add your +# controller here and delete the following section. The internal defaults are +# shown below. Any missing values will assume these displayed defaults. + +[unifi.defaults] + # URL for the UniFi Controller. Do not add any paths after the host:port. + # Do not use port 8443 if you have a UDM; just use "https://ip". + url = "{{ pillar['containers']['unpoller']['unifi_url'] }}" + + # Make a read-only user in the UniFi Admin Settings, allow it access to all sites. + user = "{{ pillar['containers']['unpoller']['unifi_user'] }}" + + # Password for UniFi controller user (above). + # If the password provided here begins with file:// then the password is read in from + # the file path that follows the file:// prefix. ex: file:///etc/unifi/password.file + # ex: file:///etc/unifi/passwd.file, windows: file://C:\\UserData\\Unifi\\Passwd.txt + pass = "{{ pillar['containers']['unpoller']['unifi_pass'] }}" + + # If the controller has more than one site, specify which sites to poll here. + # Set this to ["default"] to poll only the first site on the controller. + # A setting of ["all"] will poll all sites; this works if you only have 1 site too. + sites = ["all"] + + # Specify a timeout, leave missing to declare infinite wait. This determines the maximum + # time to wait for a response from the unifi controller on any API request. + # timeout = 60s + + # Enable collection of site data. This data powers the Network Sites dashboard. + # It's not valuable to everyone and setting this to false will save resources. + save_sites = true + + # Hash, with md5, client names and MAC addresses. This attempts to protect + # personally identifiable information. Most users won't want to enable this. + hash_pii = false + + # Enable collection of Intrusion Detection System Data (InfluxDB/Loki only). + # Only useful if IDS or IPS are enabled on one of the sites. This may store + # a lot of information. Only recommended for testing and debugging. There + # may not be any dashboards to display this data. It can be used for annotations. + # Enable this only if using InfluxDB or Loki. This will leak PII data! + save_ids = false + + # Enable collection of UniFi Events (InfluxDB/Loki only). + # This may store a lot of information. Only recommended for testing and debugging. + # There are no dashboards to display this data. It can be used for annotations. + # This is a new (June, 2020) feature. Please provide feedback if you try it out! + # Enable this only if using InfluxDB or Loki. This will leak PII data! + save_events = true + + # Enable collection of UniFi Alarms (InfluxDB/Loki only). + # There are no dashboards to display this data. It can be used for annotations. + # This is a new (June, 2020) feature. Please provide feedback if you try it out! + # Enable this only if using InfluxDB or Loki. This will leak PII data! + save_alarms = true + + # Enable collection of UniFi Anomalies (InfluxDB/Loki only). + # There are no dashboards to display this data. It can be used for annotations. + # This is a new (June, 2020) feature. Please provide feedback if you try it out! + # Enable this only if using InfluxDB or Loki. + save_anomalies = true + + # Enable collection of Deep Packet Inspection data. This data breaks down traffic + # types for each client and site, it powers a dedicated DPI dashboard. + # Enabling this adds roughly 150 data points per client. That's 6000 metrics for + # 40 clients. This adds a little bit of poller run time per interval and causes + # more API requests to your controller(s). Don't let these "cons" sway you: + # it's cool data. Please provide feedback on your experience with this feature. + save_dpi = false + + ## Enabling save_rogue stores even more data in your time series databases. + ## This saves neighboring access point metrics in a dedicated table or namespace. + save_rogue = false + + # If your UniFi controller has a valid SSL certificate (like lets encrypt), + # you can enable this option to validate it. Otherwise, any SSL certificate is + # valid. If you don't know if you have a valid SSL cert, then you don't have one. + verify_ssl = false + + ## You may provide a list of SSL cert files (PEM format) that you expect your + ## controller to use. As long as one of the certs you provide here shows up in + ## the cert trust chain the controller presents it will be accepted and allowed. + ## These files may be re-read while poller is running. + ## Example: ssl_cert_paths = ["/path/to/cert.pem", "/another/cert.pem"] + ssl_cert_paths = [] + +# The following is optional and used for configurations with multiple UniFi controllers. + +# You may repeat the following [[unifi.controller]] section as many times as needed to +# poll multiple controllers. Uncomment the entire section including [[unifi.controller]]. +# Omitted variables will have their values taken from the defaults, above. +# +#[[unifi.controller]] +# url = "https://127.0.0.1:8443" +# user = "unifipoller" +# pass = "unifipoller" +# sites = ["all"] +# save_sites = true +# hash_pii = false +# save_ids = false +# save_events = false +# save_alarms = false +# save_anomalies = false +# save_dpi = false +# save_rogue = false +# verify_ssl = false +# ssl_cert_paths = [] + diff --git a/containers/freeipa.sls b/containers/freeipa.sls new file mode 100644 index 0000000..380c0e9 --- /dev/null +++ b/containers/freeipa.sls @@ -0,0 +1,17 @@ +{% from 'lib.sls' import container_deploy with context %} + +Create freeipa root directory: + file.directory: + - name: /srv/freeipa + - user: root + - group: root + - mode: "0755" + +Create freeipa config directory: + file.directory: + - name: /srv/freeipa/data + - user: root + - group: root + - mode: "0755" + +{{ container_deploy('freeipa') }} diff --git a/containers/gitea.sls b/containers/gitea.sls new file mode 100644 index 0000000..bbd4408 --- /dev/null +++ b/containers/gitea.sls @@ -0,0 +1,10 @@ +{% from 'lib.sls' import container_deploy with context %} + +Create gitea data directory: + file.directory: + - name: /srv/gitea + - user: 1000 + - group: 1000 + - mode: "0750" + +{{ container_deploy('gitea') }} diff --git a/containers/gotify.sls b/containers/gotify.sls new file mode 100644 index 0000000..7af0caf --- /dev/null +++ b/containers/gotify.sls @@ -0,0 +1,10 @@ +{% from 'lib.sls' import container_deploy with context %} + +Create gotify data directory: + file.directory: + - name: /srv/gotify + - user: root + - group: root + - mode: "0755" + +{{ container_deploy('gotify') }} diff --git a/containers/grafana.sls b/containers/grafana.sls new file mode 100644 index 0000000..17fdb07 --- /dev/null +++ b/containers/grafana.sls @@ -0,0 +1,10 @@ +{% from 'lib.sls' import container_deploy with context %} + +Create grafana data directory: + file.directory: + - name: /srv/grafana + - user: 472 + - group: root + - mode: "0750" + +{{ container_deploy('grafana') }} diff --git a/containers/init.sls b/containers/init.sls new file mode 100644 index 0000000..35fca2c --- /dev/null +++ b/containers/init.sls @@ -0,0 +1,46 @@ +Copy check image update script: + file.managed: + - name: /root/bin/check_image_updates.sh + - source: salt://containers/files/check_image_updates.sh.jinja + - template: jinja + - user: root + - group: root + - mode: 0700 + +Create check image update service: + file.managed: + - name: /etc/systemd/system/check_image_updates.service + - source: salt://containers/files/check_image_updates.service.jinja + - template: jinja + - user: root + - group: root + - mode: "0644" + +Create check image update timer: + file.managed: + - name: /etc/systemd/system/check_image_updates.timer + - source: salt://containers/files/check_image_updates.timer.jinja + - template: jinja + - user: root + - group: root + - mode: "0644" + +Systemd daemon reload for image update: + cmd.run: + - name: systemctl daemon-reload + - watch: + - file: Create check image update service + - file: Create check image update timer + +Start check image update timer: + service.running: + - name: check_image_updates.timer + - enable: True + +{% if pillar.containers is defined %} +include: +{%- for container, args in pillar.containers.items() %} + - containers.{{ container }} +{%- endfor %} +{%- endif %} + diff --git a/containers/loki.sls b/containers/loki.sls new file mode 100644 index 0000000..ea25faf --- /dev/null +++ b/containers/loki.sls @@ -0,0 +1,32 @@ +{% from 'lib.sls' import container_deploy with context %} + +Create loki data directory: + file.directory: + - name: /srv/loki + - user: 10001 + - group: root + - mode: "0755" + +Create loki config directory: + file.directory: + - name: /etc/loki + - user: 10001 + - group: root + - mode: "0755" + +Create loki configuration file: + file.managed: + - name: /etc/loki/config.yaml + - source: salt://containers/files/loki-config.yaml + - user: root + - group: root + - mode: "0644" + +{{ container_deploy('loki') }} + +Start loki service: + service.running: + - name: loki + - enable: True + - onchanges: + - file: Create loki configuration file diff --git a/containers/mariadb.sls b/containers/mariadb.sls new file mode 100644 index 0000000..053c97d --- /dev/null +++ b/containers/mariadb.sls @@ -0,0 +1,10 @@ +{% from 'lib.sls' import container_deploy with context %} + +Create data directory for mariadb: + file.directory: + - name: /srv/mariadb + - user: 999 + - group: 999 + - mode: "0755" + +{{ container_deploy('mariadb') }} diff --git a/containers/mosquitto.sls b/containers/mosquitto.sls new file mode 100644 index 0000000..4029265 --- /dev/null +++ b/containers/mosquitto.sls @@ -0,0 +1,17 @@ +{% from 'lib.sls' import container_deploy with context %} + +Create mosquitto configuration directory: + file.directory: + - name: /etc/mosquitto + - user: 1883 + - group: 1883 + - mode: "0750" + +Create mosquitto data directory: + file.directory: + - name: /srv/mosquitto + - user: 1883 + - group: 1883 + - mode: "0750" + +{{ container_deploy('mosquitto') }} diff --git a/containers/nextcloud.sls b/containers/nextcloud.sls new file mode 100644 index 0000000..4c1e36d --- /dev/null +++ b/containers/nextcloud.sls @@ -0,0 +1,11 @@ +{% from 'lib.sls' import container_deploy with context %} + +Create nextcloud data directory: + file.directory: + - name: /srv/nextcloud + - user: 33 + - group: 33 + - mode: "0755" + +{{ container_deploy('nextcloud') }} +{{ container_deploy('nextcloud-cron') }} diff --git a/containers/node_exporter.sls b/containers/node_exporter.sls new file mode 100644 index 0000000..63f15fd --- /dev/null +++ b/containers/node_exporter.sls @@ -0,0 +1,3 @@ +{% from 'lib.sls' import container_deploy with context %} + +{{ container_deploy('node_exporter') }} diff --git a/containers/nodered.sls b/containers/nodered.sls new file mode 100644 index 0000000..bc494b5 --- /dev/null +++ b/containers/nodered.sls @@ -0,0 +1,25 @@ +{% from 'lib.sls' import container_deploy with context %} + +Create udev-rule for rfxcom usb dongel: + file.managed: + - name: /etc/udev/rules.d/99-rfxcom-serial.rules + - source: salt://containers/files/99-rfxcom-serial.rules + - user: root + - group: root + - mode: "0644" + +Relead udev rules for rfxcom dongel: + cmd.run: + - name: udevadm control --reload-rules + - onchanges: + - file: Create udev-rule for rfxcom usb dongel + +Create data folder for nodered: + file.directory: + - name: /srv/nodered + - user: 1000 + - group: 1000 + - mode: "0750" + +{{ container_deploy('nodered') }} + diff --git a/containers/npm.sls b/containers/npm.sls new file mode 100644 index 0000000..1a7f0c4 --- /dev/null +++ b/containers/npm.sls @@ -0,0 +1,24 @@ +{% from 'lib.sls' import container_deploy with context %} + +Create nextcloud root directory: + file.directory: + - name: /srv/npm + - user: root + - group: root + - mode: "0755" + +Create nextcloud data directory: + file.directory: + - name: /srv/npm/data + - user: root + - group: root + - mode: "0750" + +Create nextcloud letsencrypt directory: + file.directory: + - name: /srv/npm/letsencrypt + - user: root + - group: root + - mode: "0750" + +{{ container_deploy('npm') }} diff --git a/containers/piwigo.sls b/containers/piwigo.sls new file mode 100644 index 0000000..12cee9a --- /dev/null +++ b/containers/piwigo.sls @@ -0,0 +1,24 @@ +{% from 'lib.sls' import container_deploy with context %} + +Create piwigo root directory: + file.directory: + - name: /srv/piwigo + - user: {{ pillar.containers.piwigo.env.PUID }} + - group: {{ pillar.containers.piwigo.env.GUID }} + - mode: "0750" + +Create piwigo config directory: + file.directory: + - name: /srv/piwigo/config + - user: {{ pillar.containers.piwigo.env.PUID }} + - group: {{ pillar.containers.piwigo.env.GUID }} + - mode: "0750" + +Create piwigo gallery directory: + file.directory: + - name: /srv/piwigo/gallery + - user: {{ pillar.containers.piwigo.env.PUID }} + - group: {{ pillar.containers.piwigo.env.GUID }} + - mode: "0750" + +{{ container_deploy('piwigo') }} diff --git a/containers/prometheus.sls b/containers/prometheus.sls new file mode 100644 index 0000000..e0ddb90 --- /dev/null +++ b/containers/prometheus.sls @@ -0,0 +1,26 @@ +{% from 'lib.sls' import container_deploy with context %} + +Create prometheus config directory: + file.recurse: + - name: /etc/prometheus + - source: salt://containers/files/prometheus + - user: nobody + - group: root + - dir_mode: "0755" + - file_mode: "0644" + +Create prometheus data directory: + file.directory: + - name: /srv/prometheus + - user: nobody + - group: root + - mode: "0755" + +{{ container_deploy('prometheus') }} + +Start prometheus service: + service.running: + - name: prometheus + - enable: True + - watch: + - file: Create prometheus config directory diff --git a/containers/promtail.sls b/containers/promtail.sls new file mode 100644 index 0000000..139153a --- /dev/null +++ b/containers/promtail.sls @@ -0,0 +1,34 @@ +{%- set client_url = "http://loki.rre.nu:3100/loki/api/v1/push" %} +{%- set http_listen_port = 9080 %} + +{% from 'lib.sls' import container_deploy with context %} + +Create promtail configuration folder: + file.directory: + - name: /etc/promtail + - user: root + - group: root + - mode: "0755" + +Create promtail configuration: + file.managed: + - name: /etc/promtail/promtail.conf + - source: salt://containers/files/promtail.conf.jinja + - template: jinja + - user: root + - group: root + - mode: "0644" + - require: + - file: Create promtail configuration folder + - context: + client_url: {{ client_url }} + http_listen_port: {{ http_listen_port }} + +{{ container_deploy('promtail') }} + +Start promtail service: + service.running: + - name: promtail.service + - enable: True + - watch: + - file: Create promtail configuration diff --git a/containers/salt.sls b/containers/salt.sls new file mode 100644 index 0000000..887a539 --- /dev/null +++ b/containers/salt.sls @@ -0,0 +1,3 @@ +{% from 'lib.sls' import container_deploy with context %} + +{{ container_deploy('salt') }} diff --git a/containers/unifi.sls b/containers/unifi.sls new file mode 100644 index 0000000..3278f05 --- /dev/null +++ b/containers/unifi.sls @@ -0,0 +1,10 @@ +{% from 'lib.sls' import container_deploy with context %} + +Create unifi data directory: + file.directory: + - name: /srv/unifi + - user: root + - group: root + - mode: "0750" + +{{ container_deploy('unifi') }} diff --git a/containers/unpoller.sls b/containers/unpoller.sls new file mode 100644 index 0000000..7d7835b --- /dev/null +++ b/containers/unpoller.sls @@ -0,0 +1,20 @@ +{% from 'lib.sls' import container_deploy with context %} + +create config directory for unpoller: + file.directory: + - name: /srv/unpoller + - user: root + - group: root + - mode: "0750" + +create unpoller config: + file.managed: + - name: /srv/unpoller/up.conf + - source: salt://containers/files/unpoller.conf + - template: jinja + - user: root + - group: root + - mode: "0640" + + +{{ container_deploy('unpoller') }} diff --git a/containers/zigbee2mqtt.sls b/containers/zigbee2mqtt.sls new file mode 100644 index 0000000..1458125 --- /dev/null +++ b/containers/zigbee2mqtt.sls @@ -0,0 +1,25 @@ +{% from 'lib.sls' import container_deploy with context %} + +Create udev-rule for zigbee usb dongel: + file.managed: + - name: /etc/udev/rules.d/99-zigbee-serial.rules + - source: salt://containers/files/99-zigbee-serial.rules + - user: root + - group: root + - mode: "0644" + +Relead udev rules for zigbee dongel: + cmd.run: + - name: udevadm control --reload-rules + - onchanges: + - file: Create udev-rule for zigbee usb dongel + +Create zigbee2mqtt data folder: + file.directory: + - name: /srv/zigbee2mqtt + - user: root + - group: root + - mode: "0755" + +{{ container_deploy('zigbee2mqtt') }} + diff --git a/dnsmasq/dnsmasq.conf b/dnsmasq/dnsmasq.conf new file mode 100644 index 0000000..46e8902 --- /dev/null +++ b/dnsmasq/dnsmasq.conf @@ -0,0 +1,696 @@ +# Configuration file for dnsmasq. +# +# Format is one option per line, legal options are the same +# as the long options legal on the command line. See +# "/usr/sbin/dnsmasq --help" or "man 8 dnsmasq" for details. + +# Accept DNS queries only from hosts whose address is on a local +# subnet, ie a subnet for which an interface exists on the server. +# It is intended to be set as a default on installation, to allow +# unconfigured installations to be useful but also safe from being +# used for DNS amplification attacks. +#local-service + +# Listen on this specific port instead of the standard DNS port +# (53). Setting this to zero completely disables DNS function, +# leaving only DHCP and/or TFTP. +#port=5353 + +# The following two options make you a better netizen, since they +# tell dnsmasq to filter out queries which the public DNS cannot +# answer, and which load the servers (especially the root servers) +# unnecessarily. If you have a dial-on-demand link they also stop +# these requests from bringing up the link unnecessarily. + +# Never forward plain names (without a dot or domain part) +#domain-needed +# Never forward addresses in the non-routed address spaces. +#bogus-priv + +# Uncomment these to enable DNSSEC validation and caching: +# (Requires dnsmasq to be built with DNSSEC option.) +#conf-file=/etc/dnsmasq.d/trust-anchors.conf +#dnssec + +# Replies which are not DNSSEC signed may be legitimate, because the domain +# is unsigned, or may be forgeries. Setting this option tells dnsmasq to +# check that an unsigned reply is OK, by finding a secure proof that a DS +# record somewhere between the root and the domain does not exist. +# The cost of setting this is that even queries in unsigned domains will need +# one or more extra DNS queries to verify. +#dnssec-check-unsigned + +# Uncomment this to filter useless windows-originated DNS requests +# which can trigger dial-on-demand links needlessly. +# Note that (amongst other things) this blocks all SRV requests, +# so don't use it if you use eg Kerberos, SIP, XMMP or Google-talk. +# This option only affects forwarding, SRV records originating for +# dnsmasq (via srv-host= lines) are not suppressed by it. +#filterwin2k + +# Change this line if you want dns to get its upstream servers from +# somewhere other that /etc/resolv.conf +#resolv-file= + +# By default, dnsmasq will send queries to any of the upstream +# servers it knows about and tries to favour servers to are known +# to be up. Uncommenting this forces dnsmasq to try each query +# with each server strictly in the order they appear in +# /etc/resolv.conf +#strict-order + +# If you don't want dnsmasq to read /etc/resolv.conf or any other +# file, getting its servers from this file instead (see below), then +# uncomment this. +#no-resolv + +# If you don't want dnsmasq to poll /etc/resolv.conf or other resolv +# files for changes and re-read them then uncomment this. +#no-poll + +# Add other name servers here, with domain specs if they are for +# non-public domains. +#server=/localnet/192.168.0.1 + +# Example of routing PTR queries to nameservers: this will send all +# address->name queries for 192.168.3/24 to nameserver 10.1.2.3 +#server=/3.168.192.in-addr.arpa/10.1.2.3 + +# Add local-only domains here, queries in these domains are answered +# from /etc/hosts or DHCP only. +#local=/localnet/ + +# Add domains which you want to force to an IP address here. +# The example below send any host in double-click.net to a local +# web-server. +#address=/double-click.net/127.0.0.1 + +# --address (and --server) work with IPv6 addresses too. +#address=/www.thekelleys.org.uk/fe80::20d:60ff:fe36:f83 + +# Add the IPs of all queries to yahoo.com, google.com, and their +# subdomains to the vpn and search ipsets: +#ipset=/yahoo.com/google.com/vpn,search + +# Add the IPs of all queries to yahoo.com, google.com, and their +# subdomains to netfilters sets, which is equivalent to +# 'nft add element ip test vpn { ... }; nft add element ip test search { ... }' +#nftset=/yahoo.com/google.com/ip#test#vpn,ip#test#search + +# Use netfilters sets for both IPv4 and IPv6: +# This adds all addresses in *.yahoo.com to vpn4 and vpn6 for IPv4 and IPv6 addresses. +#nftset=/yahoo.com/4#ip#test#vpn4 +#nftset=/yahoo.com/6#ip#test#vpn6 + +# You can control how dnsmasq talks to a server: this forces +# queries to 10.1.2.3 to be routed via eth1 +# server=10.1.2.3@eth1 + +# and this sets the source (ie local) address used to talk to +# 10.1.2.3 to 192.168.1.1 port 55 (there must be an interface with that +# IP on the machine, obviously). +# server=10.1.2.3@192.168.1.1#55 + +# If you want dnsmasq to change uid and gid to something other +# than the default, edit the following lines. +#user= +#group= + +# If you want dnsmasq to listen for DHCP and DNS requests only on +# specified interfaces (and the loopback) give the name of the +# interface (eg eth0) here. +# Repeat the line for more than one interface. +#interface= +# Or you can specify which interface _not_ to listen on +#except-interface= +# Or which to listen on by address (remember to include 127.0.0.1 if +# you use this.) +#listen-address= +# If you want dnsmasq to provide only DNS service on an interface, +# configure it as shown above, and then use the following line to +# disable DHCP and TFTP on it. +#no-dhcp-interface= + +# On systems which support it, dnsmasq binds the wildcard address, +# even when it is listening on only some interfaces. It then discards +# requests that it shouldn't reply to. This has the advantage of +# working even when interfaces come and go and change address. If you +# want dnsmasq to really bind only the interfaces it is listening on, +# uncomment this option. About the only time you may need this is when +# running another nameserver on the same machine. +#bind-interfaces + +# If you don't want dnsmasq to read /etc/hosts, uncomment the +# following line. +#no-hosts +# or if you want it to read another file, as well as /etc/hosts, use +# this. +#addn-hosts=/etc/banner_add_hosts + +# Set this (and domain: see below) if you want to have a domain +# automatically added to simple names in a hosts-file. +#expand-hosts + +# Set the domain for dnsmasq. this is optional, but if it is set, it +# does the following things. +# 1) Allows DHCP hosts to have fully qualified domain names, as long +# as the domain part matches this setting. +# 2) Sets the "domain" DHCP option thereby potentially setting the +# domain of all systems configured by DHCP +# 3) Provides the domain part for "expand-hosts" +#domain=thekelleys.org.uk + +# Set a different domain for a particular subnet +#domain=wireless.thekelleys.org.uk,192.168.2.0/24 + +# Same idea, but range rather then subnet +#domain=reserved.thekelleys.org.uk,192.68.3.100,192.168.3.200 + +# Uncomment this to enable the integrated DHCP server, you need +# to supply the range of addresses available for lease and optionally +# a lease time. If you have more than one network, you will need to +# repeat this for each network on which you want to supply DHCP +# service. +#dhcp-range=192.168.0.50,192.168.0.150,12h + +# This is an example of a DHCP range where the netmask is given. This +# is needed for networks we reach the dnsmasq DHCP server via a relay +# agent. If you don't know what a DHCP relay agent is, you probably +# don't need to worry about this. +#dhcp-range=192.168.0.50,192.168.0.150,255.255.255.0,12h + +# This is an example of a DHCP range which sets a tag, so that +# some DHCP options may be set only for this network. +#dhcp-range=set:red,192.168.0.50,192.168.0.150 + +# Use this DHCP range only when the tag "green" is set. +#dhcp-range=tag:green,192.168.0.50,192.168.0.150,12h + +# Specify a subnet which can't be used for dynamic address allocation, +# is available for hosts with matching --dhcp-host lines. Note that +# dhcp-host declarations will be ignored unless there is a dhcp-range +# of some type for the subnet in question. +# In this case the netmask is implied (it comes from the network +# configuration on the machine running dnsmasq) it is possible to give +# an explicit netmask instead. +#dhcp-range=192.168.0.0,static + +# Enable DHCPv6. Note that the prefix-length does not need to be specified +# and defaults to 64 if missing/ +#dhcp-range=1234::2, 1234::500, 64, 12h + +# Do Router Advertisements, BUT NOT DHCP for this subnet. +#dhcp-range=1234::, ra-only + +# Do Router Advertisements, BUT NOT DHCP for this subnet, also try and +# add names to the DNS for the IPv6 address of SLAAC-configured dual-stack +# hosts. Use the DHCPv4 lease to derive the name, network segment and +# MAC address and assume that the host will also have an +# IPv6 address calculated using the SLAAC algorithm. +#dhcp-range=1234::, ra-names + +# Do Router Advertisements, BUT NOT DHCP for this subnet. +# Set the lifetime to 46 hours. (Note: minimum lifetime is 2 hours.) +#dhcp-range=1234::, ra-only, 48h + +# Do DHCP and Router Advertisements for this subnet. Set the A bit in the RA +# so that clients can use SLAAC addresses as well as DHCP ones. +#dhcp-range=1234::2, 1234::500, slaac + +# Do Router Advertisements and stateless DHCP for this subnet. Clients will +# not get addresses from DHCP, but they will get other configuration information. +# They will use SLAAC for addresses. +#dhcp-range=1234::, ra-stateless + +# Do stateless DHCP, SLAAC, and generate DNS names for SLAAC addresses +# from DHCPv4 leases. +#dhcp-range=1234::, ra-stateless, ra-names + +# Do router advertisements for all subnets where we're doing DHCPv6 +# Unless overridden by ra-stateless, ra-names, et al, the router +# advertisements will have the M and O bits set, so that the clients +# get addresses and configuration from DHCPv6, and the A bit reset, so the +# clients don't use SLAAC addresses. +#enable-ra + +# Supply parameters for specified hosts using DHCP. There are lots +# of valid alternatives, so we will give examples of each. Note that +# IP addresses DO NOT have to be in the range given above, they just +# need to be on the same network. The order of the parameters in these +# do not matter, it's permissible to give name, address and MAC in any +# order. + +# Always allocate the host with Ethernet address 11:22:33:44:55:66 +# The IP address 192.168.0.60 +#dhcp-host=11:22:33:44:55:66,192.168.0.60 + +# Always set the name of the host with hardware address +# 11:22:33:44:55:66 to be "fred" +#dhcp-host=11:22:33:44:55:66,fred + +# Always give the host with Ethernet address 11:22:33:44:55:66 +# the name fred and IP address 192.168.0.60 and lease time 45 minutes +#dhcp-host=11:22:33:44:55:66,fred,192.168.0.60,45m + +# Give a host with Ethernet address 11:22:33:44:55:66 or +# 12:34:56:78:90:12 the IP address 192.168.0.60. Dnsmasq will assume +# that these two Ethernet interfaces will never be in use at the same +# time, and give the IP address to the second, even if it is already +# in use by the first. Useful for laptops with wired and wireless +# addresses. +#dhcp-host=11:22:33:44:55:66,12:34:56:78:90:12,192.168.0.60 + +# Give the machine which says its name is "bert" IP address +# 192.168.0.70 and an infinite lease +#dhcp-host=bert,192.168.0.70,infinite + +# Always give the host with client identifier 01:02:02:04 +# the IP address 192.168.0.60 +#dhcp-host=id:01:02:02:04,192.168.0.60 + +# Always give the InfiniBand interface with hardware address +# 80:00:00:48:fe:80:00:00:00:00:00:00:f4:52:14:03:00:28:05:81 the +# ip address 192.168.0.61. The client id is derived from the prefix +# ff:00:00:00:00:00:02:00:00:02:c9:00 and the last 8 pairs of +# hex digits of the hardware address. +#dhcp-host=id:ff:00:00:00:00:00:02:00:00:02:c9:00:f4:52:14:03:00:28:05:81,192.168.0.61 + +# Always give the host with client identifier "marjorie" +# the IP address 192.168.0.60 +#dhcp-host=id:marjorie,192.168.0.60 + +# Enable the address given for "judge" in /etc/hosts +# to be given to a machine presenting the name "judge" when +# it asks for a DHCP lease. +#dhcp-host=judge + +# Never offer DHCP service to a machine whose Ethernet +# address is 11:22:33:44:55:66 +#dhcp-host=11:22:33:44:55:66,ignore + +# Ignore any client-id presented by the machine with Ethernet +# address 11:22:33:44:55:66. This is useful to prevent a machine +# being treated differently when running under different OS's or +# between PXE boot and OS boot. +#dhcp-host=11:22:33:44:55:66,id:* + +# Send extra options which are tagged as "red" to +# the machine with Ethernet address 11:22:33:44:55:66 +#dhcp-host=11:22:33:44:55:66,set:red + +# Send extra options which are tagged as "red" to +# any machine with Ethernet address starting 11:22:33: +#dhcp-host=11:22:33:*:*:*,set:red + +# Give a fixed IPv6 address and name to client with +# DUID 00:01:00:01:16:d2:83:fc:92:d4:19:e2:d8:b2 +# Note the MAC addresses CANNOT be used to identify DHCPv6 clients. +# Note also that the [] around the IPv6 address are obligatory. +#dhcp-host=id:00:01:00:01:16:d2:83:fc:92:d4:19:e2:d8:b2, fred, [1234::5] + +# Ignore any clients which are not specified in dhcp-host lines +# or /etc/ethers. Equivalent to ISC "deny unknown-clients". +# This relies on the special "known" tag which is set when +# a host is matched. +#dhcp-ignore=tag:!known + +# Send extra options which are tagged as "red" to any machine whose +# DHCP vendorclass string includes the substring "Linux" +#dhcp-vendorclass=set:red,Linux + +# Send extra options which are tagged as "red" to any machine one +# of whose DHCP userclass strings includes the substring "accounts" +#dhcp-userclass=set:red,accounts + +# Send extra options which are tagged as "red" to any machine whose +# MAC address matches the pattern. +#dhcp-mac=set:red,00:60:8C:*:*:* + +# If this line is uncommented, dnsmasq will read /etc/ethers and act +# on the ethernet-address/IP pairs found there just as if they had +# been given as --dhcp-host options. Useful if you keep +# MAC-address/host mappings there for other purposes. +#read-ethers + +# Send options to hosts which ask for a DHCP lease. +# See RFC 2132 for details of available options. +# Common options can be given to dnsmasq by name: +# run "dnsmasq --help dhcp" to get a list. +# Note that all the common settings, such as netmask and +# broadcast address, DNS server and default route, are given +# sane defaults by dnsmasq. You very likely will not need +# any dhcp-options. If you use Windows clients and Samba, there +# are some options which are recommended, they are detailed at the +# end of this section. + +# Override the default route supplied by dnsmasq, which assumes the +# router is the same machine as the one running dnsmasq. +#dhcp-option=3,1.2.3.4 + +# Do the same thing, but using the option name +#dhcp-option=option:router,1.2.3.4 + +# Override the default route supplied by dnsmasq and send no default +# route at all. Note that this only works for the options sent by +# default (1, 3, 6, 12, 28) the same line will send a zero-length option +# for all other option numbers. +#dhcp-option=3 + +# Set the NTP time server addresses to 192.168.0.4 and 10.10.0.5 +#dhcp-option=option:ntp-server,192.168.0.4,10.10.0.5 + +# Send DHCPv6 option. Note [] around IPv6 addresses. +#dhcp-option=option6:dns-server,[1234::77],[1234::88] + +# Send DHCPv6 option for namservers as the machine running +# dnsmasq and another. +#dhcp-option=option6:dns-server,[::],[1234::88] + +# Ask client to poll for option changes every six hours. (RFC4242) +#dhcp-option=option6:information-refresh-time,6h + +# Set option 58 client renewal time (T1). Defaults to half of the +# lease time if not specified. (RFC2132) +#dhcp-option=option:T1,1m + +# Set option 59 rebinding time (T2). Defaults to 7/8 of the +# lease time if not specified. (RFC2132) +#dhcp-option=option:T2,2m + +# Set the NTP time server address to be the same machine as +# is running dnsmasq +#dhcp-option=42,0.0.0.0 + +# Set the NIS domain name to "welly" +#dhcp-option=40,welly + +# Set the default time-to-live to 50 +#dhcp-option=23,50 + +# Set the "all subnets are local" flag +#dhcp-option=27,1 + +# Send the etherboot magic flag and then etherboot options (a string). +#dhcp-option=128,e4:45:74:68:00:00 +#dhcp-option=129,NIC=eepro100 + +# Specify an option which will only be sent to the "red" network +# (see dhcp-range for the declaration of the "red" network) +# Note that the tag: part must precede the option: part. +#dhcp-option = tag:red, option:ntp-server, 192.168.1.1 + +# The following DHCP options set up dnsmasq in the same way as is specified +# for the ISC dhcpcd in +# https://web.archive.org/web/20040313070105/http://us1.samba.org/samba/ftp/docs/textdocs/DHCP-Server-Configuration.txt +# adapted for a typical dnsmasq installation where the host running +# dnsmasq is also the host running samba. +# you may want to uncomment some or all of them if you use +# Windows clients and Samba. +#dhcp-option=19,0 # option ip-forwarding off +#dhcp-option=44,0.0.0.0 # set netbios-over-TCP/IP nameserver(s) aka WINS server(s) +#dhcp-option=45,0.0.0.0 # netbios datagram distribution server +#dhcp-option=46,8 # netbios node type + +# Send an empty WPAD option. This may be REQUIRED to get windows 7 to behave. +#dhcp-option=252,"\n" + +# Send RFC-3397 DNS domain search DHCP option. WARNING: Your DHCP client +# probably doesn't support this...... +#dhcp-option=option:domain-search,eng.apple.com,marketing.apple.com + +# Send RFC-3442 classless static routes (note the netmask encoding) +#dhcp-option=121,192.168.1.0/24,1.2.3.4,10.0.0.0/8,5.6.7.8 + +# Send vendor-class specific options encapsulated in DHCP option 43. +# The meaning of the options is defined by the vendor-class so +# options are sent only when the client supplied vendor class +# matches the class given here. (A substring match is OK, so "MSFT" +# matches "MSFT" and "MSFT 5.0"). This example sets the +# mtftp address to 0.0.0.0 for PXEClients. +#dhcp-option=vendor:PXEClient,1,0.0.0.0 + +# Send microsoft-specific option to tell windows to release the DHCP lease +# when it shuts down. Note the "i" flag, to tell dnsmasq to send the +# value as a four-byte integer - that's what microsoft wants. See +# http://technet2.microsoft.com/WindowsServer/en/library/a70f1bb7-d2d4-49f0-96d6-4b7414ecfaae1033.mspx?mfr=true +#dhcp-option=vendor:MSFT,2,1i + +# Send the Encapsulated-vendor-class ID needed by some configurations of +# Etherboot to allow is to recognise the DHCP server. +#dhcp-option=vendor:Etherboot,60,"Etherboot" + +# Send options to PXELinux. Note that we need to send the options even +# though they don't appear in the parameter request list, so we need +# to use dhcp-option-force here. +# See http://syslinux.zytor.com/pxe.php#special for details. +# Magic number - needed before anything else is recognised +#dhcp-option-force=208,f1:00:74:7e +# Configuration file name +#dhcp-option-force=209,configs/common +# Path prefix +#dhcp-option-force=210,/tftpboot/pxelinux/files/ +# Reboot time. (Note 'i' to send 32-bit value) +#dhcp-option-force=211,30i + +# Set the boot filename for netboot/PXE. You will only need +# this if you want to boot machines over the network and you will need +# a TFTP server; either dnsmasq's built-in TFTP server or an +# external one. (See below for how to enable the TFTP server.) +#dhcp-boot=pxelinux.0 + +# The same as above, but use custom tftp-server instead machine running dnsmasq +#dhcp-boot=pxelinux,server.name,192.168.1.100 + +# Boot for iPXE. The idea is to send two different +# filenames, the first loads iPXE, and the second tells iPXE what to +# load. The dhcp-match sets the ipxe tag for requests from iPXE. +#dhcp-boot=undionly.kpxe +#dhcp-match=set:ipxe,175 # iPXE sends a 175 option. +#dhcp-boot=tag:ipxe,http://boot.ipxe.org/demo/boot.php + +# Encapsulated options for iPXE. All the options are +# encapsulated within option 175 +#dhcp-option=encap:175, 1, 5b # priority code +#dhcp-option=encap:175, 176, 1b # no-proxydhcp +#dhcp-option=encap:175, 177, string # bus-id +#dhcp-option=encap:175, 189, 1b # BIOS drive code +#dhcp-option=encap:175, 190, user # iSCSI username +#dhcp-option=encap:175, 191, pass # iSCSI password + +# Test for the architecture of a netboot client. PXE clients are +# supposed to send their architecture as option 93. (See RFC 4578) +#dhcp-match=peecees, option:client-arch, 0 #x86-32 +#dhcp-match=itanics, option:client-arch, 2 #IA64 +#dhcp-match=hammers, option:client-arch, 6 #x86-64 +#dhcp-match=mactels, option:client-arch, 7 #EFI x86-64 + +# Do real PXE, rather than just booting a single file, this is an +# alternative to dhcp-boot. +#pxe-prompt="What system shall I netboot?" +# or with timeout before first available action is taken: +#pxe-prompt="Press F8 for menu.", 60 + +# Available boot services. for PXE. +#pxe-service=x86PC, "Boot from local disk" + +# Loads /pxelinux.0 from dnsmasq TFTP server. +#pxe-service=x86PC, "Install Linux", pxelinux + +# Loads /pxelinux.0 from TFTP server at 1.2.3.4. +# Beware this fails on old PXE ROMS. +#pxe-service=x86PC, "Install Linux", pxelinux, 1.2.3.4 + +# Use bootserver on network, found my multicast or broadcast. +#pxe-service=x86PC, "Install windows from RIS server", 1 + +# Use bootserver at a known IP address. +#pxe-service=x86PC, "Install windows from RIS server", 1, 1.2.3.4 + +# If you have multicast-FTP available, +# information for that can be passed in a similar way using options 1 +# to 5. See page 19 of +# http://download.intel.com/design/archives/wfm/downloads/pxespec.pdf + + +# Enable dnsmasq's built-in TFTP server +#enable-tftp + +# Set the root directory for files available via FTP. +#tftp-root=/var/ftpd + +# Do not abort if the tftp-root is unavailable +#tftp-no-fail + +# Make the TFTP server more secure: with this set, only files owned by +# the user dnsmasq is running as will be send over the net. +#tftp-secure + +# This option stops dnsmasq from negotiating a larger blocksize for TFTP +# transfers. It will slow things down, but may rescue some broken TFTP +# clients. +#tftp-no-blocksize + +# Set the boot file name only when the "red" tag is set. +#dhcp-boot=tag:red,pxelinux.red-net + +# An example of dhcp-boot with an external TFTP server: the name and IP +# address of the server are given after the filename. +# Can fail with old PXE ROMS. Overridden by --pxe-service. +#dhcp-boot=/var/ftpd/pxelinux.0,boothost,192.168.0.3 + +# If there are multiple external tftp servers having a same name +# (using /etc/hosts) then that name can be specified as the +# tftp_servername (the third option to dhcp-boot) and in that +# case dnsmasq resolves this name and returns the resultant IP +# addresses in round robin fashion. This facility can be used to +# load balance the tftp load among a set of servers. +#dhcp-boot=/var/ftpd/pxelinux.0,boothost,tftp_server_name + +# Set the limit on DHCP leases, the default is 150 +#dhcp-lease-max=150 + +# The DHCP server needs somewhere on disk to keep its lease database. +# This defaults to a sane location, but if you want to change it, use +# the line below. +#dhcp-leasefile=/var/lib/misc/dnsmasq.leases + +# Set the DHCP server to authoritative mode. In this mode it will barge in +# and take over the lease for any client which broadcasts on the network, +# whether it has a record of the lease or not. This avoids long timeouts +# when a machine wakes up on a new network. DO NOT enable this if there's +# the slightest chance that you might end up accidentally configuring a DHCP +# server for your campus/company accidentally. The ISC server uses +# the same option, and this URL provides more information: +# http://www.isc.org/files/auth.html +#dhcp-authoritative + +# Set the DHCP server to enable DHCPv4 Rapid Commit Option per RFC 4039. +# In this mode it will respond to a DHCPDISCOVER message including a Rapid Commit +# option with a DHCPACK including a Rapid Commit option and fully committed address +# and configuration information. This must only be enabled if either the server is +# the only server for the subnet, or multiple servers are present and they each +# commit a binding for all clients. +#dhcp-rapid-commit + +# Run an executable when a DHCP lease is created or destroyed. +# The arguments sent to the script are "add" or "del", +# then the MAC address, the IP address and finally the hostname +# if there is one. +#dhcp-script=/bin/echo + +# Set the cachesize here. +#cache-size=150 + +# If you want to disable negative caching, uncomment this. +#no-negcache + +# Normally responses which come from /etc/hosts and the DHCP lease +# file have Time-To-Live set as zero, which conventionally means +# do not cache further. If you are happy to trade lower load on the +# server for potentially stale date, you can set a time-to-live (in +# seconds) here. +#local-ttl= + +# If you want dnsmasq to detect attempts by Verisign to send queries +# to unregistered .com and .net hosts to its sitefinder service and +# have dnsmasq instead return the correct NXDOMAIN response, uncomment +# this line. You can add similar lines to do the same for other +# registries which have implemented wildcard A records. +#bogus-nxdomain=64.94.110.11 + +# If you want to fix up DNS results from upstream servers, use the +# alias option. This only works for IPv4. +# This alias makes a result of 1.2.3.4 appear as 5.6.7.8 +#alias=1.2.3.4,5.6.7.8 +# and this maps 1.2.3.x to 5.6.7.x +#alias=1.2.3.0,5.6.7.0,255.255.255.0 +# and this maps 192.168.0.10->192.168.0.40 to 10.0.0.10->10.0.0.40 +#alias=192.168.0.10-192.168.0.40,10.0.0.0,255.255.255.0 + +# Change these lines if you want dnsmasq to serve MX records. + +# Return an MX record named "maildomain.com" with target +# servermachine.com and preference 50 +#mx-host=maildomain.com,servermachine.com,50 + +# Set the default target for MX records created using the localmx option. +#mx-target=servermachine.com + +# Return an MX record pointing to the mx-target for all local +# machines. +#localmx + +# Return an MX record pointing to itself for all local machines. +#selfmx + +# Change the following lines if you want dnsmasq to serve SRV +# records. These are useful if you want to serve ldap requests for +# Active Directory and other windows-originated DNS requests. +# See RFC 2782. +# You may add multiple srv-host lines. +# The fields are ,,,, +# If the domain part if missing from the name (so that is just has the +# service and protocol sections) then the domain given by the domain= +# config option is used. (Note that expand-hosts does not need to be +# set for this to work.) + +# A SRV record sending LDAP for the example.com domain to +# ldapserver.example.com port 389 +#srv-host=_ldap._tcp.example.com,ldapserver.example.com,389 + +# A SRV record sending LDAP for the example.com domain to +# ldapserver.example.com port 389 (using domain=) +#domain=example.com +#srv-host=_ldap._tcp,ldapserver.example.com,389 + +# Two SRV records for LDAP, each with different priorities +#srv-host=_ldap._tcp.example.com,ldapserver.example.com,389,1 +#srv-host=_ldap._tcp.example.com,ldapserver.example.com,389,2 + +# A SRV record indicating that there is no LDAP server for the domain +# example.com +#srv-host=_ldap._tcp.example.com + +# The following line shows how to make dnsmasq serve an arbitrary PTR +# record. This is useful for DNS-SD. (Note that the +# domain-name expansion done for SRV records _does_not +# occur for PTR records.) +#ptr-record=_http._tcp.dns-sd-services,"New Employee Page._http._tcp.dns-sd-services" + +# Change the following lines to enable dnsmasq to serve TXT records. +# These are used for things like SPF and zeroconf. (Note that the +# domain-name expansion done for SRV records _does_not +# occur for TXT records.) + +#Example SPF. +#txt-record=example.com,"v=spf1 a -all" + +#Example zeroconf +#txt-record=_http._tcp.example.com,name=value,paper=A4 + +# Provide an alias for a "local" DNS name. Note that this _only_ works +# for targets which are names from DHCP or /etc/hosts. Give host +# "bert" another name, bertrand +#cname=bertand,bert + +# For debugging purposes, log each DNS query as it passes through +# dnsmasq. +#log-queries + +# Log lots of extra information about DHCP transactions. +#log-dhcp + +# Include another lot of configuration options. +#conf-file=/etc/dnsmasq.more.conf +#conf-dir=/etc/dnsmasq.d + +# Include all the files in a directory except those ending in .bak +#conf-dir=/etc/dnsmasq.d,.bak + +# Include all files in a directory which end in .conf +conf-dir=/config/dnsmasq.d/,*.conf + +# If a DHCP client claims that its name is "wpad", ignore that. +# This fixes a security hole. see CERT Vulnerability VU#598349 +#dhcp-name-match=set:wpad-ignore,wpad +#dhcp-ignore-names=tag:wpad-ignore diff --git a/dnsmasq/dnsmasq.d/rre.conf b/dnsmasq/dnsmasq.d/rre.conf new file mode 100644 index 0000000..ea73654 --- /dev/null +++ b/dnsmasq/dnsmasq.d/rre.conf @@ -0,0 +1,23 @@ +server=10.2.0.1 +local=/rre.nu/ +expand-hosts +domain=rre.nu + +# public cnames +cname=push.rre.nu,npm.rre.nu +cname=git.rre.nu,npm.rre.nu +cname=home.rre.nu,npm.rre.nu +cname=nextcloud.rre.nu,npm.rre.nu +cname=nodered.rre.nu,npm.rre.nu +cname=rre.nu,npm.rre.nu +cname=zigbee2mqtt.rre.nu,npm.rre.nu +cname=grafana.rre.nu,npm.rre.nu +cname=nv.rre.nu,npm.rre.nu +cname=esphome.rre.nu,npm.rre.nu +cname=unifi.rre.nu,npm.rre.nu +cname=prometheus.rre.nu,npm.rre.nu +cname=unpoller.rre.nu,serrano.rre.nu +cname=loki.rre.nu,serrano.rre.nu +cname=db.rre.nu,serrano.rre.nu +cname=mqtt.rre.nu,serrano.rre.nu +cname=foto.rre.nu,serrano.rre.nu diff --git a/dnsmasq/hosts/hosts b/dnsmasq/hosts/hosts new file mode 100644 index 0000000..6322060 --- /dev/null +++ b/dnsmasq/hosts/hosts @@ -0,0 +1,53 @@ +# +# hosts This file describes a number of hostname-to-address +# mappings for the TCP/IP subsystem. It is mostly +# used at boot time, when no name servers are running. +# On small systems, this file can be used instead of a +# "named" name server. +# Syntax: +# +# IP-Address Full-Qualified-Hostname Short-Hostname +# + +127.0.0.1 localhost + +# special IPv6 addresses +::1 localhost ipv6-localhost ipv6-loopback + +fe00::0 ipv6-localnet + +ff00::0 ipv6-mcastprefix +ff02::1 ipv6-allnodes +ff02::2 ipv6-allrouters +ff02::3 ipv6-allhosts + +## rre infrasturcture +10.2.0.2 ldap +10.2.0.22 poblano +10.2.0.23 npm +10.2.0.23 serrano +10.2.0.24 coronado +10.2.0.101 salt +10.2.0.102 pepper + + +#lab k3s cluster +10.2.0.250 demo-k3s +10.2.0.251 srv1 +10.2.0.252 srv2 +10.2.0.253 srv3 + +# MetalLB +# 10.2.0.60-70 +10.2.0.60 public +10.2.0.61 private +10.2.0.63 gitea + +# Client network +10.0.10.3 sw2 +10.0.10.4 ap1 +10.0.10.5 ap2 +10.0.10.6 ap3 +10.0.10.7 ap4 +10.0.10.20 woody +10.0.10.30 laserjet diff --git a/lib.sls b/lib.sls new file mode 100644 index 0000000..d9e2019 --- /dev/null +++ b/lib.sls @@ -0,0 +1,102 @@ +{% macro pod_deploy(pod) -%} + +{%- if pillar['pods'][pod]['ip'] is defined %} +Configure static IP for {{ pod }}: + cmd.run: + - name: nmcli connection modify {{ pillar['pods'][pod]['ip']['dev'] }} +ipv4.addresses "{{ pillar['pods'][pod]['ip']['address'] }}/24" + - unless: ip -brief a | grep "{{ pillar['pods'][pod]['ip']['address'] }}/24" + +Restart connection for {{ pod }} IP config: + cmd.run: + - name: nmcli connection down {{ pillar['pods'][pod]['ip']['dev'] }}; nmcli connection up {{ pillar['pods'][pod]['ip']['dev'] }} + - onchanges: + - cmd: Configure static IP for {{ pod }} +{%- endif %} + +Create pod manage file for {{ pod }}: + file.managed: + - name: "/root/bin/{{ pod }}.sh" + - source: salt://pod/files/pod.sh.jinja + - template: jinja + - user: root + - group: root + - mode: "0750" + - context: + pod: {{ pod }} + args: {{ pillar['pods'][pod] }} + +Create pod {{ pod }}: + cmd.run: + - name: "/root/bin/{{ pod }}.sh --create" + - unless: podman pod exists {{ pod }} + +Create pod {{ pod }} unit file: + cmd.run: + - name: "/root/bin/{{ pod }}.sh --generate-systemd" + - onchanges: + - cmd: Create pod {{ pod }} + +Run user daemon reload for {{ pod }} unit: + cmd.run: + - name: systemctl daemon-reload + - onchanges: + - cmd: Create pod {{ pod }} unit file + +start pod {{ pod }}: + service.running: + - name: {{ pod }}.service + - enable: True + - onchanges: + - cmd: Run user daemon reload for {{ pod }} unit +{% endmacro -%} +{% macro container_deploy(container) -%} + +{%- if pillar['containers'][container]['ip'] is defined %} +Configure static IP for {{ container }}: + cmd.run: + - name: nmcli connection modify {{ pillar['containers'][container]['ip']['dev'] }} +ipv4.addresses "{{ pillar['containers'][container]['ip']['address'] }}/24" + - unless: ip -brief a | grep "{{ pillar['containers'][container]['ip']['address'] }}/24" + +Restart connection for {{ container }} IP config: + cmd.run: + - name: nmcli connection down {{ pillar['containers'][container]['ip']['dev'] }}; nmcli connection up {{ pillar['containers'][container]['ip']['dev'] }} + - onchanges: + - cmd: Configure static IP for {{ container }} +{%- endif %} + +Create container manage file for {{ container }}: + file.managed: + - name: "/root/bin/{{ container }}.sh" + - source: salt://containers/files/container.sh.jinja + - template: jinja + - user: root + - group: root + - mode: "0750" + - context: + container: {{ container }} + args: {{ pillar['containers'][container] }} + +Create container {{ container }}: + cmd.run: + - name: "/root/bin/{{ container }}.sh --create" + - unless: podman container exists {{ container }} + +Create container {{ container }} unit file: + cmd.run: + - name: podman generate systemd --name {{ container }} > /etc/systemd/system/{{ container }}.service + - onchanges: + - cmd: Create container {{ container }} + +Run user daemon reload for {{ container }} unit: + cmd.run: + - name: systemctl daemon-reload + - onchanges: + - cmd: Create container {{ container }} unit file + +start container {{ container }}: + service.running: + - name: {{ container }}.service + - enable: True + - onchanges: + - cmd: Run user daemon reload for {{ container }} unit +{% endmacro -%} diff --git a/pod/dnsmasq.sls b/pod/dnsmasq.sls new file mode 100644 index 0000000..68a858a --- /dev/null +++ b/pod/dnsmasq.sls @@ -0,0 +1,21 @@ +{% from 'lib.sls' import pod_deploy with context %} + +Create dnsmasq config directory: + file.recurse: + - name: /srv/dnsmasq + - source: salt://dnsmasq + - template: jinja + - user: root + - group: root + - dir_mode: "0755" + - file_mode: "0644" + - include_empty: True + +{{ pod_deploy('dnsmasq') }} + +Make sure dnsmasq is running: + service.running: + - name: dnsmasq + - enable: True + - watch: + - file: Create dnsmasq config directory diff --git a/pod/files/99-rfxcom-serial.rules b/pod/files/99-rfxcom-serial.rules new file mode 100644 index 0000000..8d0fb45 --- /dev/null +++ b/pod/files/99-rfxcom-serial.rules @@ -0,0 +1 @@ +SUBSYSTEM=="tty", ATTRS{idVendor}=="0403", ATTRS{idProduct}=="6001", SYMLINK+="rfxcom", MODE="0666" diff --git a/pod/files/99-zigbee-serial.rules b/pod/files/99-zigbee-serial.rules new file mode 100644 index 0000000..25cf77a --- /dev/null +++ b/pod/files/99-zigbee-serial.rules @@ -0,0 +1 @@ +SUBSYSTEM=="tty", ATTRS{idVendor}=="0451", ATTRS{idProduct}=="16a8", SYMLINK+="zigbee-serial", MODE="0666" diff --git a/pod/files/check_image_updates.service.jinja b/pod/files/check_image_updates.service.jinja new file mode 100644 index 0000000..3a1a788 --- /dev/null +++ b/pod/files/check_image_updates.service.jinja @@ -0,0 +1,10 @@ +[Unit] +Description=Check for image updates on configured podman containers + +[Service] +Type=oneshot +User=root +ExecStart=/root/bin/check_image_updates.sh + +[Install] +WantedBy=default.target diff --git a/pod/files/check_image_updates.sh.jinja b/pod/files/check_image_updates.sh.jinja new file mode 100644 index 0000000..277cdcc --- /dev/null +++ b/pod/files/check_image_updates.sh.jinja @@ -0,0 +1,35 @@ +#!/usr/bin/env bash + +URL="{{ pillar['podman']['gotify']['url'] }}" +TOKEN="{{ pillar['podman']['gotify']['token'] }}" +TITLE="Updates on $HOSTNAME" +PRIORITY="{{ pillar['podman']['gotify']['priority'] }}" + +{% raw -%} +function check_update(){ + IFS=',' read -r -a container_info <<< "$(podman container inspect $1 --format '{{ .Name }},{{ .ImageName }},{{ .Image }}')" + + podman pull "${container_info[1]}" + if [[ "$(podman image inspect "${container_info[1]}" --format "{{.Id}}")" != "${container_info[2]}" ]];then + containers[${#containers[@]}]="${container_info[0]}" + fi +} + + +IFS=$'\n' +for line in $(podman container ls -q); do + check_update "$line" +done +if [[ "${#containers[@]}" == "0" ]]; then + exit +fi + +MESSAGE=$(cat << EOM +Following ${#containers[@]} container(s) has updates: +${containers[*]} +EOM +) + +curl "$URL/message?token=$TOKEN" -F "title=$TITLE" -F "priority=$PRIORITY" -F "message=$MESSAGE" +echo " " +{% endraw -%} diff --git a/pod/files/check_image_updates.timer.jinja b/pod/files/check_image_updates.timer.jinja new file mode 100644 index 0000000..23a5b92 --- /dev/null +++ b/pod/files/check_image_updates.timer.jinja @@ -0,0 +1,9 @@ +[Unit] +Description=Restic backup timer + +[Timer] +OnCalendar=Sun, 12:00 +Unit=check_image_updates.service + +[Install] +WantedBy=timers.target diff --git a/pod/files/pod.sh.jinja b/pod/files/pod.sh.jinja new file mode 100644 index 0000000..7eee803 --- /dev/null +++ b/pod/files/pod.sh.jinja @@ -0,0 +1,158 @@ +#!/usr/bin/env bash + +function pull_image(){ + {%- for container, cargs in pillar['pods'][pod]['containers'].items() %} + if ! podman image exists {{ cargs['image'] }}:{{ cargs['tag'] }}; then + podman pull {{ cargs['image'] }}:{{ cargs['tag'] }} + fi + {%- endfor %} +} + + + +function create_pod() { + if ! podman pod exists {{ pod }};then + podman pod create \ + --name {{ pod }} \ + --infra-name {{ pod }}-infra \ + {%- if args['ports'] is defined %} + {%- for ports in args['ports'] %} + -p {{ ports['host'] }}:{{ ports['container'] }}{% if ports['protocol'] is defined %}/{{ ports['protocol'] }}{% endif %} \ + {%- endfor %} + {%- endif %} + + fi + {%- for container, cargs in pillar['pods'][pod]['containers'].items() %} + if ! podman container exists {{ pod }}-{{ container }};then + podman container create \ + --name {{ pod }}-{{ container }} \ + --pod {{ pod }} \ + {%- if cargs['podman_options'] is defined %} + {%- for option, value in cargs['podman_options'].items() %} + --{{ option }} {{ value }} \ + {%- endfor %} + {%- endif %} + {%- if cargs['volumes'] is defined %} + {%- for volume, mount in cargs['volumes'].items() %} + -v {{ volume }}:{{ mount }} \ + {%- endfor %} + {%- endif %} + {%- if cargs['env'] is defined %} + {%- for key, value in cargs['env'].items() %} + -e {{ key }}={{ value }} \ + {%- endfor %} + {%- endif %} + {%- if cargs['devices'] is defined %} + {%- for key, value in cargs['devices'].items() %} + --device {{ key }}:{{ value}} \ + {%- endfor %} + {%- endif %} + {{ cargs['image'] }}:{{ cargs['tag'] }}{%- if cargs['run'] is defined %} \ + {{ cargs['run'] }} + {%- endif %} + fi + {%- endfor %} +} + +function generate_systemd_unit_file() { + cd /etc/systemd/system + podman generate systemd --files --name {{ pod }} --pod-prefix="" --container-prefix="" +} + +function check_update() { + ImageUpdate=0 + {%- for container, cargs in pillar['pods'][pod]['containers'].items() %} + podman pull {{ cargs['image'] }}:{{ cargs['tag'] }} + if [[ "$(podman image inspect {{ cargs['image'] }}:{{ cargs['tag'] }} --format "{% raw %}{{.Id}}{% endraw %}")" == "$(podman inspect {{ pod }}-{{ container }} --format "{% raw %}{{ .Image }}{% endraw %}")" ]];then + echo "No image updates available for {{ pod }}-{{ container }}" + else + echo "Image update available for {{ pod }}-{{ container }}" + ImageUpdate=1 + fi + {%- endfor %} + return $ImageUpdate +} + +function update() { + systemctl stop {{ pod }} + podman pod rm {{ pod }} + create_pod + generate_systemd_unit_file + systemctl daemon-reload + systemctl enable --now {{ pod }}.service +} + + + +function printHelp(){ + cat << EOF +Usage ${0##*/} [options..] +-h,-?, --help Show help and exit +-p, --pull pull container images for all containers in pod {{ pod }} +-c, --create create {{ pod }} pod +-s, --start start and enables {{ pod }} pod +-S, --stop stop {{ pod }} pod +-i, --is-running check to see if pod service is running +-u, --check-update check if there are image updates avaiable + --update perform image update if it exists +-g, --generate-systemd generate user systemd service unit file +EOF +} + + +while :; do + case $1 in + -h|-\?|--help) + printHelp + exit + ;; + -p|--pull) + pull_image + shift + ;; + -v|--volumes) + create_volumes + shift + ;; + -c|--create) + create_pod + shift + ;; + -s|--start) + systemctl --user enable --now {{ pod }}.service + shift + ;; + -S|--stop) + systemctl --user stop {{ pod }}.service + shift + ;; + -i|--is-running) + systemctl --user is-active {{ pod }}.service + exit $? + shift + ;; + -g|--generate-systemd) + generate_systemd_unit_file + shift + ;; + -u|--check-update) + check_update + shift + ;; + --update) + update + shift + ;; + --) #End of all options + shift + break + ;; + -?*) + printf "'%s' is not a valid option\n" "$1" >&2 + exit 1 + ;; + *) #Break out of case, no more options + break + esac + shift +done diff --git a/pod/gotify.sls b/pod/gotify.sls new file mode 100644 index 0000000..6642f4d --- /dev/null +++ b/pod/gotify.sls @@ -0,0 +1,11 @@ +{% from 'lib.sls' import pod_deploy with context %} + +Create gotify data directory: + file.directory: + - name: /srv/gotify + - user: root + - group: root + - mode: "0755" + +{{ pod_deploy('gotify') }} + diff --git a/pod/init.sls b/pod/init.sls new file mode 100644 index 0000000..3e05c1e --- /dev/null +++ b/pod/init.sls @@ -0,0 +1,46 @@ +Copy check image update script: + file.managed: + - name: /root/bin/check_image_updates.sh + - source: salt://pod/files/check_image_updates.sh.jinja + - template: jinja + - user: root + - group: root + - mode: 0700 + +Create check image update service: + file.managed: + - name: /etc/systemd/system/check_image_updates.service + - source: salt://pod/files/check_image_updates.service.jinja + - template: jinja + - user: root + - group: root + - mode: "0644" + +Create check image update timer: + file.managed: + - name: /etc/systemd/system/check_image_updates.timer + - source: salt://pod/files/check_image_updates.timer.jinja + - template: jinja + - user: root + - group: root + - mode: "0644" + +Systemd daemon reload for image update: + cmd.run: + - name: systemctl daemon-reload + - watch: + - file: Create check image update service + - file: Create check image update timer + +Start check image update timer: + service.running: + - name: check_image_updates.timer + - enable: True + +{% if pillar.pods is defined %} +include: +{%- for pod, args in pillar.pods.items() %} + - containers.{{ pod }} +{%- endfor %} +{%- endif %} + diff --git a/pod/loki.sls b/pod/loki.sls new file mode 100644 index 0000000..e877803 --- /dev/null +++ b/pod/loki.sls @@ -0,0 +1,24 @@ +{% from 'lib.sls' import pod_deploy with context %} + +Create loki root directory: + file.directory: + - name: /srv/loki + - user: 10001 + - group: root + - mode: "0755" + +Create loki data directory: + file.directory: + - name: /srv/loki + - user: 10001 + - group: root + - mode: "0755" + +Create loki config directory: + file.directory: + - name: /srv/loki/config + - user: 10001 + - group: root + - mode: "0755" + +{{ pod_deploy('loki') }} diff --git a/pod/mariadb.sls b/pod/mariadb.sls new file mode 100644 index 0000000..1609896 --- /dev/null +++ b/pod/mariadb.sls @@ -0,0 +1,11 @@ +{% from 'lib.sls' import pod_deploy with context %} + +Create data directory for mariadb: + file.directory: + - name: /srv/mariadb + - user: 999 + - group: 999 + - mode: "0755" + +{{ pod_deploy('mariadb') }} + diff --git a/pod/mosquitto.sls b/pod/mosquitto.sls new file mode 100644 index 0000000..41f6842 --- /dev/null +++ b/pod/mosquitto.sls @@ -0,0 +1,17 @@ +{% from 'lib.sls' import pod_deploy with context %} + +Create mosquitto configuration directory: + file.directory: + - name: /etc/mosquitto + - user: 1883 + - group: 1883 + - mode: "0750" + +Create mosquitto data directory: + file.directory: + - name: /srv/mosquitto + - user: 1883 + - group: 1883 + - mode: "0750" + +{{ pod_deploy('mosquitto') }} diff --git a/pod/nextcloud.sls b/pod/nextcloud.sls new file mode 100644 index 0000000..f37829c --- /dev/null +++ b/pod/nextcloud.sls @@ -0,0 +1,10 @@ +{% from 'lib.sls' import pod_deploy with context %} + +Create nextcloud data directory: + file.directory: + - name: /srv/nextcloud + - user: 33 + - group: 33 + - mode: "0755" + +{{ pod_deploy('nextcloud') }} diff --git a/pod/nodered.sls b/pod/nodered.sls new file mode 100644 index 0000000..52569ee --- /dev/null +++ b/pod/nodered.sls @@ -0,0 +1,25 @@ +{% from 'lib.sls' import pod_deploy with context %} + +Create udev-rule for rfxcom usb dongel: + file.managed: + - name: /etc/udev/rules.d/99-rfxcom-serial.rules + - source: salt://pod/files/99-rfxcom-serial.rules + - user: root + - group: root + - mode: "0644" + +Relead udev rules for rfxcom dongel: + cmd.run: + - name: udevadm control --reload-rules + - onchanges: + - file: Create udev-rule for rfxcom usb dongel + +Create data folder for nodered: + file.directory: + - name: /srv/nodered + - user: 1000 + - group: 1000 + - mode: "0750" + +{{ pod_deploy('nodered') }} + diff --git a/pod/piwigo.sls b/pod/piwigo.sls new file mode 100644 index 0000000..b7c3c36 --- /dev/null +++ b/pod/piwigo.sls @@ -0,0 +1,25 @@ +{% from 'lib.sls' import pod_deploy with context %} + +Create piwigo root directory: + file.directory: + - name: /srv/piwigo + - user: {{ pillar.pods.piwigo.containers.main.env.PUID }} + - group: {{ pillar.pods.piwigo.containers.main.env.GUID }} + - mode: "0750" + +Create piwigo config directory: + file.directory: + - name: /srv/piwigo/config + - user: {{ pillar.pods.piwigo.containers.main.env.PUID }} + - group: {{ pillar.pods.piwigo.containers.main.env.GUID }} + - mode: "0750" + +Create piwigo gallery directory: + file.directory: + - name: /srv/piwigo/gallery + - user: {{ pillar.pods.piwigo.containers.main.env.PUID }} + - group: {{ pillar.pods.piwigo.containers.main.env.GUID }} + - mode: "0750" + +{{ pod_deploy('piwigo') }} + diff --git a/pod/prometheus.sls b/pod/prometheus.sls new file mode 100644 index 0000000..7566447 --- /dev/null +++ b/pod/prometheus.sls @@ -0,0 +1,47 @@ +{% from 'lib.sls' import pod_deploy with context %} + +Create prometheus root directory: + file.directory: + - name: /srv/prometheus + - user: nobody + - group: root + - mode: "0755" + +Create prometheus data directory: + file.directory: + - name: /srv/prometheus/data + - user: nobody + - group: root + - mode: "0755" + +Create prometheus config directory: + file.directory: + - name: /srv/prometheus/config + - user: nobody + - group: root + - mode: "0755" + +Create alertmanager root directory: + file.directory: + - name: /srv/alertmanager + - source: salt://containers/files/alertmanager + - user: nobody + - group: root + - dir_mode: "0755" + - file_mode: "0644" + +Create alertmanager data directory: + file.directory: + - name: /srv/alertmanager/data + - user: nobody + - group: nobody + - mode: "0755" + +Create alertmanager config directory: + file.directory: + - name: /srv/alertmanager/config + - user: nobody + - group: nobody + - mode: "0755" + +{{ pod_deploy('prometheus') }} diff --git a/pod/unifi.sls b/pod/unifi.sls new file mode 100644 index 0000000..0be79bb --- /dev/null +++ b/pod/unifi.sls @@ -0,0 +1,10 @@ +{% from 'lib.sls' import pod_deploy with context %} + +Create unifi data directory: + file.directory: + - name: /srv/unifi + - user: root + - group: root + - mode: "0750" + +{{ pod_deploy('unifi') }} diff --git a/restic/files/backup.sh.jinja b/restic/files/backup.sh.jinja new file mode 100644 index 0000000..3d3232a --- /dev/null +++ b/restic/files/backup.sh.jinja @@ -0,0 +1,99 @@ +#!/bin/bash + +{%- if pillar['pods']['mariadb'] is defined %} +umask 0077 +BACKUP_DIR={{ pillar.pods.mariadb.containers.main.backup_dir }} +databases=$(podman exec -it mariadb-main mysql -B -u root -p{{ pillar.pods.mariadb.containers.main.env.MYSQL_ROOT_PASSWORD }} -e "SHOW DATABASES;" | tr -d "| " | grep -v Database) + +for db in ${databases[@]}; do + db=${db::-1} + if [[ "$db" != "information_schema" ]] && [[ "$db" != "performance_schema" ]] && [[ "$db" != "mysql" ]] && [[ "$db" != _* ]] && [[ "$db" != "sys" ]]; then + echo "Dumping database: $db" + podman exec -it mariadb-main mysqldump -u root -p{{ pillar.pods.mariadb.containers.main.env.MYSQL_ROOT_PASSWORD }} --databases $db | gzip > ${BACKUP_DIR}/$(date +"%Y-%m-%d_%H-%M-%S")_$db-sql.gz + fi +done +# Delete the files older than 3 days +echo "removing old mysql dumps" +find $BACKUP_DIR/* -type f -name *-sql.gz -mtime +3 -exec rm {} \; +umask 0022 +{%- endif %} + +{%- if pillar['containers']['freeipa'] is defined %} +echo "Stopping FreeIPA" +systemctl stop freeipa.service +{%- endif %} + +podman run --rm \ + --name=restic \ + --hostname="$HOSTNAME" \ + -v /root/.restic.password:/restic-password:ro \ + -v /root:/root \ +{%- if pillar.restic.repository is defined %} + -v {{ pillar.restic.mount }}{{ pillar.restic.suffix }}:/repo \ +{%- endif %} +{%- for target in pillar.restic.targets %} + -v {{ target }}:{{ target }} \ +{%- endfor %} + -v /root/.restic.password:/root/.restic.password \ + {{ pillar.restic.image.url }}:{{ pillar.restic.image.tag }} \ + restic \ + backup \ + --password-file=/restic-password \ +{%- if pillar.restic.repository is defined %} + --repo=/repo \ +{%- else %} + -r sftp:{{ pillar.restic.user }}@{{ pillar.restic.host }}:{{ pillar.restic.mount }}{{ pillar.restic.suffix }} \ +{%- endif %} + --exclude="*.tmp" \ + --exclude="lost+found" \ + --exclude="Cache" \ + --exclude="cache" \ + --exclude=".cache" \ + --exclude="tmp" \ + --exclude="temp" \ + --exclude="Temp" \ + --exclude="/home/*/go" \ + --exclude="/home/*/.local/share/virtualenv" \ + --exclude="/home/*/.local/share/virtualenvs" \ + --exclude="/home/*/VirtualBox VMs" \ + --exclude="/home/*/.mozillla/firefox/*/minidumps" \ + --exclude="/home/*/.mozillla/firefox/*/storage" \ + --exclude="/home/*/.mozillla/firefox/*/extensions.sqlite" \ + --exclude="/home/*/.mozillla/firefox/*/urlclassifier3.sqlite" \ + --exclude="/home/*/.config/google-chrome/*/Local Storage" \ + --exclude="/home/*/.config/google-chrome/*/Session Storage" \ + --exclude="/home/*/.config/google-chrome/*/Application Cache" \ + --exclude="/home/*/.config/google-chrome/*/History" \ + --exclude="/home/*/.config/google-chrome/*/History-journal" \ + --exclude="/home/*/.config/google-chrome/*/History Provider Cache" \ + --exclude="/home/*/.local/share/flatpak" \ + --exclude="/home/*/.var/app/com.slack.Slack" \ + --exclude="/home/*/.local/share/Trash" \ + --exclude="/home/*/.config/Microsoft/Microsoft Teams" \ + --exclude="/home/*/.wine" \ + --exclude="/home/*/.vim/bundle" \ + --exclude="/home/*/snap" \ + --exclude="/home/*/Downloads" \ + --exclude="/home/*/Nextcloud" \ + --exclude="/home/*/git" \ + --exclude="/srv/backup" \ + --verbose \ +{%- for target in pillar.restic.targets %} + {{ target }} \ +{%- endfor %} + +return_code=$? + +{%- if pillar['containers']['freeipa'] is defined %} +echo "Starting FreeIPA" +systemctl start freeipa.service +{%- endif %} + +if [[ $return_code -eq 0 ]]; then + exit 0 +fi + + +MESSAGE="$(journalctl -u restic-backup.service -p 5 --since today)" +curl "$GOTIFY_URL/message?token=$GOTIFY_TOKEN" -F "title=$GOTIFY_TITLE" -F "priority=$GOTIFY_PRIO" -F "message=$MESSAGE" + diff --git a/restic/files/mysql-backup.sh b/restic/files/mysql-backup.sh new file mode 100644 index 0000000..da88d92 --- /dev/null +++ b/restic/files/mysql-backup.sh @@ -0,0 +1,16 @@ +#!/bin/bash +umask 0077 +BACKUP_DIR={{ pillar.containers.mariadb.backup_dir }} +databases=$(podman exec -it mariadb mysql -B -u root -p{{ pillar.containers.mariadb.env.MYSQL_ROOT_PASSWORD }} -e "SHOW DATABASES;" | tr -d "| " | grep -v Database) + +for db in ${databases[@]}; do + db=${db::-1} + if [[ "$db" != "information_schema" ]] && [[ "$db" != "performance_schema" ]] && [[ "$db" != "mysql" ]] && [[ "$db" != _* ]] && [[ "$db" != "sys" ]]; then + echo "Dumping database: $db" + podman exec -it mariadb mysqldump -u root -p{{ pillar.containers.mariadb.env.MYSQL_ROOT_PASSWORD }} --databases $db | gzip > ${BACKUP_DIR}/$(date +"%Y-%m-%d_%H-%M-%S")_$db-sql.gz + fi +done +# Delete the files older than 3 days +echo "removing old mysql dumps" +find $BACKUP_DIR/* -type f -name *-sql.gz -mtime +3 -exec rm {} \; +umask 0022 diff --git a/restic/files/restic-backup.service.jinja b/restic/files/restic-backup.service.jinja new file mode 100644 index 0000000..cfac81d --- /dev/null +++ b/restic/files/restic-backup.service.jinja @@ -0,0 +1,11 @@ +[Unit] +Description=Run restic backup + +[Service] +Type=oneshot +User={{ pillar.restic.user }} +ExecStart={{ salt['user.info'](pillar.restic.user).home }}/bin/backup.sh + +[Install] +WantedBy=default.target + diff --git a/restic/files/restic-backup.timer.jinja b/restic/files/restic-backup.timer.jinja new file mode 100644 index 0000000..bf3654a --- /dev/null +++ b/restic/files/restic-backup.timer.jinja @@ -0,0 +1,11 @@ +[Unit] +Description=Restic backup timer + +[Timer] +OnCalendar={{ pillar.restic.OnCalendar }} +RandomizedDelaySec=300 +Unit=restic-backup.service + +[Install] +WantedBy=timers.target + diff --git a/restic/files/restic.automount.jinja b/restic/files/restic.automount.jinja new file mode 100644 index 0000000..0ec3a96 --- /dev/null +++ b/restic/files/restic.automount.jinja @@ -0,0 +1,8 @@ +[Unit] +Description=Automount for restic repository + +[Automount] +Where={{ pillar['restic']['mount'] }} + +[Install] +WantedBy=multi-user.target diff --git a/restic/files/restic.jinja b/restic/files/restic.jinja new file mode 100644 index 0000000..c1d8e10 --- /dev/null +++ b/restic/files/restic.jinja @@ -0,0 +1,23 @@ +#!/bin/bash +if [[ "$EUID" -ne 0 ]]; then + echo "Needs to be run as root" + exit 1 +fi + +podman run --rm \ + --name=restic \ + --hostname="$HOSTNAME" \ + -v /root/.restic.password:/restic-password:ro \ +{%- if pillar.restic.repository is defined %} + -v {{ pillar.restic.mount }}{{ pillar.restic.suffix }}:/repo \ +{%- endif %} + -v /root:/root \ + {{ pillar.restic.image.url }}:{{ pillar.restic.image.tag }} \ + restic \ + --password-file=/restic-password \ +{%- if pillar.restic.repository is defined %} + --repo=/repo \ +{%- else %} + -r sftp:{{ pillar.restic.user }}@{{ pillar.restic.host }}:{{ pillar.restic.mount }}{{ pillar.restic.suffix }} \ +{%- endif %} + "$@" diff --git a/restic/files/restic.mount.jinja b/restic/files/restic.mount.jinja new file mode 100644 index 0000000..08730b3 --- /dev/null +++ b/restic/files/restic.mount.jinja @@ -0,0 +1,11 @@ +[Unit] +Description=Mount the USB disk used as restic repository + +[Mount] +What=/dev/disk/by-uuid/{{ pillar['restic']['repository']['disk_uuid'] }} +Where={{ pillar['restic']['mount'] }} +Type=auto +Options=defaults + +[Install] +WantedBy=multi-user.target diff --git a/restic/init.sls b/restic/init.sls new file mode 100644 index 0000000..3f512cd --- /dev/null +++ b/restic/init.sls @@ -0,0 +1,77 @@ +{%- if pillar['containers']['mariadb'] is defined %} +Create mariadb dump script: + file.managed: + - name: /root/bin/mysql-backup.sh + - source: salt://restic/files/mysql-backup.sh + - template: jinja + - user: root + - group: root + - mode: "0700" + +Create backup destination for mariadn: + file.directory: + - name: {{ pillar.containers.mariadb.backup_dir }} + - user: root + - group: root + - mode: "0700" + +{%- endif %} + +Create restiv password-file: + file.managed: + - name: /root/.restic.password + - contents: + - {{ pillar.restic.password }} + - user: root + - group: root + - mode: "0600" + +Create restic script: + file.managed: + - name: /usr/local/bin/restic + - source: salt://restic/files/restic.jinja + - template: jinja + - user: root + - group: root + - mode: "0755" + +Create backup script: + file.managed: + - name: /root/bin/backup.sh + - source: salt://restic/files/backup.sh.jinja + - template: jinja + - user: root + - group: root + - mode: "0700" +Create the restic backup service unit: + file.managed: + - name: /etc/systemd/system/restic-backup.service + - source: salt://restic/files/restic-backup.service.jinja + - template: jinja + - user: root + - group: root + - mode: "0644" + +Create the restic backup timer: + file.managed: + - name: /etc/systemd/system/restic-backup.timer + - source: salt://restic/files/restic-backup.timer.jinja + - template: jinja + - user: root + - group: root + - mode: "0644" + +Run systemctl daemon reload for restic: + cmd.run: + - name: systemctl daemon-reload + - onchanges: + - file: Create the restic backup service unit + - file: Create the restic backup timer + +Start the restic backup timer: + service.running: + - name: restic-backup.timer + - enable: True + - onchanges: + - cmd: Run systemctl daemon reload for restic + diff --git a/restic/repository.sls b/restic/repository.sls new file mode 100644 index 0000000..abda925 --- /dev/null +++ b/restic/repository.sls @@ -0,0 +1,39 @@ +{% set escape_command = [ "systemd-escape -p", pillar['restic']['mount']]|join(" ") %} +{% set unit_name = salt['cmd.shell'](escape_command) %} + +Create restic repository mount unit: + file.managed: + - name: /etc/systemd/system/{{ unit_name }}.mount + - source: salt://restic/files/restic.mount.jinja + - template: jinja + - user: root + - group: root + - mode: "0644" + +Create restic repository automount unit: + file.managed: + - name: /etc/systemd/system/{{ unit_name }}.automount + - source: salt://restic/files/restic.automount.jinja + - template: jinja + - user: root + - group: root + - mode: "0644" + +Run systemd daemon reload for repository: + cmd.run: + - name: systemctl daemon-reload + - onchanges: + - file: Create restic repository mount unit + - file: Create restic repository automount unit + +Start restic repository automount unit: + service.running: + - name: {{ unit_name }}.automount + - enable: True + +Add autherized keys for root-restic: + ssh_auth.present: + - user: root + - enc: ssh-ed25519 + - comment: restic-backup + - names: {{ pillar.restic.repository.auth_keys }} diff --git a/top.sls b/top.sls new file mode 100644 index 0000000..fc6c940 --- /dev/null +++ b/top.sls @@ -0,0 +1,8 @@ +base: + '*': + - common.users + - pod + - containers + - restic + 'woody': + - wifi-voucher diff --git a/wifi-voucher/files/voucher.env.jinja b/wifi-voucher/files/voucher.env.jinja new file mode 100644 index 0000000..813ad65 --- /dev/null +++ b/wifi-voucher/files/voucher.env.jinja @@ -0,0 +1,3 @@ +{% for key, value in pillar['voucher']['config'].items() -%} +{{ key }}="{{ value }}" +{% endfor -%} diff --git a/wifi-voucher/files/wifi-voucher.service b/wifi-voucher/files/wifi-voucher.service new file mode 100644 index 0000000..389d2b0 --- /dev/null +++ b/wifi-voucher/files/wifi-voucher.service @@ -0,0 +1,13 @@ +[Unit] +Description=WIFI-Voucher system for my guest network +After=network.target + +[Service] +Type=simple +Restart=always +User=voucher +ExecStart=/usr/local/bin/voucher +EnvironmentFile=/etc/default/wifi-voucher + +[Install] +WantedBy=multi-user.target diff --git a/wifi-voucher/init.sls b/wifi-voucher/init.sls new file mode 100644 index 0000000..b320d2d --- /dev/null +++ b/wifi-voucher/init.sls @@ -0,0 +1,4 @@ +include: + - wifi-voucher.packages + - wifi-voucher.user + - wifi-voucher.service diff --git a/wifi-voucher/packages.sls b/wifi-voucher/packages.sls new file mode 100644 index 0000000..e183e1a --- /dev/null +++ b/wifi-voucher/packages.sls @@ -0,0 +1,9 @@ +Install CUPS and bsd tools: + pkg.installed: + - pkgs: + - cups + - cups-bsd + +Install wifi-voucher: + pip.installed: + - name: https://git.rre.nu/jonas/wifi-voucher/raw/branch/main/dist/voucher-{{ pillar['voucher']['version'] }}-py39-none-any.whl diff --git a/wifi-voucher/remove.sls b/wifi-voucher/remove.sls new file mode 100644 index 0000000..c1f6c70 --- /dev/null +++ b/wifi-voucher/remove.sls @@ -0,0 +1,4 @@ +remove voucher user: + user.absent: + - name: voucher + - purge: True diff --git a/wifi-voucher/service.sls b/wifi-voucher/service.sls new file mode 100644 index 0000000..d8ac9ba --- /dev/null +++ b/wifi-voucher/service.sls @@ -0,0 +1,31 @@ +WIFI-Voucher Unit environment file: + file.managed: + - name: /etc/default/wifi-voucher + - source: salt://wifi-voucher/files/voucher.env.jinja + - template: jinja + - user: root + - group: voucher + - mode: "0640" + +WIFI-Voucher unit service file: + file.managed: + - name: /etc/systemd/system/wifi-voucher.service + - source: salt://wifi-voucher/files/wifi-voucher.service + - user: root + - group: root + - mode: "0644" + +SystemD Reload: + cmd.run: + - name: systemctl --system daemon-reload + - onchanges: + - file: WIFI-Voucher unit service file + +Start wifi-voucher: + service.running: + - name: wifi-voucher + - enable: True + - watch: + - file: WIFI-Voucher Unit environment file + - file: WIFI-Voucher unit service file + - cmd: SystemD Reload diff --git a/wifi-voucher/user.sls b/wifi-voucher/user.sls new file mode 100644 index 0000000..e37e689 --- /dev/null +++ b/wifi-voucher/user.sls @@ -0,0 +1,8 @@ +WIFI-Voucher user: + user.present: + - name: voucher + - fullname: wifi-voucher system + - home: /home/voucher + - shell: /bin/bash + - groups: + - gpio