Compare commits
37 commits
b384357636
...
3602c50ae0
Author | SHA1 | Date | |
---|---|---|---|
3602c50ae0 | |||
ffb4c5015c | |||
9d75c90e36 | |||
ba1fdb69c1 | |||
b6a9759dc7 | |||
7b13d9d4b6 | |||
bff3b9546e | |||
1e0e4fce65 | |||
68355bb656 | |||
39294744df | |||
db3a2375b4 | |||
87047957aa | |||
f1200e9788 | |||
9a6c3bbc3f | |||
1748cbf2de | |||
c533bf261d | |||
37b1f69e10 | |||
f27cd14e50 | |||
cc67692a95 | |||
68d5959776 | |||
9be7531243 | |||
58b3be4cfe | |||
8d7d178d9d | |||
ff0744f600 | |||
a61f97cccf | |||
2a45b0b8e9 | |||
ca4564f25d | |||
c57976a299 | |||
e87a1b23e9 | |||
fe89d1d3c3 | |||
204d3f77eb | |||
3b1e82b33f | |||
c92f1c5ed8 | |||
54ba01d8cd | |||
d015c77ffa | |||
d1c0e9d7f9 | |||
4f6ea4eb8c |
46 changed files with 1013 additions and 546 deletions
|
@ -46,6 +46,7 @@ in
|
||||||
};
|
};
|
||||||
}) // (if secretConfig.shared then let
|
}) // (if secretConfig.shared then let
|
||||||
secretFile = "${svcName}-${secretName}.age";
|
secretFile = "${svcName}-${secretName}.age";
|
||||||
|
snakeoilFile = "${svcName}-${secretName}-snakeoil.txt";
|
||||||
in {
|
in {
|
||||||
editSecret = {
|
editSecret = {
|
||||||
description = "Edit this secret";
|
description = "Edit this secret";
|
||||||
|
@ -54,7 +55,14 @@ in
|
||||||
agenix -e '${secretFile}'
|
agenix -e '${secretFile}'
|
||||||
'';
|
'';
|
||||||
};
|
};
|
||||||
} else lib.mapAttrs' (name: lib.nameValuePair "editSecretInstance-${name}") (lib.genAttrs secretConfig.nodes (node: let
|
editSnakeoil = {
|
||||||
|
description = "Edit this secret's snakeoil";
|
||||||
|
command = ''
|
||||||
|
$EDITOR "$PRJ_ROOT/cluster/secrets"/'${snakeoilFile}'
|
||||||
|
'';
|
||||||
|
};
|
||||||
|
} else lib.mkMerge [
|
||||||
|
(lib.mapAttrs' (name: lib.nameValuePair "editSecretInstance-${name}") (lib.genAttrs secretConfig.nodes (node: let
|
||||||
secretFile = "${svcName}-${secretName}-${node}.age";
|
secretFile = "${svcName}-${secretName}-${node}.age";
|
||||||
in {
|
in {
|
||||||
description = "Edit this secret for '${node}'";
|
description = "Edit this secret for '${node}'";
|
||||||
|
@ -62,7 +70,16 @@ in
|
||||||
${setupCommands secretFile [ node ]}
|
${setupCommands secretFile [ node ]}
|
||||||
agenix -e '${secretFile}'
|
agenix -e '${secretFile}'
|
||||||
'';
|
'';
|
||||||
})));
|
})))
|
||||||
|
(lib.mapAttrs' (name: lib.nameValuePair "editSnakeoilInstance-${name}") (lib.genAttrs secretConfig.nodes (node: let
|
||||||
|
snakeoilFile = "${svcName}-${secretName}-${node}-snakeoil.txt";
|
||||||
|
in {
|
||||||
|
description = "Edit this secret's snakeoil for '${node}'";
|
||||||
|
command = ''
|
||||||
|
$EDITOR "$PRJ_ROOT/cluster/secrets"/'${snakeoilFile}'
|
||||||
|
'';
|
||||||
|
})))
|
||||||
|
]);
|
||||||
};
|
};
|
||||||
}) svcConfig.secrets))
|
}) svcConfig.secrets))
|
||||||
lib.concatLists
|
lib.concatLists
|
||||||
|
|
|
@ -1,13 +0,0 @@
|
||||||
age-encryption.org/v1
|
|
||||||
-> ssh-ed25519 NO562A YQQrnpQI/qyEZugiRwsrPbW4oMYK/rlmRKAdD3JjYz4
|
|
||||||
JRGFqNc4BVflfR4WUuEOym39IhZlUI778NtOFtxE8eY
|
|
||||||
-> ssh-ed25519 5/zT0w utH25Xa9WQK9hXbKWsEWK5LJtCbhjpDX6JaomxnRaCI
|
|
||||||
2MfxxDjs0doUTVsGP9942rx1tyCYsDxhlDo1542BhKQ
|
|
||||||
-> ssh-ed25519 d3WGuA 6qD02cluQEBqEvupHf93Onlpv8QJJSl/bJm/XqyD+gQ
|
|
||||||
bLz/ULSaIW6HnPXDKD5dxCbQWv0VC2R+E5wlj7VxOc0
|
|
||||||
-> Ovax-grease ^1$]}H G4 FpDF XKHkj{
|
|
||||||
IVdVFYcVe9PoHCCqM3GG1pM6xgTZ5r8XWlkBjlQimgaDArotF4dPpsSTpyc
|
|
||||||
--- wdTYr6EpFPFsDJI0qQf74c6ce+v5ek6j+mgAx2CI9uI
|
|
||||||
ÜA³×oÈð:±‹`ÜVd±å(Kät:fk¼’}3*#MJš<4A>Áõ]ê,¤éÐÈÍ69i›l`ÛÆJKwAè8y@Ýœ¯à+&ðÖ©s]ÅÓ–›Ç>~Ší„+Úô
|
|
||||||
üÁ»<C381>qa©h<C2A9>( YÕ<17>eÇjýI•ê·/ð^å~Ý’wÊ
|
|
||||||
ÆÜßÌZî!^þRˆéÿv¾…ïk‹Êp»ÛPÌ)ý̆ÍpÓV5²F΄ÆÚÙÚÞhBÇ»ßb#Š<>´ùºãi”»¸9ìQy¹¾<C2B9>Êè‹}€ß ƒ¬E}~ZHûjmyq{òxŠ–Éôß<C3B4>"”éÀ´C#šójÿÐ.ò§yÔ£¸v¦
<0A>ÉÐòê<1“Œúâ¾ìßzâš#/êGñ?që
|
|
60
cluster/services/acme-client/augment.nix
Normal file
60
cluster/services/acme-client/augment.nix
Normal file
|
@ -0,0 +1,60 @@
|
||||||
|
{ config, pkgs, ... }:
|
||||||
|
|
||||||
|
let
|
||||||
|
lift = config;
|
||||||
|
in
|
||||||
|
|
||||||
|
{
|
||||||
|
nowhere.names = {
|
||||||
|
"acme-v02.api.letsencrypt.org" = "stepCa";
|
||||||
|
"api.buypass.com" = "stepCa";
|
||||||
|
};
|
||||||
|
|
||||||
|
nodes.nowhere = { config, ... }: {
|
||||||
|
links.stepCa.protocol = "https";
|
||||||
|
|
||||||
|
environment.etc.step-ca-password.text = "";
|
||||||
|
|
||||||
|
services = {
|
||||||
|
step-ca = {
|
||||||
|
enable = true;
|
||||||
|
address = config.links.stepCa.ipv4;
|
||||||
|
inherit (config.links.stepCa) port;
|
||||||
|
intermediatePasswordFile = "/etc/step-ca-password";
|
||||||
|
settings = {
|
||||||
|
root = "${lift.nowhere.certs.ca}/ca.pem";
|
||||||
|
crt = "${lift.nowhere.certs.intermediate}/cert.pem";
|
||||||
|
key = "${lift.nowhere.certs.intermediate}/cert-key.pem";
|
||||||
|
address = config.links.stepCa.tuple;
|
||||||
|
db = {
|
||||||
|
type = "badgerv2";
|
||||||
|
dataSource = "/var/lib/step-ca/db";
|
||||||
|
};
|
||||||
|
authority.provisioners = [
|
||||||
|
{
|
||||||
|
type = "ACME";
|
||||||
|
name = "snakeoil";
|
||||||
|
challenges = [
|
||||||
|
"dns-01"
|
||||||
|
"http-01"
|
||||||
|
];
|
||||||
|
}
|
||||||
|
];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
|
||||||
|
nginx.virtualHosts = {
|
||||||
|
"acme-v02.api.letsencrypt.org".locations."/".extraConfig = ''
|
||||||
|
rewrite /directory /acme/snakeoil/directory break;
|
||||||
|
'';
|
||||||
|
"api.buypass.com".locations."/".extraConfig = ''
|
||||||
|
rewrite /acme/directory /acme/snakeoil/directory break;
|
||||||
|
'';
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
|
||||||
|
defaults.environment.etc."dummy-secrets/acmeDnsApiKey".text = "ACME_DNS_DIRECT_STATIC_KEY=simulacrum";
|
||||||
|
defaults.environment.etc."dummy-secrets/acmeDnsDirectKey".text = "ACME_DNS_DIRECT_STATIC_KEY=simulacrum";
|
||||||
|
defaults.environment.etc."dummy-secrets/acmeDnsDbCredentials".text = "PGPASSWORD=simulacrum";
|
||||||
|
}
|
|
@ -2,5 +2,6 @@
|
||||||
services.acme-client = {
|
services.acme-client = {
|
||||||
nodes.client = [ "checkmate" "grail" "thunderskin" "VEGAS" "prophet" ];
|
nodes.client = [ "checkmate" "grail" "thunderskin" "VEGAS" "prophet" ];
|
||||||
nixos.client = ./client.nix;
|
nixos.client = ./client.nix;
|
||||||
|
simulacrum.augments = ./augment.nix;
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
|
@ -33,7 +33,7 @@
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
|
|
||||||
garage = {
|
garage = config.lib.forService "attic" {
|
||||||
keys.attic.locksmith = {
|
keys.attic.locksmith = {
|
||||||
nodes = config.services.attic.nodes.server;
|
nodes = config.services.attic.nodes.server;
|
||||||
owner = "atticd";
|
owner = "atticd";
|
||||||
|
@ -48,7 +48,7 @@
|
||||||
serverAddrs = map
|
serverAddrs = map
|
||||||
(node: depot.hours.${node}.interfaces.primary.addrPublic)
|
(node: depot.hours.${node}.interfaces.primary.addrPublic)
|
||||||
config.services.attic.nodes.server;
|
config.services.attic.nodes.server;
|
||||||
in {
|
in config.lib.forService "attic" {
|
||||||
cache.target = serverAddrs;
|
cache.target = serverAddrs;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
|
@ -6,5 +6,6 @@
|
||||||
nixos.listener = [
|
nixos.listener = [
|
||||||
./listener.nix
|
./listener.nix
|
||||||
];
|
];
|
||||||
|
simulacrum.deps = [ "consul" ];
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
|
@ -10,6 +10,13 @@ let
|
||||||
in
|
in
|
||||||
|
|
||||||
{
|
{
|
||||||
|
system.extraIncantations = {
|
||||||
|
runConsul = i: script: i.execShellWith [ config.services.consul.package ] ''
|
||||||
|
export CONSUL_HTTP_ADDR='${config.links.consulAgent.tuple}'
|
||||||
|
${script}
|
||||||
|
'';
|
||||||
|
};
|
||||||
|
|
||||||
links.consulAgent.protocol = "http";
|
links.consulAgent.protocol = "http";
|
||||||
|
|
||||||
services.consul = {
|
services.consul = {
|
||||||
|
|
|
@ -43,9 +43,6 @@ in {
|
||||||
links.localAuthoritativeDNS = {};
|
links.localAuthoritativeDNS = {};
|
||||||
|
|
||||||
age.secrets = {
|
age.secrets = {
|
||||||
acmeDnsDbCredentials = {
|
|
||||||
file = ./acme-dns-db-credentials.age;
|
|
||||||
};
|
|
||||||
acmeDnsDirectKey = {
|
acmeDnsDirectKey = {
|
||||||
file = ./acme-dns-direct-key.age;
|
file = ./acme-dns-direct-key.age;
|
||||||
};
|
};
|
||||||
|
@ -78,8 +75,12 @@ in {
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
|
|
||||||
|
services.locksmith.waitForSecrets.acme-dns = [
|
||||||
|
"patroni-acmedns"
|
||||||
|
];
|
||||||
|
|
||||||
systemd.services.acme-dns.serviceConfig.EnvironmentFile = with config.age.secrets; [
|
systemd.services.acme-dns.serviceConfig.EnvironmentFile = with config.age.secrets; [
|
||||||
acmeDnsDbCredentials.path
|
"/run/locksmith/patroni-acmedns"
|
||||||
acmeDnsDirectKey.path
|
acmeDnsDirectKey.path
|
||||||
];
|
];
|
||||||
|
|
||||||
|
|
|
@ -58,6 +58,16 @@ in
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
|
|
||||||
|
patroni = {
|
||||||
|
databases.acmedns = {};
|
||||||
|
users.acmedns = {
|
||||||
|
locksmith = {
|
||||||
|
nodes = config.services.dns.nodes.authoritative;
|
||||||
|
format = "envFile";
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
|
||||||
dns.records = {
|
dns.records = {
|
||||||
securedns.consulService = "securedns";
|
securedns.consulService = "securedns";
|
||||||
"acme-dns-challenge.internal".consulService = "acme-dns";
|
"acme-dns-challenge.internal".consulService = "acme-dns";
|
||||||
|
|
|
@ -13,25 +13,36 @@
|
||||||
nodes = server;
|
nodes = server;
|
||||||
owner = "forgejo";
|
owner = "forgejo";
|
||||||
};
|
};
|
||||||
dbCredentials.nodes = server;
|
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
|
|
||||||
ways.forge.target = let
|
ways.forge = let
|
||||||
host = builtins.head config.services.forge.nodes.server;
|
host = builtins.head config.services.forge.nodes.server;
|
||||||
in config.hostLinks.${host}.forge.url;
|
in config.lib.forService "forge" {
|
||||||
|
target = config.hostLinks.${host}.forge.url;
|
||||||
|
};
|
||||||
|
|
||||||
garage = {
|
patroni = config.lib.forService "forge" {
|
||||||
|
databases.forge = {};
|
||||||
|
users.forge.locksmith = {
|
||||||
|
nodes = config.services.forge.nodes.server;
|
||||||
|
format = "raw";
|
||||||
|
};
|
||||||
|
};
|
||||||
|
|
||||||
|
garage = config.lib.forService "forge" {
|
||||||
keys.forgejo.locksmith.nodes = config.services.forge.nodes.server;
|
keys.forgejo.locksmith.nodes = config.services.forge.nodes.server;
|
||||||
buckets.forgejo.allow.forgejo = [ "read" "write" ];
|
buckets.forgejo.allow.forgejo = [ "read" "write" ];
|
||||||
};
|
};
|
||||||
|
|
||||||
monitoring.blackbox.targets.forge = {
|
monitoring.blackbox.targets.forge = config.lib.forService "forge" {
|
||||||
address = "https://forge.${depot.lib.meta.domain}/api/v1/version";
|
address = "https://forge.${depot.lib.meta.domain}/api/v1/version";
|
||||||
module = "https2xx";
|
module = "https2xx";
|
||||||
};
|
};
|
||||||
|
|
||||||
dns.records."ssh.forge".target = map
|
dns.records = config.lib.forService "forge" {
|
||||||
|
"ssh.forge".target = map
|
||||||
(node: depot.hours.${node}.interfaces.primary.addrPublic)
|
(node: depot.hours.${node}.interfaces.primary.addrPublic)
|
||||||
config.services.forge.nodes.server;
|
config.services.forge.nodes.server;
|
||||||
|
};
|
||||||
}
|
}
|
||||||
|
|
|
@ -26,6 +26,7 @@ in
|
||||||
services.locksmith.waitForSecrets.forgejo = [
|
services.locksmith.waitForSecrets.forgejo = [
|
||||||
"garage-forgejo-id"
|
"garage-forgejo-id"
|
||||||
"garage-forgejo-secret"
|
"garage-forgejo-secret"
|
||||||
|
"patroni-forge"
|
||||||
];
|
];
|
||||||
|
|
||||||
services.forgejo = {
|
services.forgejo = {
|
||||||
|
@ -39,7 +40,7 @@ in
|
||||||
inherit (patroni) port;
|
inherit (patroni) port;
|
||||||
name = "forge";
|
name = "forge";
|
||||||
user = "forge";
|
user = "forge";
|
||||||
passwordFile = secrets.dbCredentials.path;
|
passwordFile = "/run/locksmith/patroni-forge";
|
||||||
};
|
};
|
||||||
settings = {
|
settings = {
|
||||||
DEFAULT = {
|
DEFAULT = {
|
||||||
|
|
|
@ -62,7 +62,7 @@
|
||||||
lib.unique
|
lib.unique
|
||||||
(map (x: "hci-agent-${x}"))
|
(map (x: "hci-agent-${x}"))
|
||||||
];
|
];
|
||||||
in {
|
in config.lib.forService "hercules-ci-multi-agent" {
|
||||||
keys = lib.genAttrs hciAgentKeys (lib.const {});
|
keys = lib.genAttrs hciAgentKeys (lib.const {});
|
||||||
buckets.nix-store = {
|
buckets.nix-store = {
|
||||||
allow = lib.genAttrs hciAgentKeys (lib.const [ "read" "write" ]);
|
allow = lib.genAttrs hciAgentKeys (lib.const [ "read" "write" ]);
|
||||||
|
|
25
cluster/services/incandescence/default.nix
Normal file
25
cluster/services/incandescence/default.nix
Normal file
|
@ -0,0 +1,25 @@
|
||||||
|
{ config, ... }:
|
||||||
|
|
||||||
|
{
|
||||||
|
imports = [
|
||||||
|
./options.nix
|
||||||
|
./simulacrum/test-data.nix
|
||||||
|
];
|
||||||
|
|
||||||
|
services.incandescence = {
|
||||||
|
nodes = {
|
||||||
|
provider = config.services.consul.nodes.agent;
|
||||||
|
};
|
||||||
|
nixos = {
|
||||||
|
provider = [
|
||||||
|
./provider.nix
|
||||||
|
./provider-options.nix
|
||||||
|
];
|
||||||
|
};
|
||||||
|
simulacrum = {
|
||||||
|
enable = true;
|
||||||
|
deps = [ "consul" "locksmith" ];
|
||||||
|
settings = ./simulacrum/test.nix;
|
||||||
|
};
|
||||||
|
};
|
||||||
|
}
|
22
cluster/services/incandescence/options.nix
Normal file
22
cluster/services/incandescence/options.nix
Normal file
|
@ -0,0 +1,22 @@
|
||||||
|
{ lib, ... }:
|
||||||
|
|
||||||
|
let
|
||||||
|
inherit (lib) mkOption;
|
||||||
|
inherit (lib.types) attrsOf listOf submodule str;
|
||||||
|
in
|
||||||
|
|
||||||
|
{
|
||||||
|
options.incandescence = {
|
||||||
|
providers = mkOption {
|
||||||
|
type = attrsOf (submodule ({ name, ... }: {
|
||||||
|
options = {
|
||||||
|
objects = mkOption {
|
||||||
|
type = attrsOf (listOf str);
|
||||||
|
default = { };
|
||||||
|
};
|
||||||
|
};
|
||||||
|
}));
|
||||||
|
default = { };
|
||||||
|
};
|
||||||
|
};
|
||||||
|
}
|
72
cluster/services/incandescence/provider-options.nix
Normal file
72
cluster/services/incandescence/provider-options.nix
Normal file
|
@ -0,0 +1,72 @@
|
||||||
|
{ lib, ... }:
|
||||||
|
|
||||||
|
let
|
||||||
|
inherit (lib) mkEnableOption mkOption;
|
||||||
|
inherit (lib.types) attrsOf functionTo ints listOf nullOr package submodule str;
|
||||||
|
in
|
||||||
|
|
||||||
|
{
|
||||||
|
options.services.incandescence = {
|
||||||
|
providers = mkOption {
|
||||||
|
type = attrsOf (submodule ({ name, ... }: {
|
||||||
|
options = {
|
||||||
|
locksmith = mkEnableOption "Locksmith integration";
|
||||||
|
|
||||||
|
wantedBy = mkOption {
|
||||||
|
type = listOf str;
|
||||||
|
};
|
||||||
|
|
||||||
|
partOf = mkOption {
|
||||||
|
type = listOf str;
|
||||||
|
};
|
||||||
|
|
||||||
|
wants = mkOption {
|
||||||
|
type = listOf str;
|
||||||
|
default = [ ];
|
||||||
|
};
|
||||||
|
|
||||||
|
after = mkOption {
|
||||||
|
type = listOf str;
|
||||||
|
default = [ ];
|
||||||
|
};
|
||||||
|
|
||||||
|
packages = mkOption {
|
||||||
|
type = listOf package;
|
||||||
|
default = [ ];
|
||||||
|
};
|
||||||
|
|
||||||
|
formulae = mkOption {
|
||||||
|
type = attrsOf (submodule ({ ... }: {
|
||||||
|
options = {
|
||||||
|
deps = mkOption {
|
||||||
|
type = listOf str;
|
||||||
|
default = [ ];
|
||||||
|
};
|
||||||
|
|
||||||
|
create = mkOption {
|
||||||
|
type = functionTo str;
|
||||||
|
};
|
||||||
|
|
||||||
|
change = mkOption {
|
||||||
|
type = nullOr (functionTo str);
|
||||||
|
default = null;
|
||||||
|
};
|
||||||
|
|
||||||
|
destroy = mkOption {
|
||||||
|
type = str;
|
||||||
|
};
|
||||||
|
|
||||||
|
destroyAfterDays = mkOption {
|
||||||
|
type = ints.unsigned;
|
||||||
|
default = 0;
|
||||||
|
};
|
||||||
|
};
|
||||||
|
}));
|
||||||
|
default = { };
|
||||||
|
};
|
||||||
|
};
|
||||||
|
}));
|
||||||
|
default = { };
|
||||||
|
};
|
||||||
|
};
|
||||||
|
}
|
144
cluster/services/incandescence/provider.nix
Normal file
144
cluster/services/incandescence/provider.nix
Normal file
|
@ -0,0 +1,144 @@
|
||||||
|
{ cluster, config, lib, ... }:
|
||||||
|
|
||||||
|
let
|
||||||
|
inherit (lib) concatStringsSep escapeShellArg flatten filter filterAttrs length mapAttrs mapAttrs' mapAttrsToList mkIf mkMerge pipe stringToCharacters;
|
||||||
|
|
||||||
|
cfg = config.services.incandescence;
|
||||||
|
clusterCfg = cluster.config.incandescence;
|
||||||
|
in
|
||||||
|
|
||||||
|
{
|
||||||
|
systemd.services = pipe cfg.providers [
|
||||||
|
(mapAttrsToList (provider: providerConfig: pipe providerConfig.formulae [
|
||||||
|
(mapAttrsToList (formula: formulaConfig: let
|
||||||
|
kvRoot = "services/incandescence/providers/${provider}/formulae/${formula}";
|
||||||
|
time = "$(date +%s)";
|
||||||
|
in {
|
||||||
|
"ignite-${provider}-${formula}-create" = {
|
||||||
|
description = "Ignite Creation: ${provider} - ${formula}";
|
||||||
|
wantedBy = [ "incandescence-${provider}.target" ];
|
||||||
|
before = [ "incandescence-${provider}.target" ];
|
||||||
|
wants = providerConfig.wants ++ map (dep: "ignite-${provider}-${dep}-create.service") formulaConfig.deps;
|
||||||
|
after = providerConfig.after ++ map (dep: "ignite-${provider}-${dep}-create.service") formulaConfig.deps;
|
||||||
|
serviceConfig.Type = "oneshot";
|
||||||
|
distributed.enable = true;
|
||||||
|
path = [ config.services.consul.package ] ++ providerConfig.packages;
|
||||||
|
script = pipe clusterCfg.providers.${provider}.objects.${formula} [
|
||||||
|
(map (object: ''
|
||||||
|
if ! consul kv get ${kvRoot}/${object}/alive >/dev/null; then
|
||||||
|
echo "Create ${formula}: ${object}"
|
||||||
|
if (
|
||||||
|
${formulaConfig.create object}
|
||||||
|
)
|
||||||
|
then
|
||||||
|
consul kv put ${kvRoot}/${object}/alive true
|
||||||
|
consul kv delete ${kvRoot}/${object}/destroyOn
|
||||||
|
else
|
||||||
|
echo "Creation failed: ${object}"
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
''))
|
||||||
|
(concatStringsSep "\n")
|
||||||
|
(script: if script == "" then ''
|
||||||
|
echo "Nothing to create"
|
||||||
|
'' else script)
|
||||||
|
];
|
||||||
|
};
|
||||||
|
"ignite-${provider}-${formula}-change" = mkIf (formulaConfig.change != null) {
|
||||||
|
description = "Ignite Change: ${provider} - ${formula}";
|
||||||
|
wantedBy = [ "incandescence-${provider}.target" ];
|
||||||
|
before = [ "incandescence-${provider}.target" ];
|
||||||
|
wants = providerConfig.wants ++ [ "ignite-${provider}-${formula}-create.service" ] ++ map (dep: "ignite-${provider}-${dep}-change.service") formulaConfig.deps;
|
||||||
|
after = providerConfig.after ++ [ "ignite-${provider}-${formula}-create.service" ] ++ map (dep: "ignite-${provider}-${dep}-change.service") formulaConfig.deps;
|
||||||
|
serviceConfig.Type = "oneshot";
|
||||||
|
distributed.enable = true;
|
||||||
|
path = [ config.services.consul.package ] ++ providerConfig.packages;
|
||||||
|
script = pipe clusterCfg.providers.${provider}.objects.${formula} [
|
||||||
|
(map (object: ''
|
||||||
|
echo "Change ${formula}: ${object}"
|
||||||
|
(
|
||||||
|
${formulaConfig.change object}
|
||||||
|
) || echo "Change failed: ${object}"
|
||||||
|
''))
|
||||||
|
(concatStringsSep "\n")
|
||||||
|
(script: if script == "" then ''
|
||||||
|
echo "Nothing to change"
|
||||||
|
'' else script)
|
||||||
|
];
|
||||||
|
};
|
||||||
|
"ignite-${provider}-${formula}-destroy" = {
|
||||||
|
description = "Ignite Destruction: ${provider} - ${formula}";
|
||||||
|
wantedBy = [ "incandescence-${provider}.target" ] ++ map (dep: "ignite-${provider}-${dep}-destroy.service") formulaConfig.deps;
|
||||||
|
before = [ "incandescence-${provider}.target" ] ++ map (dep: "ignite-${provider}-${dep}-destroy.service") formulaConfig.deps;
|
||||||
|
wants = providerConfig.wants ++ [ "ignite-${provider}-${formula}-change.service" ];
|
||||||
|
after = providerConfig.after ++ [ "ignite-${provider}-${formula}-change.service" ];
|
||||||
|
serviceConfig.Type = "oneshot";
|
||||||
|
distributed.enable = true;
|
||||||
|
path = [ config.services.consul.package ] ++ providerConfig.packages;
|
||||||
|
script = let
|
||||||
|
fieldNum = pipe kvRoot [
|
||||||
|
stringToCharacters
|
||||||
|
(filter (x: x == "/"))
|
||||||
|
length
|
||||||
|
(builtins.add 2)
|
||||||
|
toString
|
||||||
|
];
|
||||||
|
keyFilter = pipe clusterCfg.providers.${provider}.objects.${formula} [
|
||||||
|
(map (x: escapeShellArg "^${x}$"))
|
||||||
|
(concatStringsSep " \\\n -e ")
|
||||||
|
];
|
||||||
|
destroyAfterDays = toString formulaConfig.destroyAfterDays;
|
||||||
|
in ''
|
||||||
|
consul kv get --keys ${kvRoot}/ | cut -d/ -f${fieldNum} | grep -v -e ${keyFilter} | while read object; do
|
||||||
|
if consul kv get ${kvRoot}/$object/alive >/dev/null; then
|
||||||
|
destroyOn="$(consul kv get ${kvRoot}/$object/destroyOn || true)"
|
||||||
|
if [[ -z "$destroyOn" && "${destroyAfterDays}" -ne 0 ]]; then
|
||||||
|
echo "Schedule ${formula} for destruction in ${destroyAfterDays} days: $object"
|
||||||
|
consul kv put ${kvRoot}/$object/destroyOn "$((${time} + 86400 * ${destroyAfterDays}))"
|
||||||
|
elif [[ "${destroyAfterDays}" -eq 0 || "${time}" -ge "$destroyOn" ]]; then
|
||||||
|
echo "Destroy ${formula}: $object"
|
||||||
|
export OBJECT="$object"
|
||||||
|
if (
|
||||||
|
${formulaConfig.destroy}
|
||||||
|
)
|
||||||
|
then
|
||||||
|
consul kv delete --recurse ${kvRoot}/$object
|
||||||
|
else
|
||||||
|
echo "Destruction failed: $object"
|
||||||
|
fi
|
||||||
|
else
|
||||||
|
echo "Scheduled for destruction on $destroyOn (now: ${time})"
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
'';
|
||||||
|
};
|
||||||
|
}))
|
||||||
|
]))
|
||||||
|
flatten
|
||||||
|
mkMerge
|
||||||
|
];
|
||||||
|
|
||||||
|
systemd.targets = mapAttrs' (provider: providerConfig: {
|
||||||
|
name = "incandescence-${provider}";
|
||||||
|
value = {
|
||||||
|
description = "An Incandescence | ${provider}";
|
||||||
|
inherit (providerConfig) wantedBy partOf;
|
||||||
|
};
|
||||||
|
}) cfg.providers;
|
||||||
|
|
||||||
|
services.locksmith.providers = mapAttrs (provider: providerConfig: {
|
||||||
|
wantedBy = [ "incandescence-${provider}.target" ];
|
||||||
|
after = [ "incandescence-${provider}.target" ];
|
||||||
|
}) (filterAttrs (_: providerConfig: providerConfig.locksmith) cfg.providers);
|
||||||
|
|
||||||
|
system.ascensions = mapAttrs' (provider: providerConfig: {
|
||||||
|
name = "incandescence-${provider}";
|
||||||
|
value = {
|
||||||
|
distributed = true;
|
||||||
|
requiredBy = map (formula: "ignite-${provider}-${formula}-create.service") (lib.attrNames providerConfig.formulae);
|
||||||
|
before = map (formula: "ignite-${provider}-${formula}-create.service") (lib.attrNames providerConfig.formulae);
|
||||||
|
incantations = lib.mkDefault (i: []);
|
||||||
|
};
|
||||||
|
}) cfg.providers;
|
||||||
|
}
|
8
cluster/services/incandescence/simulacrum/test-data.nix
Normal file
8
cluster/services/incandescence/simulacrum/test-data.nix
Normal file
|
@ -0,0 +1,8 @@
|
||||||
|
{ config, lib, ... }:
|
||||||
|
{
|
||||||
|
incandescence = lib.mkIf config.simulacrum {
|
||||||
|
providers = config.lib.forService "incandescence" {
|
||||||
|
test.objects.example = [ "example1" "example2" ];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
}
|
47
cluster/services/incandescence/simulacrum/test.nix
Normal file
47
cluster/services/incandescence/simulacrum/test.nix
Normal file
|
@ -0,0 +1,47 @@
|
||||||
|
{ cluster, lib, ... }:
|
||||||
|
|
||||||
|
let
|
||||||
|
providers = lib.take 2 cluster.config.services.incandescence.nodes.provider;
|
||||||
|
in
|
||||||
|
|
||||||
|
{
|
||||||
|
nodes = lib.genAttrs providers (lib.const {
|
||||||
|
services.incandescence.providers.test = {
|
||||||
|
wantedBy = [ "multi-user.target" ];
|
||||||
|
partOf = [ ];
|
||||||
|
formulae.example = {
|
||||||
|
create = x: "consul kv put testData/${x} ${x}";
|
||||||
|
destroy = "consul kv delete testData/$OBJECT";
|
||||||
|
};
|
||||||
|
};
|
||||||
|
});
|
||||||
|
|
||||||
|
testScript = ''
|
||||||
|
import json
|
||||||
|
nodeNames = json.loads('${builtins.toJSON providers}')
|
||||||
|
nodes = [ n for n in machines if n.name in nodeNames ]
|
||||||
|
|
||||||
|
start_all()
|
||||||
|
|
||||||
|
consulConfig = json.loads(nodes[0].succeed("cat /etc/consul.json"))
|
||||||
|
addr = consulConfig["addresses"]["http"]
|
||||||
|
port = consulConfig["ports"]["http"]
|
||||||
|
setEnv = f"CONSUL_HTTP_ADDR={addr}:{port}"
|
||||||
|
|
||||||
|
with subtest("should create objects"):
|
||||||
|
for node in nodes:
|
||||||
|
node.wait_for_unit("incandescence-test.target")
|
||||||
|
nodes[0].succeed(f"[[ $({setEnv} consul kv get testData/example1) == example1 ]]")
|
||||||
|
nodes[0].succeed(f"[[ $({setEnv} consul kv get testData/example2) == example2 ]]")
|
||||||
|
|
||||||
|
with subtest("should destroy objects"):
|
||||||
|
nodes[0].succeed(f"{setEnv} consul kv put testData/example3 example3")
|
||||||
|
nodes[0].succeed(f"{setEnv} consul kv put services/incandescence/providers/test/formulae/example/example3/alive true")
|
||||||
|
nodes[1].succeed(f"{setEnv} consul kv get testData/example3")
|
||||||
|
for node in nodes:
|
||||||
|
node.systemctl("isolate default")
|
||||||
|
for node in nodes:
|
||||||
|
node.wait_for_unit("incandescence-test.target")
|
||||||
|
nodes[0].fail(f"{setEnv} consul kv get testData/example3")
|
||||||
|
'';
|
||||||
|
}
|
|
@ -14,5 +14,6 @@
|
||||||
./provider.nix
|
./provider.nix
|
||||||
];
|
];
|
||||||
};
|
};
|
||||||
|
simulacrum.deps = [ "chant" "consul" ];
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
|
@ -28,6 +28,10 @@ in
|
||||||
command = mkOption {
|
command = mkOption {
|
||||||
type = types.coercedTo types.package (package: "${package}") types.str;
|
type = types.coercedTo types.package (package: "${package}") types.str;
|
||||||
};
|
};
|
||||||
|
checkUpdate = mkOption {
|
||||||
|
type = types.coercedTo types.package (package: "${package}") types.str;
|
||||||
|
default = "true";
|
||||||
|
};
|
||||||
owner = mkOption {
|
owner = mkOption {
|
||||||
type = types.str;
|
type = types.str;
|
||||||
default = "root";
|
default = "root";
|
||||||
|
@ -72,20 +76,27 @@ in
|
||||||
activeNodes = lib.unique (lib.flatten (lib.mapAttrsToList (_: secret: secret.nodes) activeSecrets));
|
activeNodes = lib.unique (lib.flatten (lib.mapAttrsToList (_: secret: secret.nodes) activeSecrets));
|
||||||
secretNames = map (name: "${providerRoot}-${name}/") (lib.attrNames activeSecrets);
|
secretNames = map (name: "${providerRoot}-${name}/") (lib.attrNames activeSecrets);
|
||||||
|
|
||||||
createSecret = { path, nodes, owner, mode, group, command }: ''
|
createSecret = { path, nodes, owner, mode, group, command, checkUpdate }: ''
|
||||||
|
if (${checkUpdate}); then
|
||||||
consul kv put ${lib.escapeShellArg path}/mode ${lib.escapeShellArg mode}
|
consul kv put ${lib.escapeShellArg path}/mode ${lib.escapeShellArg mode}
|
||||||
consul kv put ${lib.escapeShellArg path}/owner ${lib.escapeShellArg owner}
|
consul kv put ${lib.escapeShellArg path}/owner ${lib.escapeShellArg owner}
|
||||||
consul kv put ${lib.escapeShellArg path}/group ${lib.escapeShellArg group}
|
consul kv put ${lib.escapeShellArg path}/group ${lib.escapeShellArg group}
|
||||||
|
secret="$(mktemp -ut)"
|
||||||
|
(${command}) > "$secret"
|
||||||
${lib.concatStringsSep "\n" (map (node: ''
|
${lib.concatStringsSep "\n" (map (node: ''
|
||||||
consul kv put ${lib.escapeShellArg path}/recipient/${node} "$( (${command}) | age --encrypt --armor -r ${lib.escapeShellArg depot.hours.${node}.ssh.id.publicKey})"
|
consul kv put ${lib.escapeShellArg path}/recipient/${node} "$(age < "$secret" --encrypt --armor -r ${lib.escapeShellArg depot.hours.${node}.ssh.id.publicKey})"
|
||||||
'') nodes)}
|
'') nodes)}
|
||||||
|
else
|
||||||
|
echo Skipping update for ${lib.escapeShellArg path}
|
||||||
|
fi
|
||||||
'';
|
'';
|
||||||
in ''
|
in ''
|
||||||
# create/update secrets
|
# create/update secrets
|
||||||
|
umask 77
|
||||||
${lib.pipe activeSecrets [
|
${lib.pipe activeSecrets [
|
||||||
(lib.mapAttrsToList (secretName: secretConfig: createSecret {
|
(lib.mapAttrsToList (secretName: secretConfig: createSecret {
|
||||||
path = "${providerRoot}-${secretName}";
|
path = "${providerRoot}-${secretName}";
|
||||||
inherit (secretConfig) nodes mode owner group command;
|
inherit (secretConfig) nodes mode owner group command checkUpdate;
|
||||||
}))
|
}))
|
||||||
(lib.concatStringsSep "\n")
|
(lib.concatStringsSep "\n")
|
||||||
]}
|
]}
|
||||||
|
|
|
@ -72,7 +72,7 @@ in
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
|
|
||||||
garage = {
|
garage = config.lib.forService "monitoring" {
|
||||||
keys = {
|
keys = {
|
||||||
loki-ingest.locksmith = {
|
loki-ingest.locksmith = {
|
||||||
nodes = config.services.monitoring.nodes.logging;
|
nodes = config.services.monitoring.nodes.logging;
|
||||||
|
|
91
cluster/services/patroni/create-databases.nix
Normal file
91
cluster/services/patroni/create-databases.nix
Normal file
|
@ -0,0 +1,91 @@
|
||||||
|
{ cluster, config, lib, pkgs, ... }:
|
||||||
|
|
||||||
|
let
|
||||||
|
inherit (cluster.config.services.patroni) secrets;
|
||||||
|
|
||||||
|
patroni = cluster.config.links.patroni-pg-access;
|
||||||
|
|
||||||
|
cfg = cluster.config.patroni;
|
||||||
|
|
||||||
|
writeQueryFile = pkgs.writeText "patroni-query.sql";
|
||||||
|
|
||||||
|
psqlRunFile = file: ''
|
||||||
|
export PGPASSWORD="$(< ${secrets.PATRONI_SUPERUSER_PASSWORD.path})"
|
||||||
|
while ! ${config.services.patroni.postgresqlPackage}/bin/psql 'host=${patroni.ipv4} port=${patroni.portStr} dbname=postgres user=postgres' --tuples-only --csv --file="${file}"; do
|
||||||
|
sleep 3
|
||||||
|
done
|
||||||
|
'';
|
||||||
|
|
||||||
|
psql = query: psqlRunFile (writeQueryFile query);
|
||||||
|
|
||||||
|
psqlSecret = getSecret: queryTemplate: let
|
||||||
|
queryTemplateFile = writeQueryFile queryTemplate;
|
||||||
|
in ''
|
||||||
|
umask 77
|
||||||
|
secretFile="$(mktemp -ut patroniSecret.XXXXXXXXXXXXXXXX)"
|
||||||
|
queryFile="$(mktemp -ut patroniQuery.XXXXXXXXXXXXXXXX)"
|
||||||
|
trap "rm -f $secretFile $queryFile" EXIT
|
||||||
|
${getSecret} > "$secretFile"
|
||||||
|
cp --no-preserve=mode ${queryTemplateFile} "$queryFile"
|
||||||
|
${pkgs.replace-secret}/bin/replace-secret '@SECRET@' "$secretFile" "$queryFile"
|
||||||
|
${psqlRunFile "$queryFile"}
|
||||||
|
'';
|
||||||
|
|
||||||
|
genPassword = pkgs.writeShellScript "patroni-generate-user-password" ''
|
||||||
|
umask 77
|
||||||
|
base64 -w0 /dev/urandom | tr -d /+ | head -c256 | tee "/run/keys/locksmith-provider-patroni-$1"
|
||||||
|
'';
|
||||||
|
in
|
||||||
|
|
||||||
|
{
|
||||||
|
services.incandescence.providers.patroni = lib.mkIf config.services.haproxy.enable {
|
||||||
|
locksmith = true;
|
||||||
|
wantedBy = [ "patroni.service" "multi-user.target" ];
|
||||||
|
partOf = [ "patroni.service" ];
|
||||||
|
wants = [ "postgresql.service" ];
|
||||||
|
after = [ "postgresql.service" ];
|
||||||
|
|
||||||
|
formulae = {
|
||||||
|
user = {
|
||||||
|
destroyAfterDays = 0;
|
||||||
|
create = user: psqlSecret "${genPassword} ${user}" ''
|
||||||
|
CREATE USER ${user} PASSWORD '@SECRET@';
|
||||||
|
'';
|
||||||
|
destroy = psqlSecret "printenv OBJECT" ''
|
||||||
|
DROP USER @SECRET@;
|
||||||
|
'';
|
||||||
|
};
|
||||||
|
database = {
|
||||||
|
destroyAfterDays = 30;
|
||||||
|
deps = [ "user" ];
|
||||||
|
create = db: psql ''
|
||||||
|
CREATE DATABASE ${db} OWNER ${cfg.databases.${db}.owner};
|
||||||
|
'';
|
||||||
|
destroy = psqlSecret "printenv OBJECT" ''
|
||||||
|
DROP DATABASE @SECRET@;
|
||||||
|
'';
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
|
||||||
|
services.locksmith.providers.patroni = lib.mkIf config.services.haproxy.enable {
|
||||||
|
secrets = lib.mapAttrs (user: userConfig: {
|
||||||
|
command = {
|
||||||
|
envFile = ''
|
||||||
|
echo "PGPASSWORD=$(cat /run/keys/locksmith-provider-patroni-${user})"
|
||||||
|
rm -f /run/keys/locksmith-provider-patroni-${user}
|
||||||
|
'';
|
||||||
|
pgpass = ''
|
||||||
|
echo "*:*:*:${user}:$(cat /run/keys/locksmith-provider-patroni-${user})"
|
||||||
|
rm -f /run/keys/locksmith-provider-patroni-${user}
|
||||||
|
'';
|
||||||
|
raw = ''
|
||||||
|
cat /run/keys/locksmith-provider-patroni-${user}
|
||||||
|
rm -f /run/keys/locksmith-provider-patroni-${user}
|
||||||
|
'';
|
||||||
|
}.${userConfig.locksmith.format};
|
||||||
|
checkUpdate = "test -e /run/keys/locksmith-provider-patroni-${user}";
|
||||||
|
inherit (userConfig.locksmith) nodes;
|
||||||
|
}) cfg.users;
|
||||||
|
};
|
||||||
|
}
|
|
@ -1,6 +1,12 @@
|
||||||
{ config, lib, ... }:
|
{ config, ... }:
|
||||||
|
|
||||||
{
|
{
|
||||||
|
imports = [
|
||||||
|
./options.nix
|
||||||
|
./incandescence.nix
|
||||||
|
./simulacrum/test-data.nix
|
||||||
|
];
|
||||||
|
|
||||||
links = {
|
links = {
|
||||||
patroni-pg-internal.ipv4 = "0.0.0.0";
|
patroni-pg-internal.ipv4 = "0.0.0.0";
|
||||||
patroni-api.ipv4 = "0.0.0.0";
|
patroni-api.ipv4 = "0.0.0.0";
|
||||||
|
@ -15,6 +21,7 @@
|
||||||
worker = [
|
worker = [
|
||||||
./worker.nix
|
./worker.nix
|
||||||
./metrics.nix
|
./metrics.nix
|
||||||
|
./create-databases.nix
|
||||||
];
|
];
|
||||||
haproxy = ./haproxy.nix;
|
haproxy = ./haproxy.nix;
|
||||||
};
|
};
|
||||||
|
@ -30,5 +37,10 @@
|
||||||
PATRONI_REWIND_PASSWORD = default;
|
PATRONI_REWIND_PASSWORD = default;
|
||||||
metricsCredentials.nodes = nodes.worker;
|
metricsCredentials.nodes = nodes.worker;
|
||||||
};
|
};
|
||||||
|
simulacrum = {
|
||||||
|
enable = true;
|
||||||
|
deps = [ "consul" "incandescence" "locksmith" ];
|
||||||
|
settings = ./simulacrum/test.nix;
|
||||||
|
};
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
10
cluster/services/patroni/incandescence.nix
Normal file
10
cluster/services/patroni/incandescence.nix
Normal file
|
@ -0,0 +1,10 @@
|
||||||
|
{ config, lib, ... }:
|
||||||
|
|
||||||
|
{
|
||||||
|
incandescence.providers.patroni = {
|
||||||
|
objects = {
|
||||||
|
user = lib.attrNames config.patroni.users;
|
||||||
|
database = lib.attrNames config.patroni.databases;
|
||||||
|
};
|
||||||
|
};
|
||||||
|
}
|
39
cluster/services/patroni/options.nix
Normal file
39
cluster/services/patroni/options.nix
Normal file
|
@ -0,0 +1,39 @@
|
||||||
|
{ lib, ... }:
|
||||||
|
|
||||||
|
let
|
||||||
|
inherit (lib) mkOption;
|
||||||
|
inherit (lib.types) attrsOf enum listOf submodule str;
|
||||||
|
in
|
||||||
|
|
||||||
|
{
|
||||||
|
options.patroni = {
|
||||||
|
databases = mkOption {
|
||||||
|
type = attrsOf (submodule ({ name, ... }: {
|
||||||
|
options = {
|
||||||
|
owner = mkOption {
|
||||||
|
type = str;
|
||||||
|
default = name;
|
||||||
|
};
|
||||||
|
};
|
||||||
|
}));
|
||||||
|
default = {};
|
||||||
|
};
|
||||||
|
users = mkOption {
|
||||||
|
type = attrsOf (submodule ({ ... }: {
|
||||||
|
options = {
|
||||||
|
locksmith = {
|
||||||
|
nodes = mkOption {
|
||||||
|
type = listOf str;
|
||||||
|
default = [];
|
||||||
|
};
|
||||||
|
format = mkOption {
|
||||||
|
type = enum [ "pgpass" "envFile" "raw" ];
|
||||||
|
default = "pgpass";
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
}));
|
||||||
|
default = {};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
}
|
14
cluster/services/patroni/simulacrum/test-data.nix
Normal file
14
cluster/services/patroni/simulacrum/test-data.nix
Normal file
|
@ -0,0 +1,14 @@
|
||||||
|
{ config, lib, ... }:
|
||||||
|
{
|
||||||
|
patroni = lib.mkIf config.simulacrum {
|
||||||
|
databases = config.lib.forService "patroni" {
|
||||||
|
testdb.owner = "testuser";
|
||||||
|
};
|
||||||
|
users = config.lib.forService "patroni" {
|
||||||
|
testuser.locksmith = {
|
||||||
|
nodes = config.services.patroni.nodes.haproxy;
|
||||||
|
format = "pgpass";
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
}
|
91
cluster/services/patroni/simulacrum/test.nix
Normal file
91
cluster/services/patroni/simulacrum/test.nix
Normal file
|
@ -0,0 +1,91 @@
|
||||||
|
{ cluster, ... }:
|
||||||
|
|
||||||
|
let
|
||||||
|
clusterName = "poseidon";
|
||||||
|
link = cluster.config.links.patroni-pg-access;
|
||||||
|
in
|
||||||
|
{
|
||||||
|
defaults = { depot, pkgs, ... }: {
|
||||||
|
environment.systemPackages = [
|
||||||
|
pkgs.jq
|
||||||
|
depot.packages.postgresql
|
||||||
|
];
|
||||||
|
services.patroni.settings.postgresql.pg_hba = [
|
||||||
|
"host postgres postgres 0.0.0.0/0 trust"
|
||||||
|
];
|
||||||
|
};
|
||||||
|
|
||||||
|
# taken from https://github.com/phfroidmont/nixpkgs/blob/patroni-module/nixos/tests/patroni.nix
|
||||||
|
testScript = ''
|
||||||
|
import json
|
||||||
|
nodeNames = json.loads('${builtins.toJSON cluster.config.services.patroni.nodes.worker}')
|
||||||
|
clientNames = json.loads('${builtins.toJSON cluster.config.services.patroni.nodes.haproxy}')
|
||||||
|
nodes = [ n for n in machines if n.name in nodeNames ]
|
||||||
|
clients = [ n for n in machines if n.name in clientNames ]
|
||||||
|
|
||||||
|
def booted(nodes):
|
||||||
|
return filter(lambda node: node.booted, nodes)
|
||||||
|
|
||||||
|
def wait_for_all_nodes_ready(expected_replicas=2):
|
||||||
|
booted_nodes = booted(nodes)
|
||||||
|
for node in booted_nodes:
|
||||||
|
node.wait_for_unit("patroni.service")
|
||||||
|
print(node.succeed("patronictl list ${clusterName}"))
|
||||||
|
node.wait_until_succeeds(f"[ $(patronictl list -f json ${clusterName} | jq 'length') == {expected_replicas + 1} ]")
|
||||||
|
node.wait_until_succeeds("[ $(patronictl list -f json ${clusterName} | jq 'map(select(.Role | test(\"^Leader$\"))) | map(select(.State | test(\"^running$\"))) | length') == 1 ]")
|
||||||
|
node.wait_until_succeeds(f"[ $(patronictl list -f json ${clusterName} | jq 'map(select(.Role | test(\"^Replica$\"))) | map(select(.State | test(\"^streaming$\"))) | length') == {expected_replicas} ]")
|
||||||
|
print(node.succeed("patronictl list ${clusterName}"))
|
||||||
|
for client in booted(clients):
|
||||||
|
client.wait_until_succeeds("psql -h ${link.ipv4} -p ${link.portStr} -U postgres --command='select 1;'")
|
||||||
|
|
||||||
|
def run_dummy_queries():
|
||||||
|
for client in booted(clients):
|
||||||
|
client.succeed("psql -h ${link.ipv4} -p ${link.portStr} -U postgres --pset='pager=off' --tuples-only --command='insert into dummy(val) values (101);'")
|
||||||
|
client.succeed("test $(psql -h ${link.ipv4} -p ${link.portStr} -U postgres --pset='pager=off' --tuples-only --command='select val from dummy where val = 101;') -eq 101")
|
||||||
|
client.succeed("psql -h ${link.ipv4} -p ${link.portStr} -U postgres --pset='pager=off' --tuples-only --command='delete from dummy where val = 101;'")
|
||||||
|
|
||||||
|
start_all()
|
||||||
|
|
||||||
|
with subtest("should bootstrap a new patroni cluster"):
|
||||||
|
wait_for_all_nodes_ready()
|
||||||
|
|
||||||
|
with subtest("should be able to insert and select"):
|
||||||
|
booted_clients = list(booted(clients))
|
||||||
|
booted_clients[0].succeed("psql -h ${link.ipv4} -p ${link.portStr} -U postgres --command='create table dummy as select * from generate_series(1, 100) as val;'")
|
||||||
|
for client in booted_clients:
|
||||||
|
client.succeed("test $(psql -h ${link.ipv4} -p ${link.portStr} -U postgres --pset='pager=off' --tuples-only --command='select count(distinct val) from dummy;') -eq 100")
|
||||||
|
|
||||||
|
with subtest("should restart after all nodes are crashed"):
|
||||||
|
for node in nodes:
|
||||||
|
node.crash()
|
||||||
|
for node in nodes:
|
||||||
|
node.start()
|
||||||
|
wait_for_all_nodes_ready()
|
||||||
|
|
||||||
|
with subtest("should be able to run queries while any one node is crashed"):
|
||||||
|
masterNodeName = nodes[0].succeed("patronictl list -f json ${clusterName} | jq '.[] | select(.Role | test(\"^Leader$\")) | .Member' -r").strip()
|
||||||
|
masterNodeIndex = next((i for i, v in enumerate(nodes) if v.name == masterNodeName))
|
||||||
|
|
||||||
|
# Move master node at the end of the list to avoid multiple failovers (makes the test faster and more consistent)
|
||||||
|
nodes.append(nodes.pop(masterNodeIndex))
|
||||||
|
|
||||||
|
for node in nodes:
|
||||||
|
node.crash()
|
||||||
|
wait_for_all_nodes_ready(1)
|
||||||
|
|
||||||
|
# Execute some queries while a node is down.
|
||||||
|
run_dummy_queries()
|
||||||
|
|
||||||
|
# Restart crashed node.
|
||||||
|
node.start()
|
||||||
|
wait_for_all_nodes_ready()
|
||||||
|
|
||||||
|
# Execute some queries with the node back up.
|
||||||
|
run_dummy_queries()
|
||||||
|
|
||||||
|
with subtest("should create databases and users via incandescence"):
|
||||||
|
for client in clients:
|
||||||
|
client.succeed(f"PGPASSFILE=/run/locksmith/patroni-testuser psql -h ${link.ipv4} -p ${link.portStr} -U testuser -d testdb --command='create table test_table_{client.name} as select * from generate_series(1, 10) as val;'")
|
||||||
|
client.fail("PGPASSFILE=/run/locksmith/patroni-testuser psql -h ${link.ipv4} -p ${link.portStr} -U testuser -d postgres --command='select * from dummy;'")
|
||||||
|
'';
|
||||||
|
}
|
|
@ -25,6 +25,10 @@ in
|
||||||
"d '${baseDir}' 0700 patroni patroni - -"
|
"d '${baseDir}' 0700 patroni patroni - -"
|
||||||
"d '${walDir}' 0700 patroni patroni - -"
|
"d '${walDir}' 0700 patroni patroni - -"
|
||||||
];
|
];
|
||||||
|
systemd.services.patroni = {
|
||||||
|
requires = [ "consul-ready.service" ];
|
||||||
|
after = [ "consul-ready.service" ];
|
||||||
|
};
|
||||||
services.patroni = {
|
services.patroni = {
|
||||||
enable = true;
|
enable = true;
|
||||||
name = hostName;
|
name = hostName;
|
||||||
|
@ -57,6 +61,7 @@ in
|
||||||
};
|
};
|
||||||
use_pg_rewind = true;
|
use_pg_rewind = true;
|
||||||
use_slots = true;
|
use_slots = true;
|
||||||
|
synchronous_mode = true;
|
||||||
authentication = {
|
authentication = {
|
||||||
replication.username = "patronirep";
|
replication.username = "patronirep";
|
||||||
rewind.username = "patronirew";
|
rewind.username = "patronirew";
|
||||||
|
@ -67,6 +72,7 @@ in
|
||||||
wal_level = "replica";
|
wal_level = "replica";
|
||||||
hot_standby_feedback = "on";
|
hot_standby_feedback = "on";
|
||||||
unix_socket_directories = "/tmp";
|
unix_socket_directories = "/tmp";
|
||||||
|
synchronous_commit = "on";
|
||||||
};
|
};
|
||||||
pg_hba = [
|
pg_hba = [
|
||||||
"host replication patronirep ${net} scram-sha-256"
|
"host replication patronirep ${net} scram-sha-256"
|
||||||
|
|
|
@ -7,6 +7,8 @@ in
|
||||||
{
|
{
|
||||||
imports = [
|
imports = [
|
||||||
./options.nix
|
./options.nix
|
||||||
|
./incandescence.nix
|
||||||
|
./simulacrum/test-data.nix
|
||||||
];
|
];
|
||||||
|
|
||||||
services.storage = {
|
services.storage = {
|
||||||
|
@ -30,11 +32,16 @@ in
|
||||||
heresy = [
|
heresy = [
|
||||||
./heresy.nix
|
./heresy.nix
|
||||||
./s3ql-upgrades.nix
|
./s3ql-upgrades.nix
|
||||||
|
] ++ lib.optionals config.simulacrum [
|
||||||
|
./simulacrum/snakeoil-heresy-passphrase.nix
|
||||||
];
|
];
|
||||||
garage = [
|
garage = [
|
||||||
./garage.nix
|
./garage.nix
|
||||||
./garage-options.nix
|
./garage-options.nix
|
||||||
./garage-layout.nix
|
./garage-layout.nix
|
||||||
|
./incandescence-ascensions.nix
|
||||||
|
] ++ lib.optionals config.simulacrum [
|
||||||
|
./simulacrum/snakeoil-rpc-secret.nix
|
||||||
];
|
];
|
||||||
garageConfig = [
|
garageConfig = [
|
||||||
./garage-gateway.nix
|
./garage-gateway.nix
|
||||||
|
@ -48,6 +55,11 @@ in
|
||||||
garageInternal = [ ./garage-internal.nix ];
|
garageInternal = [ ./garage-internal.nix ];
|
||||||
garageExternal = [ ./garage-external.nix ];
|
garageExternal = [ ./garage-external.nix ];
|
||||||
};
|
};
|
||||||
|
simulacrum = {
|
||||||
|
enable = true;
|
||||||
|
deps = [ "consul" "locksmith" "incandescence" "patroni" "ways" ];
|
||||||
|
settings = ./test.nix;
|
||||||
|
};
|
||||||
};
|
};
|
||||||
|
|
||||||
links = {
|
links = {
|
||||||
|
@ -86,7 +98,10 @@ in
|
||||||
};
|
};
|
||||||
|
|
||||||
garage = {
|
garage = {
|
||||||
keys.storage-prophet = {};
|
keys.storage-prophet.locksmith = {
|
||||||
|
nodes = [ "prophet" ];
|
||||||
|
format = "s3ql";
|
||||||
|
};
|
||||||
buckets.storage-prophet = {
|
buckets.storage-prophet = {
|
||||||
allow.storage-prophet = [ "read" "write" ];
|
allow.storage-prophet = [ "read" "write" ];
|
||||||
};
|
};
|
||||||
|
|
|
@ -8,7 +8,7 @@ in
|
||||||
services.external-storage = {
|
services.external-storage = {
|
||||||
fileSystems.external = {
|
fileSystems.external = {
|
||||||
mountpoint = "/srv/storage";
|
mountpoint = "/srv/storage";
|
||||||
authFile = ./secrets/external-storage-auth-${hostName}.age;
|
locksmithSecret = "garage-storage-${hostName}";
|
||||||
backend = "s3c4://${cluster.config.links.garageS3.hostname}/storage-${hostName}";
|
backend = "s3c4://${cluster.config.links.garageS3.hostname}/storage-${hostName}";
|
||||||
backendOptions = [ "disable-expect100" ];
|
backendOptions = [ "disable-expect100" ];
|
||||||
};
|
};
|
||||||
|
|
|
@ -26,62 +26,6 @@ let
|
||||||
sleep 1
|
sleep 1
|
||||||
done
|
done
|
||||||
}
|
}
|
||||||
# FIXME: returns bogus empty string when one of the lists is empty
|
|
||||||
diffAdded() {
|
|
||||||
comm -13 <(printf '%s\n' $1 | sort) <(printf '%s\n' $2 | sort)
|
|
||||||
}
|
|
||||||
diffRemoved() {
|
|
||||||
comm -23 <(printf '%s\n' $1 | sort) <(printf '%s\n' $2 | sort)
|
|
||||||
}
|
|
||||||
# FIXME: this does not handle list items with spaces
|
|
||||||
listKeys() {
|
|
||||||
garage key list | tail -n +2 | grep -ow '[^ ]*$' || true
|
|
||||||
}
|
|
||||||
ensureKeys() {
|
|
||||||
old="$(listKeys)"
|
|
||||||
if [[ -z "$1" ]]; then
|
|
||||||
for key in $old; do
|
|
||||||
garage key delete --yes "$key"
|
|
||||||
done
|
|
||||||
elif [[ -z "$old" ]]; then
|
|
||||||
for key in $1; do
|
|
||||||
# don't print secret key
|
|
||||||
garage key new --name "$key" >/dev/null
|
|
||||||
echo Key "$key" was created.
|
|
||||||
done
|
|
||||||
else
|
|
||||||
diffAdded "$old" "$1" | while read key; do
|
|
||||||
# don't print secret key
|
|
||||||
garage key new --name "$key" >/dev/null
|
|
||||||
echo Key "$key" was created.
|
|
||||||
done
|
|
||||||
diffRemoved "$old" "$1" | while read key; do
|
|
||||||
garage key delete --yes "$key"
|
|
||||||
done
|
|
||||||
fi
|
|
||||||
}
|
|
||||||
listBuckets() {
|
|
||||||
garage bucket list | tail -n +2 | grep -ow '^ *[^ ]*' | tr -d ' ' || true
|
|
||||||
}
|
|
||||||
ensureBuckets() {
|
|
||||||
old="$(listBuckets)"
|
|
||||||
if [[ -z "$1" ]]; then
|
|
||||||
for bucket in $old; do
|
|
||||||
garage bucket delete --yes "$bucket"
|
|
||||||
done
|
|
||||||
elif [[ -z "$old" ]]; then
|
|
||||||
for bucket in $1; do
|
|
||||||
garage bucket create "$bucket"
|
|
||||||
done
|
|
||||||
else
|
|
||||||
diffAdded "$old" "$1" | while read bucket; do
|
|
||||||
garage bucket create "$bucket"
|
|
||||||
done
|
|
||||||
diffRemoved "$old" "$1" | while read bucket; do
|
|
||||||
garage bucket delete --yes "$bucket"
|
|
||||||
done
|
|
||||||
fi
|
|
||||||
}
|
|
||||||
'';
|
'';
|
||||||
in
|
in
|
||||||
|
|
||||||
|
@ -118,7 +62,7 @@ in
|
||||||
};
|
};
|
||||||
format = mkOption {
|
format = mkOption {
|
||||||
description = "Locksmith secret format.";
|
description = "Locksmith secret format.";
|
||||||
type = enum [ "files" "aws" "envFile" ];
|
type = enum [ "files" "aws" "envFile" "s3ql" ];
|
||||||
default = "files";
|
default = "files";
|
||||||
};
|
};
|
||||||
owner = mkOption {
|
owner = mkOption {
|
||||||
|
@ -203,9 +147,7 @@ in
|
||||||
garage layout apply --version 1
|
garage layout apply --version 1
|
||||||
'';
|
'';
|
||||||
};
|
};
|
||||||
garage-apply = {
|
garage-ready = {
|
||||||
distributed.enable = true;
|
|
||||||
wantedBy = [ "garage.service" "multi-user.target" ];
|
|
||||||
wants = [ "garage.service" ];
|
wants = [ "garage.service" ];
|
||||||
after = [ "garage.service" "garage-layout-init.service" ];
|
after = [ "garage.service" "garage-layout-init.service" ];
|
||||||
path = [ config.services.garage.package ];
|
path = [ config.services.garage.package ];
|
||||||
|
@ -219,54 +161,70 @@ in
|
||||||
script = ''
|
script = ''
|
||||||
source ${garageShellLibrary}
|
source ${garageShellLibrary}
|
||||||
waitForGarageOperational
|
waitForGarageOperational
|
||||||
|
|
||||||
ensureKeys '${lib.concatStringsSep " " (lib.attrNames cfg.keys)}'
|
|
||||||
ensureBuckets '${lib.concatStringsSep " " (lib.attrNames cfg.buckets)}'
|
|
||||||
|
|
||||||
# key permissions
|
|
||||||
${lib.pipe cfg.keys [
|
|
||||||
(lib.mapAttrsToList (key: kCfg: ''
|
|
||||||
garage key ${if kCfg.allow.createBucket then "allow" else "deny"} '${key}' --create-bucket >/dev/null
|
|
||||||
''))
|
|
||||||
(lib.concatStringsSep "\n")
|
|
||||||
]}
|
|
||||||
|
|
||||||
# bucket permissions
|
|
||||||
${lib.pipe cfg.buckets [
|
|
||||||
(lib.mapAttrsToList (bucket: bCfg:
|
|
||||||
lib.mapAttrsToList (key: perms: ''
|
|
||||||
garage bucket allow '${bucket}' --key '${key}' ${lib.escapeShellArgs (map (x: "--${x}") perms)}
|
|
||||||
garage bucket deny '${bucket}' --key '${key}' ${lib.escapeShellArgs (map (x: "--${x}") (lib.subtractLists perms [ "read" "write" "owner" ]))}
|
|
||||||
'') bCfg.allow
|
|
||||||
))
|
|
||||||
lib.flatten
|
|
||||||
(lib.concatStringsSep "\n")
|
|
||||||
]}
|
|
||||||
|
|
||||||
# bucket quotas
|
|
||||||
${lib.pipe cfg.buckets [
|
|
||||||
(lib.mapAttrsToList (bucket: bCfg: ''
|
|
||||||
garage bucket set-quotas '${bucket}' \
|
|
||||||
--max-objects '${if bCfg.quotas.maxObjects == null then "none" else toString bCfg.quotas.maxObjects}' \
|
|
||||||
--max-size '${if bCfg.quotas.maxSize == null then "none" else toString bCfg.quotas.maxSize}'
|
|
||||||
''))
|
|
||||||
(lib.concatStringsSep "\n")
|
|
||||||
]}
|
|
||||||
|
|
||||||
# bucket website access
|
|
||||||
${lib.pipe cfg.buckets [
|
|
||||||
(lib.mapAttrsToList (bucket: bCfg: ''
|
|
||||||
garage bucket website ${if bCfg.web.enable then "--allow" else "--deny"} '${bucket}'
|
|
||||||
''))
|
|
||||||
(lib.concatStringsSep "\n")
|
|
||||||
]}
|
|
||||||
'';
|
'';
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
|
|
||||||
|
services.incandescence.providers.garage = {
|
||||||
|
locksmith = true;
|
||||||
|
wantedBy = [ "garage.service" "multi-user.target" ];
|
||||||
|
partOf = [ "garage.service" ];
|
||||||
|
wants = [ "garage-ready.service" ];
|
||||||
|
after = [ "garage-ready.service" ];
|
||||||
|
|
||||||
|
packages = [
|
||||||
|
config.services.garage.package
|
||||||
|
];
|
||||||
|
formulae = {
|
||||||
|
key = {
|
||||||
|
destroyAfterDays = 0;
|
||||||
|
create = key: ''
|
||||||
|
# don't print secret key
|
||||||
|
garage key new --name ${lib.escapeShellArg key} >/dev/null
|
||||||
|
echo Key ${lib.escapeShellArg key} was created.
|
||||||
|
'';
|
||||||
|
destroy = ''
|
||||||
|
garage key delete --yes "$OBJECT"
|
||||||
|
'';
|
||||||
|
change = key: let
|
||||||
|
kCfg = cfg.keys.${key};
|
||||||
|
in ''
|
||||||
|
garage key ${if kCfg.allow.createBucket then "allow" else "deny"} ${lib.escapeShellArg key} --create-bucket >/dev/null
|
||||||
|
'';
|
||||||
|
};
|
||||||
|
bucket = {
|
||||||
|
deps = [ "key" ];
|
||||||
|
destroyAfterDays = 30;
|
||||||
|
create = bucket: ''
|
||||||
|
garage bucket create ${lib.escapeShellArg bucket}
|
||||||
|
'';
|
||||||
|
destroy = ''
|
||||||
|
garage bucket delete --yes "$OBJECT"
|
||||||
|
'';
|
||||||
|
change = bucket: let
|
||||||
|
bCfg = cfg.buckets.${bucket};
|
||||||
|
in ''
|
||||||
|
# permissions
|
||||||
|
${lib.concatStringsSep "\n" (lib.flatten (
|
||||||
|
lib.mapAttrsToList (key: perms: ''
|
||||||
|
garage bucket allow ${lib.escapeShellArg bucket} --key ${lib.escapeShellArg key} ${lib.escapeShellArgs (map (x: "--${x}") perms)}
|
||||||
|
garage bucket deny ${lib.escapeShellArg bucket} --key ${lib.escapeShellArg key} ${lib.escapeShellArgs (map (x: "--${x}") (lib.subtractLists perms [ "read" "write" "owner" ]))}
|
||||||
|
'') bCfg.allow
|
||||||
|
))}
|
||||||
|
|
||||||
|
# quotas
|
||||||
|
garage bucket set-quotas ${lib.escapeShellArg bucket} \
|
||||||
|
--max-objects '${if bCfg.quotas.maxObjects == null then "none" else toString bCfg.quotas.maxObjects}' \
|
||||||
|
--max-size '${if bCfg.quotas.maxSize == null then "none" else toString bCfg.quotas.maxSize}'
|
||||||
|
|
||||||
|
# website access
|
||||||
|
garage bucket website ${if bCfg.web.enable then "--allow" else "--deny"} ${lib.escapeShellArg bucket}
|
||||||
|
'';
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
|
||||||
services.locksmith.providers.garage = {
|
services.locksmith.providers.garage = {
|
||||||
wantedBy = [ "garage-apply.service" ];
|
|
||||||
after = [ "garage-apply.service" ];
|
|
||||||
secrets = lib.mkMerge (lib.mapAttrsToList (key: kCfg: let
|
secrets = lib.mkMerge (lib.mapAttrsToList (key: kCfg: let
|
||||||
common = {
|
common = {
|
||||||
inherit (kCfg.locksmith) mode owner group nodes;
|
inherit (kCfg.locksmith) mode owner group nodes;
|
||||||
|
@ -291,6 +249,12 @@ in
|
||||||
AWS_ACCESS_KEY_ID=@@GARAGE_KEY_ID@@
|
AWS_ACCESS_KEY_ID=@@GARAGE_KEY_ID@@
|
||||||
AWS_SECRET_ACCESS_KEY=@@GARAGE_SECRET_KEY@@
|
AWS_SECRET_ACCESS_KEY=@@GARAGE_SECRET_KEY@@
|
||||||
'';
|
'';
|
||||||
|
s3ql = ''
|
||||||
|
[s3c]
|
||||||
|
storage-url: s3c4://
|
||||||
|
backend-login: @@GARAGE_KEY_ID@@
|
||||||
|
backend-password: @@GARAGE_SECRET_KEY@@
|
||||||
|
'';
|
||||||
}.${kCfg.locksmith.format};
|
}.${kCfg.locksmith.format};
|
||||||
in {
|
in {
|
||||||
${key} = common // {
|
${key} = common // {
|
||||||
|
|
|
@ -63,6 +63,8 @@ in
|
||||||
};
|
};
|
||||||
|
|
||||||
systemd.services.garage = {
|
systemd.services.garage = {
|
||||||
|
requires = [ "consul-ready.service" ];
|
||||||
|
after = [ "consul-ready.service" ];
|
||||||
unitConfig = {
|
unitConfig = {
|
||||||
RequiresMountsFor = [ cfg.settings.data_dir ];
|
RequiresMountsFor = [ cfg.settings.data_dir ];
|
||||||
};
|
};
|
||||||
|
|
|
@ -11,6 +11,7 @@
|
||||||
unitDescription = "Heresy Filesystem";
|
unitDescription = "Heresy Filesystem";
|
||||||
authFile = ./secrets/heresy-encryption-key.age;
|
authFile = ./secrets/heresy-encryption-key.age;
|
||||||
underlay = "heresy";
|
underlay = "heresy";
|
||||||
|
encrypt = true;
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
18
cluster/services/storage/incandescence-ascensions.nix
Normal file
18
cluster/services/storage/incandescence-ascensions.nix
Normal file
|
@ -0,0 +1,18 @@
|
||||||
|
{ config, lib, ... }:
|
||||||
|
|
||||||
|
{
|
||||||
|
system.ascensions = {
|
||||||
|
incandescence-garage = lib.mkIf (config.services.incandescence.providers ? garage) {
|
||||||
|
incantations = i: [
|
||||||
|
(i.runGarage /*bash*/ ''
|
||||||
|
garage bucket list | tail -n +2 | cut -d' ' -f3 | while read bucket; do
|
||||||
|
${i.runConsul /*bash*/ ''consul kv put "services/incandescence/providers/garage/formulae/bucket/$1/alive" true''} "$bucket"
|
||||||
|
done
|
||||||
|
garage key list | tail -n +2 | cut -d' ' -f5 | while read key; do
|
||||||
|
${i.runConsul /*bash*/ ''consul kv put "services/incandescence/providers/garage/formulae/key/$1/alive" true''} "$key"
|
||||||
|
done
|
||||||
|
'')
|
||||||
|
];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
}
|
10
cluster/services/storage/incandescence.nix
Normal file
10
cluster/services/storage/incandescence.nix
Normal file
|
@ -0,0 +1,10 @@
|
||||||
|
{ config, lib, ... }:
|
||||||
|
|
||||||
|
{
|
||||||
|
incandescence.providers.garage = {
|
||||||
|
objects = {
|
||||||
|
key = lib.attrNames config.garage.keys;
|
||||||
|
bucket = lib.attrNames config.garage.buckets;
|
||||||
|
};
|
||||||
|
};
|
||||||
|
}
|
|
@ -1,11 +0,0 @@
|
||||||
age-encryption.org/v1
|
|
||||||
-> ssh-ed25519 NO562A tC8lfwNJIXjVJImBq25v/NGIQ1Ns24NpCzksbw/eb3w
|
|
||||||
2hQltUYSO2Gpjd+49IQR1UJOhy33xWvNH6dx+uGDvFA
|
|
||||||
-> ssh-ed25519 5/zT0w dapxQ/VV0peQKMwghQJ91wQVahYOqxw2QrXqQCau82c
|
|
||||||
0DnIF5ISoB5htYA3X5DSTgLJXLSkqjz1O0CMcmnnrjQ
|
|
||||||
-> ssh-ed25519 YIaSKQ ehv+WWCLC/co9lhpa+cAdqJUG33L/Vkn6lUXOwNRV2w
|
|
||||||
LEobbvvpq6lPNbzasGeXf9NabN150ZVe5n5OJNgbyD4
|
|
||||||
--- FrT2CFmuWQ+vKGbBY2pGT90Mu8WzXfpbIAzYdR3Vb2w
|
|
||||||
™ªg¬NÑ8´¨\K!p
«ï…7ù¶›käõ¯#ŒÏu›µ*{}Tþ0·|@ÉÿàE>z„'-RxK¸zB£ÿä©n*0¢÷~OVû®4¦qûÁ]^(ìì>-‡3ÌÙe0aí<61>¥ì.oòÙC)†‡4g¶ð»7NzÉ”ºnÒÃî®Mª†x6àöãö×'[Ô6ãw?ÿª€ãi=†vèEJˆB
|
|
||||||
µÿÂ9gÏi"Q
–ÿ
|
|
||||||
™›Ù®à
|
|
|
@ -0,0 +1,8 @@
|
||||||
|
{
|
||||||
|
environment.etc."dummy-secrets/storageAuth-heresy".text = ''
|
||||||
|
[local]
|
||||||
|
storage-url: local://
|
||||||
|
fs-passphrase: simulacrum
|
||||||
|
'';
|
||||||
|
}
|
||||||
|
|
|
@ -0,0 +1,3 @@
|
||||||
|
{
|
||||||
|
environment.etc."dummy-secrets/garageRpcSecret".text = "aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa";
|
||||||
|
}
|
8
cluster/services/storage/simulacrum/test-data.nix
Normal file
8
cluster/services/storage/simulacrum/test-data.nix
Normal file
|
@ -0,0 +1,8 @@
|
||||||
|
{ config, lib, ... }:
|
||||||
|
|
||||||
|
{
|
||||||
|
garage = lib.mkIf config.simulacrum {
|
||||||
|
keys.testkey = {};
|
||||||
|
buckets.testbucket.allow.testKey = [ "read" "write" ];
|
||||||
|
};
|
||||||
|
}
|
105
cluster/services/storage/test.nix
Normal file
105
cluster/services/storage/test.nix
Normal file
|
@ -0,0 +1,105 @@
|
||||||
|
{ cluster, lib, ... }:
|
||||||
|
|
||||||
|
let
|
||||||
|
inherit (cluster.config.services.storage) nodes;
|
||||||
|
|
||||||
|
firstGarageNode = lib.elemAt nodes.garage 0;
|
||||||
|
in
|
||||||
|
|
||||||
|
{
|
||||||
|
nodes = lib.genAttrs nodes.garage (node: {
|
||||||
|
services.garage = {
|
||||||
|
layout.initial = lib.genAttrs nodes.garage (_: {
|
||||||
|
capacity = lib.mkOverride 51 1000;
|
||||||
|
});
|
||||||
|
};
|
||||||
|
specialisation.modifiedLayout = {
|
||||||
|
inheritParentConfig = true;
|
||||||
|
configuration = {
|
||||||
|
services.garage = {
|
||||||
|
layout.initial.${firstGarageNode}.capacity = lib.mkForce 2000;
|
||||||
|
};
|
||||||
|
system.ascensions.garage-layout.incantations = lib.mkForce (i: [
|
||||||
|
(i.runGarage ''
|
||||||
|
garage layout assign -z eu-central -c 2000 "$(garage node id -q | cut -d@ -f1)"
|
||||||
|
garage layout apply --version 2
|
||||||
|
'')
|
||||||
|
]);
|
||||||
|
};
|
||||||
|
};
|
||||||
|
});
|
||||||
|
|
||||||
|
testScript = ''
|
||||||
|
import json
|
||||||
|
nodes = [n for n in machines if n.name in json.loads('${builtins.toJSON nodes.garage}')]
|
||||||
|
garage1 = nodes[0]
|
||||||
|
|
||||||
|
start_all()
|
||||||
|
|
||||||
|
with subtest("should bootstrap new cluster"):
|
||||||
|
for node in nodes:
|
||||||
|
node.wait_for_unit("garage.service")
|
||||||
|
|
||||||
|
for node in nodes:
|
||||||
|
node.wait_until_fails("garage status | grep 'NO ROLE ASSIGNED'")
|
||||||
|
|
||||||
|
with subtest("should apply new layout with ascension"):
|
||||||
|
for node in nodes:
|
||||||
|
node.wait_until_succeeds('test "$(systemctl list-jobs | wc -l)" -eq 1')
|
||||||
|
|
||||||
|
for node in nodes:
|
||||||
|
node.succeed("/run/current-system/specialisation/modifiedLayout/bin/switch-to-configuration test")
|
||||||
|
|
||||||
|
for node in nodes:
|
||||||
|
node.wait_until_succeeds("garage layout show | grep -w 2000")
|
||||||
|
assert "1" in node.succeed("garage layout show | grep -w 2000 | wc -l")
|
||||||
|
assert "2" in node.succeed("garage layout show | grep -w 1000 | wc -l")
|
||||||
|
|
||||||
|
consulConfig = json.loads(garage1.succeed("cat /etc/consul.json"))
|
||||||
|
addr = consulConfig["addresses"]["http"]
|
||||||
|
port = consulConfig["ports"]["http"]
|
||||||
|
setEnv = f"CONSUL_HTTP_ADDR={addr}:{port}"
|
||||||
|
with subtest("should apply new layout from scratch"):
|
||||||
|
for node in nodes:
|
||||||
|
node.systemctl("stop garage.service")
|
||||||
|
node.succeed("rm -rf /var/lib/garage-metadata")
|
||||||
|
garage1.succeed(f"{setEnv} consul kv delete --recurse services/incandescence/providers/garage")
|
||||||
|
|
||||||
|
for node in nodes:
|
||||||
|
node.systemctl("start garage.service")
|
||||||
|
|
||||||
|
for node in nodes:
|
||||||
|
node.wait_for_unit("garage.service")
|
||||||
|
|
||||||
|
for node in nodes:
|
||||||
|
node.wait_until_fails("garage status | grep 'NO ROLE ASSIGNED'")
|
||||||
|
|
||||||
|
for node in nodes:
|
||||||
|
node.wait_until_succeeds("garage layout show | grep -w 2000")
|
||||||
|
assert "1" in node.succeed("garage layout show | grep -w 2000 | wc -l")
|
||||||
|
assert "${toString ((lib.length nodes.garage) - 1)}" in node.succeed("garage layout show | grep -w 1000 | wc -l")
|
||||||
|
|
||||||
|
with subtest("should create specified buckets and keys"):
|
||||||
|
for node in nodes:
|
||||||
|
node.wait_for_unit("incandescence-garage.target")
|
||||||
|
garage1.succeed("garage key list | grep testkey")
|
||||||
|
garage1.succeed("garage bucket list | grep testbucket")
|
||||||
|
|
||||||
|
with subtest("should delete unspecified keys"):
|
||||||
|
garage1.succeed("garage bucket create unwantedbucket")
|
||||||
|
garage1.succeed("garage key new --name unwantedkey")
|
||||||
|
garage1.succeed(f"{setEnv} consul kv put services/incandescence/providers/garage/formulae/key/unwantedkey/alive true")
|
||||||
|
garage1.succeed(f"{setEnv} consul kv put services/incandescence/providers/garage/formulae/bucket/unwantedbucket/alive true")
|
||||||
|
garage1.succeed("systemctl restart garage.service")
|
||||||
|
garage1.wait_for_unit("incandescence-garage.target")
|
||||||
|
garage1.fail("garage key list | grep unwantedkey")
|
||||||
|
garage1.succeed("garage bucket list | grep unwantedbucket")
|
||||||
|
|
||||||
|
with subtest("should delete unspecified buckets after grace period"):
|
||||||
|
garage1.succeed(f"{setEnv} consul kv put services/incandescence/providers/garage/formulae/bucket/unwantedbucket/destroyOn 1")
|
||||||
|
garage1.succeed("systemctl restart garage.service")
|
||||||
|
garage1.wait_for_unit("incandescence-garage.target")
|
||||||
|
garage1.fail("garage bucket list | grep unwantedbucket")
|
||||||
|
'';
|
||||||
|
}
|
||||||
|
|
|
@ -8,6 +8,7 @@
|
||||||
services.ways = {
|
services.ways = {
|
||||||
nodes.host = config.services.websites.nodes.host;
|
nodes.host = config.services.websites.nodes.host;
|
||||||
nixos.host = ./host.nix;
|
nixos.host = ./host.nix;
|
||||||
|
simulacrum.deps = [ "nginx" "acme-client" "dns" "certificates" "consul" ];
|
||||||
};
|
};
|
||||||
|
|
||||||
dns.records = lib.mapAttrs'
|
dns.records = lib.mapAttrs'
|
||||||
|
|
|
@ -8,6 +8,8 @@ let
|
||||||
cfgAge = config.age;
|
cfgAge = config.age;
|
||||||
|
|
||||||
create = lib.flip lib.mapAttrs';
|
create = lib.flip lib.mapAttrs';
|
||||||
|
|
||||||
|
createFiltered = pred: attrs: f: create (lib.filterAttrs pred attrs) f;
|
||||||
in
|
in
|
||||||
|
|
||||||
{
|
{
|
||||||
|
@ -20,12 +22,17 @@ in
|
||||||
fileSystems = lib.mkOption {
|
fileSystems = lib.mkOption {
|
||||||
description = "S3QL-based filesystems on top of CIFS mountpoints.";
|
description = "S3QL-based filesystems on top of CIFS mountpoints.";
|
||||||
default = {};
|
default = {};
|
||||||
type = with lib.types; lazyAttrsOf (submodule ({ config, name, ... }: {
|
type = with lib.types; lazyAttrsOf (submodule ({ config, name, ... }: let
|
||||||
|
authFile = if config.locksmithSecret != null then
|
||||||
|
"/run/locksmith/${config.locksmithSecret}"
|
||||||
|
else
|
||||||
|
cfgAge.secrets."storageAuth-${name}".path;
|
||||||
|
in {
|
||||||
imports = [ ./filesystem-type.nix ];
|
imports = [ ./filesystem-type.nix ];
|
||||||
backend = lib.mkIf (config.underlay != null) "local://${cfg.underlays.${config.underlay}.mountpoint}";
|
backend = lib.mkIf (config.underlay != null) "local://${cfg.underlays.${config.underlay}.mountpoint}";
|
||||||
commonArgs = [
|
commonArgs = [
|
||||||
"--cachedir" config.cacheDir
|
"--cachedir" config.cacheDir
|
||||||
"--authfile" cfgAge.secrets."storageAuth-${name}".path
|
"--authfile" authFile
|
||||||
] ++ (lib.optionals (config.backendOptions != []) [ "--backend-options" (lib.concatStringsSep "," config.backendOptions) ]);
|
] ++ (lib.optionals (config.backendOptions != []) [ "--backend-options" (lib.concatStringsSep "," config.backendOptions) ]);
|
||||||
}));
|
}));
|
||||||
};
|
};
|
||||||
|
@ -57,9 +64,14 @@ in
|
||||||
|
|
||||||
age.secrets = lib.mkMerge [
|
age.secrets = lib.mkMerge [
|
||||||
(create cfg.underlays (name: ul: lib.nameValuePair "cifsCredentials-${name}" { file = ul.credentialsFile; }))
|
(create cfg.underlays (name: ul: lib.nameValuePair "cifsCredentials-${name}" { file = ul.credentialsFile; }))
|
||||||
(create cfg.fileSystems (name: fs: lib.nameValuePair "storageAuth-${name}" { file = fs.authFile; }))
|
(createFiltered (_: fs: fs.locksmithSecret == null) cfg.fileSystems (name: fs: lib.nameValuePair "storageAuth-${name}" { file = fs.authFile; }))
|
||||||
];
|
];
|
||||||
|
|
||||||
|
services.locksmith.waitForSecrets = createFiltered (_: fs: fs.locksmithSecret != null) cfg.fileSystems (name: fs: {
|
||||||
|
name = fs.unitName;
|
||||||
|
value = [ fs.locksmithSecret ];
|
||||||
|
});
|
||||||
|
|
||||||
fileSystems = create cfg.underlays (name: ul: {
|
fileSystems = create cfg.underlays (name: ul: {
|
||||||
name = ul.mountpoint;
|
name = ul.mountpoint;
|
||||||
value = {
|
value = {
|
||||||
|
@ -97,7 +109,13 @@ in
|
||||||
value = let
|
value = let
|
||||||
isUnderlay = fs.underlay != null;
|
isUnderlay = fs.underlay != null;
|
||||||
|
|
||||||
fsType = if isUnderlay then "local" else lib.head (lib.strings.match "([a-z0-9]*)://.*" fs.backend);
|
backendParts = lib.strings.match "([a-z0-9]*)://([^/]*)/([^/]*)(/.*)?" fs.backend;
|
||||||
|
|
||||||
|
fsType = if isUnderlay then "local" else lib.head backendParts;
|
||||||
|
|
||||||
|
s3Endpoint = assert fsType == "s3c4"; lib.elemAt backendParts 1;
|
||||||
|
|
||||||
|
s3Bucket = assert fsType == "s3c4"; lib.elemAt backendParts 2;
|
||||||
|
|
||||||
localBackendPath = if isUnderlay then cfg.underlays.${fs.underlay}.mountpoint else lib.head (lib.strings.match "[a-z0-9]*://(/.*)" fs.backend);
|
localBackendPath = if isUnderlay then cfg.underlays.${fs.underlay}.mountpoint else lib.head (lib.strings.match "[a-z0-9]*://(/.*)" fs.backend);
|
||||||
in {
|
in {
|
||||||
|
@ -120,8 +138,12 @@ in
|
||||||
ExecStartPre = map lib.escapeShellArgs [
|
ExecStartPre = map lib.escapeShellArgs [
|
||||||
[
|
[
|
||||||
(let
|
(let
|
||||||
|
authFile = if fs.locksmithSecret != null then
|
||||||
|
"/run/locksmith/${fs.locksmithSecret}"
|
||||||
|
else
|
||||||
|
cfgAge.secrets."storageAuth-${name}".path;
|
||||||
mkfsEncrypted = ''
|
mkfsEncrypted = ''
|
||||||
${pkgs.gnugrep}/bin/grep -m1 fs-passphrase: '${config.age.secrets."storageAuth-${name}".path}' \
|
${pkgs.gnugrep}/bin/grep -m1 fs-passphrase: '${authFile}' \
|
||||||
| cut -d' ' -f2- \
|
| cut -d' ' -f2- \
|
||||||
| ${s3ql}/bin/mkfs.s3ql ${lib.escapeShellArgs fs.commonArgs} -L '${name}' '${fs.backend}'
|
| ${s3ql}/bin/mkfs.s3ql ${lib.escapeShellArgs fs.commonArgs} -L '${name}' '${fs.backend}'
|
||||||
'';
|
'';
|
||||||
|
@ -132,6 +154,11 @@ in
|
||||||
|
|
||||||
detectFs = {
|
detectFs = {
|
||||||
local = "test -e ${localBackendPath}/s3ql_metadata";
|
local = "test -e ${localBackendPath}/s3ql_metadata";
|
||||||
|
s3c4 = pkgs.writeShellScript "detect-s3ql-filesystem" ''
|
||||||
|
export AWS_ACCESS_KEY_ID="$(${pkgs.gnugrep}/bin/grep -m1 backend-login: '${authFile}' | cut -d' ' -f2-)"
|
||||||
|
export AWS_SECRET_ACCESS_KEY="$(${pkgs.gnugrep}/bin/grep -m1 backend-password: '${authFile}' | cut -d' ' -f2-)"
|
||||||
|
${pkgs.s5cmd}/bin/s5cmd --endpoint-url https://${s3Endpoint}/ ls 's3://${s3Bucket}/s3ql_params' >/dev/null
|
||||||
|
'';
|
||||||
}.${fsType} or null;
|
}.${fsType} or null;
|
||||||
in pkgs.writeShellScript "create-s3ql-filesystem" (lib.optionalString (detectFs != null) ''
|
in pkgs.writeShellScript "create-s3ql-filesystem" (lib.optionalString (detectFs != null) ''
|
||||||
if ! ${detectFs}; then
|
if ! ${detectFs}; then
|
||||||
|
|
|
@ -22,6 +22,10 @@ with lib;
|
||||||
authFile = mkOption {
|
authFile = mkOption {
|
||||||
type = types.path;
|
type = types.path;
|
||||||
};
|
};
|
||||||
|
locksmithSecret = mkOption {
|
||||||
|
type = with types; nullOr str;
|
||||||
|
default = null;
|
||||||
|
};
|
||||||
cacheDir = mkOption {
|
cacheDir = mkOption {
|
||||||
type = types.path;
|
type = types.path;
|
||||||
default = "/var/cache/remote-storage/${name}";
|
default = "/var/cache/remote-storage/${name}";
|
||||||
|
|
|
@ -15,12 +15,6 @@ in
|
||||||
inherit (config) cluster;
|
inherit (config) cluster;
|
||||||
};
|
};
|
||||||
|
|
||||||
garage = pkgs.callPackage ./garage.nix {
|
|
||||||
inherit (self'.packages) garage consul;
|
|
||||||
inherit (self) nixosModules;
|
|
||||||
inherit (config) cluster;
|
|
||||||
};
|
|
||||||
|
|
||||||
ipfs-cluster-upgrade = pkgs.callPackage ./ipfs-cluster-upgrade.nix {
|
ipfs-cluster-upgrade = pkgs.callPackage ./ipfs-cluster-upgrade.nix {
|
||||||
inherit (self) nixosModules;
|
inherit (self) nixosModules;
|
||||||
previous = timeMachine.preUnstable;
|
previous = timeMachine.preUnstable;
|
||||||
|
@ -35,11 +29,6 @@ in
|
||||||
inherit (self'.packages) keycloak;
|
inherit (self'.packages) keycloak;
|
||||||
};
|
};
|
||||||
|
|
||||||
patroni = pkgs.callPackage ./patroni.nix {
|
|
||||||
inherit (self) nixosModules;
|
|
||||||
inherit (self'.packages) postgresql;
|
|
||||||
};
|
|
||||||
|
|
||||||
s3ql-upgrade = pkgs.callPackage ./s3ql-upgrade.nix {
|
s3ql-upgrade = pkgs.callPackage ./s3ql-upgrade.nix {
|
||||||
inherit (self'.packages) s3ql;
|
inherit (self'.packages) s3ql;
|
||||||
inherit (self) nixosModules;
|
inherit (self) nixosModules;
|
||||||
|
|
|
@ -1,155 +0,0 @@
|
||||||
{ testers, nixosModules, cluster, garage, consul }:
|
|
||||||
|
|
||||||
testers.runNixOSTest {
|
|
||||||
name = "garage";
|
|
||||||
|
|
||||||
imports = [
|
|
||||||
./modules/consul.nix
|
|
||||||
];
|
|
||||||
|
|
||||||
extraBaseModules.services.consul.package = consul;
|
|
||||||
|
|
||||||
nodes = let
|
|
||||||
common = { config, lib, ... }: let
|
|
||||||
inherit (config.networking) hostName primaryIPAddress;
|
|
||||||
in {
|
|
||||||
imports = lib.flatten [
|
|
||||||
./modules/nixos/age-dummy-secrets
|
|
||||||
./modules/nixos/age-dummy-secrets/options.nix
|
|
||||||
nixosModules.ascensions
|
|
||||||
nixosModules.systemd-extras
|
|
||||||
nixosModules.consul-distributed-services
|
|
||||||
nixosModules.port-magic
|
|
||||||
cluster.config.services.storage.nixos.garage
|
|
||||||
cluster.config.services.storage.nixos.garageInternal
|
|
||||||
cluster.config.services.consul.nixos.ready
|
|
||||||
];
|
|
||||||
options.services.locksmith.providers = lib.mkOption {
|
|
||||||
type = lib.types.raw;
|
|
||||||
};
|
|
||||||
config = {
|
|
||||||
links.consulAgent = {
|
|
||||||
protocol = "http";
|
|
||||||
hostname = "consul";
|
|
||||||
port = 8500;
|
|
||||||
};
|
|
||||||
_module.args = {
|
|
||||||
depot.packages = { inherit garage; };
|
|
||||||
cluster.config = {
|
|
||||||
hostLinks.${hostName} = {
|
|
||||||
garageRpc.tuple = "${primaryIPAddress}:3901";
|
|
||||||
garageS3.tuple = "${primaryIPAddress}:8080";
|
|
||||||
garageWeb.tuple = "${primaryIPAddress}:8081";
|
|
||||||
};
|
|
||||||
links.garageWeb.hostname = "web.garage.example.com";
|
|
||||||
vars.meshNet.cidr = "192.168.0.0/16";
|
|
||||||
};
|
|
||||||
};
|
|
||||||
environment.etc."dummy-secrets/garageRpcSecret".text = "aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa";
|
|
||||||
networking.firewall.allowedTCPPorts = [ 3901 8080 ];
|
|
||||||
services.garage = {
|
|
||||||
layout.initial = lib.mkOverride 51 {
|
|
||||||
garage1 = { zone = "dc1"; capacity = 1000; };
|
|
||||||
garage2 = { zone = "dc1"; capacity = 1000; };
|
|
||||||
garage3 = { zone = "dc1"; capacity = 1000; };
|
|
||||||
};
|
|
||||||
};
|
|
||||||
system.ascensions.garage-layout.incantations = lib.mkOverride 51 (i: [ ]);
|
|
||||||
specialisation.modifiedLayout = {
|
|
||||||
inheritParentConfig = true;
|
|
||||||
configuration = {
|
|
||||||
services.garage = {
|
|
||||||
layout.initial = lib.mkForce {
|
|
||||||
garage1 = { zone = "dc1"; capacity = 2000; };
|
|
||||||
garage2 = { zone = "dc1"; capacity = 1000; };
|
|
||||||
garage3 = { zone = "dc1"; capacity = 1000; };
|
|
||||||
};
|
|
||||||
keys.testKey.allow.createBucket = true;
|
|
||||||
buckets = {
|
|
||||||
bucket1 = {
|
|
||||||
allow.testKey = [ "read" "write" ];
|
|
||||||
quotas = {
|
|
||||||
maxObjects = 300;
|
|
||||||
maxSize = 400 * 1024 * 1024;
|
|
||||||
};
|
|
||||||
};
|
|
||||||
bucket2 = {
|
|
||||||
allow.testKey = [ "read" ];
|
|
||||||
};
|
|
||||||
};
|
|
||||||
};
|
|
||||||
system.ascensions.garage-layout.incantations = lib.mkForce (i: [
|
|
||||||
(i.runGarage ''
|
|
||||||
garage layout assign -z dc1 -c 2000 "$(garage node id -q | cut -d@ -f1)"
|
|
||||||
garage layout apply --version 2
|
|
||||||
'')
|
|
||||||
]);
|
|
||||||
};
|
|
||||||
};
|
|
||||||
};
|
|
||||||
};
|
|
||||||
in {
|
|
||||||
garage1.imports = [ common ];
|
|
||||||
garage2.imports = [ common ];
|
|
||||||
garage3.imports = [ common ];
|
|
||||||
};
|
|
||||||
|
|
||||||
testScript = { nodes, ... }: /*python*/ ''
|
|
||||||
nodes = [garage1, garage2, garage3]
|
|
||||||
|
|
||||||
start_all()
|
|
||||||
|
|
||||||
with subtest("should bootstrap new cluster"):
|
|
||||||
for node in nodes:
|
|
||||||
node.wait_for_unit("garage.service")
|
|
||||||
|
|
||||||
for node in nodes:
|
|
||||||
node.wait_until_fails("garage status | grep 'NO ROLE ASSIGNED'")
|
|
||||||
|
|
||||||
with subtest("should apply new layout with ascension"):
|
|
||||||
for node in nodes:
|
|
||||||
node.wait_until_succeeds('test "$(systemctl list-jobs | wc -l)" -eq 1')
|
|
||||||
|
|
||||||
for node in nodes:
|
|
||||||
node.succeed("/run/current-system/specialisation/modifiedLayout/bin/switch-to-configuration test")
|
|
||||||
|
|
||||||
for node in nodes:
|
|
||||||
node.wait_until_succeeds("garage layout show | grep -w 2000")
|
|
||||||
assert "1" in node.succeed("garage layout show | grep -w 2000 | wc -l")
|
|
||||||
assert "2" in node.succeed("garage layout show | grep -w 1000 | wc -l")
|
|
||||||
|
|
||||||
with subtest("should apply new layout from scratch"):
|
|
||||||
for node in nodes:
|
|
||||||
node.systemctl("stop garage.service")
|
|
||||||
node.succeed("rm -rf /var/lib/garage-metadata")
|
|
||||||
|
|
||||||
for node in nodes:
|
|
||||||
node.systemctl("start garage.service")
|
|
||||||
|
|
||||||
for node in nodes:
|
|
||||||
node.wait_for_unit("garage.service")
|
|
||||||
|
|
||||||
for node in nodes:
|
|
||||||
node.wait_until_fails("garage status | grep 'NO ROLE ASSIGNED'")
|
|
||||||
|
|
||||||
for node in nodes:
|
|
||||||
node.wait_until_succeeds("garage layout show | grep -w 2000")
|
|
||||||
assert "1" in node.succeed("garage layout show | grep -w 2000 | wc -l")
|
|
||||||
assert "2" in node.succeed("garage layout show | grep -w 1000 | wc -l")
|
|
||||||
|
|
||||||
with subtest("should create specified buckets and keys"):
|
|
||||||
for node in nodes:
|
|
||||||
node.wait_until_succeeds('test "$(systemctl is-active garage-apply)" != activating')
|
|
||||||
garage1.succeed("garage key list | grep testKey")
|
|
||||||
garage1.succeed("garage bucket list | grep bucket1")
|
|
||||||
garage1.succeed("garage bucket list | grep bucket2")
|
|
||||||
|
|
||||||
with subtest("should delete unspecified buckets and keys"):
|
|
||||||
garage1.succeed("garage bucket create unwantedbucket")
|
|
||||||
garage1.succeed("garage key new --name unwantedkey")
|
|
||||||
garage1.succeed("systemctl restart garage-apply.service")
|
|
||||||
|
|
||||||
garage1.fail("garage key list | grep unwantedkey")
|
|
||||||
garage1.fail("garage bucket list | grep unwantedbucket")
|
|
||||||
'';
|
|
||||||
}
|
|
|
@ -1,211 +0,0 @@
|
||||||
{ nixosTest, nixosModules, postgresql }:
|
|
||||||
|
|
||||||
# taken from https://github.com/phfroidmont/nixpkgs/blob/patroni-module/nixos/tests/patroni.nix
|
|
||||||
nixosTest (
|
|
||||||
let
|
|
||||||
nodesIps = [
|
|
||||||
"192.168.1.1"
|
|
||||||
"192.168.1.2"
|
|
||||||
"192.168.1.3"
|
|
||||||
];
|
|
||||||
|
|
||||||
createNode = index: { pkgs, ... }:
|
|
||||||
let
|
|
||||||
ip = builtins.elemAt nodesIps index; # since we already use IPs to identify servers
|
|
||||||
in
|
|
||||||
{
|
|
||||||
imports = [
|
|
||||||
nixosModules.patroni
|
|
||||||
nixosModules.systemd-extras
|
|
||||||
];
|
|
||||||
|
|
||||||
networking.interfaces.eth1.ipv4.addresses = pkgs.lib.mkOverride 0 [
|
|
||||||
{ address = ip; prefixLength = 16; }
|
|
||||||
];
|
|
||||||
|
|
||||||
networking.firewall.allowedTCPPorts = [ 5432 8008 5010 ];
|
|
||||||
|
|
||||||
environment.systemPackages = [ pkgs.jq ];
|
|
||||||
|
|
||||||
services.patroni = {
|
|
||||||
|
|
||||||
enable = true;
|
|
||||||
|
|
||||||
postgresqlPackage = postgresql.withPackages (p: [ p.pg_safeupdate ]);
|
|
||||||
|
|
||||||
scope = "cluster1";
|
|
||||||
name = "node${toString(index + 1)}";
|
|
||||||
nodeIp = ip;
|
|
||||||
otherNodesIps = builtins.filter (h: h != ip) nodesIps;
|
|
||||||
softwareWatchdog = true;
|
|
||||||
|
|
||||||
settings = {
|
|
||||||
bootstrap = {
|
|
||||||
dcs = {
|
|
||||||
ttl = 30;
|
|
||||||
loop_wait = 10;
|
|
||||||
retry_timeout = 10;
|
|
||||||
maximum_lag_on_failover = 1048576;
|
|
||||||
};
|
|
||||||
initdb = [
|
|
||||||
{ encoding = "UTF8"; }
|
|
||||||
"data-checksums"
|
|
||||||
];
|
|
||||||
};
|
|
||||||
|
|
||||||
postgresql = {
|
|
||||||
use_pg_rewind = true;
|
|
||||||
use_slots = true;
|
|
||||||
authentication = {
|
|
||||||
replication = {
|
|
||||||
username = "replicator";
|
|
||||||
};
|
|
||||||
superuser = {
|
|
||||||
username = "postgres";
|
|
||||||
};
|
|
||||||
rewind = {
|
|
||||||
username = "rewind";
|
|
||||||
};
|
|
||||||
};
|
|
||||||
parameters = {
|
|
||||||
listen_addresses = "${ip}";
|
|
||||||
wal_level = "replica";
|
|
||||||
hot_standby_feedback = "on";
|
|
||||||
unix_socket_directories = "/tmp";
|
|
||||||
};
|
|
||||||
pg_hba = [
|
|
||||||
"host replication replicator 192.168.1.0/24 md5"
|
|
||||||
# Unsafe, do not use for anything other than tests
|
|
||||||
"host all all 0.0.0.0/0 trust"
|
|
||||||
];
|
|
||||||
};
|
|
||||||
|
|
||||||
etcd3 = {
|
|
||||||
host = "192.168.1.4:2379";
|
|
||||||
};
|
|
||||||
};
|
|
||||||
|
|
||||||
environmentFiles = {
|
|
||||||
PATRONI_REPLICATION_PASSWORD = pkgs.writeText "replication-password" "postgres";
|
|
||||||
PATRONI_SUPERUSER_PASSWORD = pkgs.writeText "superuser-password" "postgres";
|
|
||||||
PATRONI_REWIND_PASSWORD = pkgs.writeText "rewind-password" "postgres";
|
|
||||||
};
|
|
||||||
};
|
|
||||||
|
|
||||||
# We always want to restart so the tests never hang
|
|
||||||
systemd.services.patroni.serviceConfig.StartLimitIntervalSec = 0;
|
|
||||||
};
|
|
||||||
in
|
|
||||||
{
|
|
||||||
name = "patroni";
|
|
||||||
|
|
||||||
nodes = {
|
|
||||||
node1 = createNode 0;
|
|
||||||
node2 = createNode 1;
|
|
||||||
node3 = createNode 2;
|
|
||||||
|
|
||||||
etcd = { pkgs, ... }: {
|
|
||||||
|
|
||||||
networking.interfaces.eth1.ipv4.addresses = pkgs.lib.mkOverride 0 [
|
|
||||||
{ address = "192.168.1.4"; prefixLength = 16; }
|
|
||||||
];
|
|
||||||
|
|
||||||
services.etcd = {
|
|
||||||
enable = true;
|
|
||||||
listenClientUrls = [ "http://192.168.1.4:2379" ];
|
|
||||||
};
|
|
||||||
|
|
||||||
networking.firewall.allowedTCPPorts = [ 2379 ];
|
|
||||||
};
|
|
||||||
|
|
||||||
client = { pkgs, ... }: {
|
|
||||||
environment.systemPackages = [ postgresql ];
|
|
||||||
|
|
||||||
networking.interfaces.eth1.ipv4.addresses = pkgs.lib.mkOverride 0 [
|
|
||||||
{ address = "192.168.2.1"; prefixLength = 16; }
|
|
||||||
];
|
|
||||||
|
|
||||||
services.haproxy = {
|
|
||||||
enable = true;
|
|
||||||
config = ''
|
|
||||||
global
|
|
||||||
maxconn 100
|
|
||||||
|
|
||||||
defaults
|
|
||||||
log global
|
|
||||||
mode tcp
|
|
||||||
retries 2
|
|
||||||
timeout client 30m
|
|
||||||
timeout connect 4s
|
|
||||||
timeout server 30m
|
|
||||||
timeout check 5s
|
|
||||||
|
|
||||||
listen cluster1
|
|
||||||
bind 127.0.0.1:5432
|
|
||||||
option httpchk
|
|
||||||
http-check expect status 200
|
|
||||||
default-server inter 3s fall 3 rise 2 on-marked-down shutdown-sessions
|
|
||||||
${builtins.concatStringsSep "\n" (map (ip: "server postgresql_${ip}_5432 ${ip}:5432 maxconn 100 check port 8008") nodesIps)}
|
|
||||||
'';
|
|
||||||
};
|
|
||||||
};
|
|
||||||
};
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
testScript = ''
|
|
||||||
nodes = [node1, node2, node3]
|
|
||||||
|
|
||||||
def wait_for_all_nodes_ready(expected_replicas=2):
|
|
||||||
booted_nodes = filter(lambda node: node.booted, nodes)
|
|
||||||
for node in booted_nodes:
|
|
||||||
print(node.succeed("patronictl list cluster1"))
|
|
||||||
node.wait_until_succeeds(f"[ $(patronictl list -f json cluster1 | jq 'length') == {expected_replicas + 1} ]")
|
|
||||||
node.wait_until_succeeds("[ $(patronictl list -f json cluster1 | jq 'map(select(.Role | test(\"^Leader$\"))) | map(select(.State | test(\"^running$\"))) | length') == 1 ]")
|
|
||||||
node.wait_until_succeeds(f"[ $(patronictl list -f json cluster1 | jq 'map(select(.Role | test(\"^Replica$\"))) | map(select(.State | test(\"^streaming$\"))) | length') == {expected_replicas} ]")
|
|
||||||
print(node.succeed("patronictl list cluster1"))
|
|
||||||
client.wait_until_succeeds("psql -h 127.0.0.1 -U postgres --command='select 1;'")
|
|
||||||
|
|
||||||
def run_dummy_queries():
|
|
||||||
client.succeed("psql -h 127.0.0.1 -U postgres --pset='pager=off' --tuples-only --command='insert into dummy(val) values (101);'")
|
|
||||||
client.succeed("test $(psql -h 127.0.0.1 -U postgres --pset='pager=off' --tuples-only --command='select val from dummy where val = 101;') -eq 101")
|
|
||||||
client.succeed("psql -h 127.0.0.1 -U postgres --pset='pager=off' --tuples-only --command='delete from dummy where val = 101;'")
|
|
||||||
|
|
||||||
start_all()
|
|
||||||
|
|
||||||
with subtest("should bootstrap a new patroni cluster"):
|
|
||||||
wait_for_all_nodes_ready()
|
|
||||||
|
|
||||||
with subtest("should be able to insert and select"):
|
|
||||||
client.succeed("psql -h 127.0.0.1 -U postgres --command='create table dummy as select * from generate_series(1, 100) as val;'")
|
|
||||||
client.succeed("test $(psql -h 127.0.0.1 -U postgres --pset='pager=off' --tuples-only --command='select count(distinct val) from dummy;') -eq 100")
|
|
||||||
|
|
||||||
with subtest("should restart after all nodes are crashed"):
|
|
||||||
for node in nodes:
|
|
||||||
node.crash()
|
|
||||||
for node in nodes:
|
|
||||||
node.start()
|
|
||||||
wait_for_all_nodes_ready()
|
|
||||||
|
|
||||||
with subtest("should be able to run queries while any one node is crashed"):
|
|
||||||
masterNodeName = node1.succeed("patronictl list -f json cluster1 | jq '.[] | select(.Role | test(\"^Leader$\")) | .Member' -r").strip()
|
|
||||||
masterNodeIndex = int(masterNodeName[len(masterNodeName)-1]) - 1
|
|
||||||
|
|
||||||
# Move master node at the end of the list to avoid multiple failovers (makes the test faster and more consistent)
|
|
||||||
nodes.append(nodes.pop(masterNodeIndex))
|
|
||||||
|
|
||||||
for node in nodes:
|
|
||||||
node.crash()
|
|
||||||
wait_for_all_nodes_ready(1)
|
|
||||||
|
|
||||||
# Execute some queries while a node is down.
|
|
||||||
run_dummy_queries()
|
|
||||||
|
|
||||||
# Restart crashed node.
|
|
||||||
node.start()
|
|
||||||
wait_for_all_nodes_ready()
|
|
||||||
|
|
||||||
# Execute some queries with the node back up.
|
|
||||||
run_dummy_queries()
|
|
||||||
'';
|
|
||||||
})
|
|
Loading…
Add table
Reference in a new issue