forked from input-output-hk/sanchonet-demo
-
Notifications
You must be signed in to change notification settings - Fork 0
/
Justfile
974 lines (838 loc) · 30 KB
/
Justfile
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
import? 'scripts/recipes-aws.just'
import? 'scripts/recipes-custom.just'
set shell := ["bash", "-uc"]
set positional-arguments
alias terraform := tofu
alias tf := tofu
# Defaults
null := ""
stateDir := "STATEDIR=" + statePrefix / "$(basename $(git remote get-url origin))"
statePrefix := "~/.local/share"
# Environment variables can be used to change the default template diff and path comparison sources.
# If TEMPLATE_PATH is set, it will have precedence, otherwise git url will be used for source templates.
templateBranch := env_var_or_default("TEMPLATE_BRANCH","main")
templatePath := env_var_or_default("TEMPLATE_PATH","no-path-given")
templateRepo := env_var_or_default("TEMPLATE_REPO","cardano-parts")
templateUrl := "https://raw.githubusercontent.com/input-output-hk/" + templateRepo + "/" + templateBranch + "/templates/cardano-parts-project"
# Common code
checkEnv := '''
TESTNET_MAGIC="${2:-""}"
''' + checkEnvWithoutOverride + '''
# Allow a magic override if the just recipe optional var is provided
if ! [ -z "${TESTNET_MAGIC:-}" ]; then
MAGIC="$TESTNET_MAGIC"
fi
'''
checkEnvWithoutOverride := '''
ENV="${1:-}"
if ! [[ "$ENV" =~ mainnet$|preprod$|preview$|private$|sanchonet$|shelley-qa$|demo$ ]]; then
echo "Error: only node environments for demo, mainnet, preprod, preview, private, sanchonet and shelley-qa are supported"
exit 1
fi
if [ "$ENV" = "mainnet" ]; then
MAGIC="764824073"
elif [ "$ENV" = "preprod" ]; then
MAGIC="1"
elif [ "$ENV" = "preview" ]; then
MAGIC="2"
elif [ "$ENV" = "shelley-qa" ]; then
MAGIC="3"
elif [ "$ENV" = "sanchonet" ]; then
MAGIC="4"
elif [ "$ENV" = "private" ]; then
MAGIC="5"
elif [ "$ENV" = "demo" ]; then
MAGIC="42"
fi
'''
checkSshConfig := '''
if not ('.ssh_config' | path exists) {
print "Please run `just save-ssh-config` first to create the .ssh_config file"
exit 1
}
'''
checkSshKey := '''
if not ('.ssh_key' | path exists) {
just save-bootstrap-ssh-key
}
'''
sopsConfigSetup := '''
# To support searching for sops config files from the target path rather than cwd up,
# implement a userland solution until natively sops supported.
#
# This enables $NO_DEPLOY_DIR to be separate from the default $STAKE_POOL_DIR/no-deploy default location.
# Ref: https://github.com/getsops/sops/issues/242#issuecomment-999809670
function sops_config() {
# Suppress xtrace on this fn as the return string is observed from the caller's output
{ SHOPTS="$-"; set +x; } 2> /dev/null
FILE="$1"
CONFIG_DIR=$(dirname "$(realpath "$FILE")")
while ! [ -f "$CONFIG_DIR/.sops.yaml" ]; do
if [ "$CONFIG_DIR" = "/" ]; then
>&2 echo "error: no .sops.yaml file was found while walking the directory structure upwards from the target file: \"$FILE\""
exit 1
fi
CONFIG_DIR=$(dirname "$CONFIG_DIR")
done
echo "$CONFIG_DIR/.sops.yaml"
# Reset the xtrace option to its state prior to suppression
[ -n "${SHOPTS//[^x]/}" ] && set -x
}
'''
# List all just recipes available
default:
@just --list
# Deploy select machines
apply *ARGS:
colmena apply --verbose --on {{ARGS}}
# Deploy all machines
apply-all *ARGS:
colmena apply --verbose {{ARGS}}
# Deploy select machines with the bootstrap key
apply-bootstrap *ARGS:
SSH_CONFIG=<(sed '6i IdentityFile .ssh_key' .ssh_config) colmena apply --verbose --on {{ARGS}}
# Build the prod cardano book
build-book-prod:
#!/usr/bin/env bash
set -e
cd mdbook
ln -sf book-prod.toml book.toml
cd -
mdbook build mdbook/
echo
nu -c 'echo $"(ansi bg_light_purple)REMINDER:(ansi reset) Ensure node version statement and link for each environment are up to date."'
# Build the staging cardano book
build-book-staging:
#!/usr/bin/env bash
set -e
cd mdbook
ln -sf book-staging.toml book.toml
cd -
mdbook build mdbook/
echo
nu -c 'echo $"(ansi bg_light_purple)REMINDER:(ansi reset) Ensure node version statement and link for each environment are up to date."'
# Build a nixos configuration
build-machine MACHINE *ARGS:
nix build -L .#nixosConfigurations.{{MACHINE}}.config.system.build.toplevel {{ARGS}}
# Build all nixosConfigurations
build-machines *ARGS:
#!/usr/bin/env nu
let nodes = (nix eval --json '.#nixosConfigurations' --apply builtins.attrNames | from json)
for node in $nodes {just build-machine $node {{ARGS}}}
# Deploy a cloudFormation stack
cf STACKNAME:
#!/usr/bin/env nu
mkdir cloudFormation
nix eval --json '.#cloudFormation.{{STACKNAME}}' | from json | save --force 'cloudFormation/{{STACKNAME}}.json'
rain deploy --debug --termination-protection --yes ./cloudFormation/{{STACKNAME}}.json
# Prep dbsync for delegation analysis
dbsync-prep ENV HOST ACCTS="500":
#!/usr/bin/env bash
set -euo pipefail
TMPFILE="/tmp/create-faucet-stake-keys-table-{{ENV}}.sql"
echo "Creating stake key sql injection command for environment {{ENV}} (this will take a minute)..."
NOMENU=true \
scripts/setup-delegation-accounts.py \
--print-only \
--wallet-mnemonic <(sops -d secrets/envs/{{ENV}}/utxo-keys/faucet.mnemonic) \
--num-accounts {{ACCTS}} \
> "$TMPFILE"
echo
echo "Pushing stake key sql injection command for environment {{ENV}}..."
just scp "$TMPFILE" {{HOST}}:"$TMPFILE"
echo
echo "Executing stake key sql injection command for environment {{ENV}}..."
just ssh {{HOST}} -t "psql -XU cexplorer cexplorer < \"$TMPFILE\""
# Start a remote dbsync psql session
dbsync-psql HOSTNAME:
#!/usr/bin/env bash
just ssh {{HOSTNAME}} -t 'psql -U cexplorer cexplorer'
# Analyze pool performance
dbsync-pool-analyze HOSTNAME:
#!/usr/bin/env bash
set -euo pipefail
echo "Pushing pool analysis sql command on {{HOSTNAME}}..."
just scp scripts/dbsync-pool-perf.sql {{HOSTNAME}}:/tmp/
echo
echo "Executing pool analysis sql command on host {{HOSTNAME}}..."
QUERY=$(just ssh {{HOSTNAME}} -t 'psql -P pager=off -xXU cexplorer cexplorer < /tmp/dbsync-pool-perf.sql')
echo
echo "Query output:"
echo "$QUERY" | tail -n +2
echo
JSON=$(grep -oP '^faucet_pool_summary_json[[:space:]]+\| \K{.*$' <<< "$QUERY" | jq .)
echo "$JSON"
echo
echo "Faucet pools to de-delegate are:"
jq '.faucet_to_dedelegate' <<< "$JSON"
echo
echo "The string of indexes of faucet pools to de-delegate from the JSON above are:"
jq '.faucet_to_dedelegate | to_entries | map(.key) | join(" ")' <<< "$JSON"
echo
MAX_SHIFT=$(grep -oP '^faucet_pool_to_dedelegate_shift_pct[[:space:]]+\| \K.*$' <<< "$QUERY")
echo "The maximum percentage difference de-delegation of all these pools will make in chain density is: $MAX_SHIFT"
# De-delegation pools for given faucet stake indexes
dedelegate-pools ENV *IDXS=null:
#!/usr/bin/env bash
set -euo pipefail
{{checkEnvWithoutOverride}}
if [ "{{ENV}}" = "mainnet" ]; then
echo "Dedelegation cannot be performed on the mainnet environment"
exit 1
fi
just set-default-cardano-env {{ENV}} "$MAGIC" "$PPID"
if [ "$(jq -re .syncProgress <<< "$(just query-tip {{ENV}})")" != "100.00" ]; then
echo "Please wait until the local tip of environment {{ENV}} is 100.00 before dedelegation"
exit 1
fi
echo
read -p "Press any key to start de-delegating {{ENV}} faucet pool delegations for stake key indexes {{IDXS}}" -n 1 -r -s
echo
echo "Starting de-delegation of the following stake key indexes: {{IDXS}}"
for i in {{IDXS}}; do
echo "De-delegating index $i"
NOMENU=true scripts/restore-delegation-accounts.py \
--testnet-magic "$MAGIC" \
--signing-key-file <(just sops-decrypt-binary secrets/envs/{{ENV}}/utxo-keys/rich-utxo.skey) \
--wallet-mnemonic <(just sops-decrypt-binary secrets/envs/{{ENV}}/utxo-keys/faucet.mnemonic) \
--delegation-index "$i"
echo "Sleeping 2 minutes until $(date -d @$(($(date +%s) + 120)))"
sleep 120
echo
echo
done
# Get a wallet address from mnemonic file
gen-payment-address FILE OFFSET="0":
cardano-address key from-recovery-phrase Shelley < {{FILE}} \
| cardano-address key child 1852H/1815H/0H/0/{{OFFSET}} \
| cardano-address key public --with-chain-code \
| cardano-address address payment --network-tag testnet
# Standard lint check
lint:
deadnix -f
statix check
# List machines
list-machines:
#!/usr/bin/env nu
let nixosNodes = (do -i { ^nix eval --json '.#nixosConfigurations' --apply 'builtins.attrNames' } | complete)
if $nixosNodes.exit_code != 0 {
print "Nixos failed to evaluate the .#nixosConfigurations attribute."
print "The output was:"
print
print $nixosNodes
exit 1
}
{{checkSshConfig}}
let sshNodes = (do -i { ^scj dump /dev/stdout -c .ssh_config } | complete)
if $sshNodes.exit_code != 0 {
print "Ssh-config-json failed to evaluate the .ssh_config file."
print "The output was:"
print
print $sshNodes
exit 1
}
let nixosNodesDfr = (
let nodeList = ($nixosNodes.stdout | from json);
let sanitizedList = (if ($nodeList | is-empty) {$nodeList | insert 0 ""} else {$nodeList});
$sanitizedList
| insert 0 "machine"
| each {|i| [$i] | into record}
| headers
| each {|i| insert inNixosCfg {"yes"}}
| dfr into-df
)
let sshNodesDfr = (
let sshTable = ($sshNodes.stdout | from json | where ('HostName' in $it));
if ($sshTable | is-empty) {
[[Host IP]; ["" ""]] | dfr into-df
}
else {
$sshTable | rename Host IP | dfr into-df
}
)
(
$nixosNodesDfr
| dfr join -o $sshNodesDfr machine Host
| dfr sort-by machine
| dfr into-nu
| update cells {|v| if $v == null {"Missing"} else {$v}}
| where machine != ""
)
# Check mimir required config
mimir-alertmanager-bootstrap:
#!/usr/bin/env bash
set -euo pipefail
echo "Enter the mimir admin username: "
read -s MIMIR_USER
echo
echo "Enter the mimir admin token: "
read -s MIMIR_TOKEN
echo
echo "Enter the mimir base monitoring fqdn without the HTTPS:// scheme: "
read URL
echo
echo "Obtaining current mimir alertmanager config:"
echo "-----------"
mimirtool alertmanager get --address "https://$MIMIR_USER:$MIMIR_TOKEN@$URL/mimir" --id 1
echo "-----------"
echo
echo "If the output between the dashed lines above is blank, you may need to preload an initial alertmanager ruleset"
echo "for the mimir TF plugin to succeed, where the command to preload alertmanager is:"
echo
echo "mimirtool alertmanager load --address \"https://\$MIMIR_USER:\$MIMIR_TOKEN@$URL/mimir\" --id 1 alertmanager-bootstrap-config.yaml"
echo
echo "The contents of alertmanager-bootstrap-config.yaml can be:"
echo
echo "route:"
echo " group_wait: 0s"
echo " receiver: empty-receiver"
echo "receivers:"
echo " - name: 'empty-receiver'"
# Query the tip of all running envs
query-tip-all:
#!/usr/bin/env bash
set -euo pipefail
QUERIED=0
for i in mainnet preprod preview private shelley-qa sanchonet demo; do
TIP=$(just query-tip $i 2>&1) && {
echo "Environment: $i"
echo "$TIP"
echo
QUERIED=$((QUERIED + 1))
}
done
[ "$QUERIED" = "0" ] && echo "No environments running." || true
# Query the current envs tip
query-tip ENV TESTNET_MAGIC=null:
#!/usr/bin/env bash
set -euo pipefail
{{checkEnv}}
{{stateDir}}
cardano-cli query tip \
--socket-path "$STATEDIR/node-{{ENV}}.socket" \
--testnet-magic "$MAGIC"
# Save the cluster bootstrap ssh key
save-bootstrap-ssh-key:
#!/usr/bin/env nu
print "Retrieving ssh key from tofu..."
nix build ".#opentofu.cluster" --out-link terraform.tf.json
tofu init -reconfigure
tofu workspace select -or-create cluster
let tf = (tofu show -json | from json)
let key = ($tf.values.root_module.resources | where type == tls_private_key and name == bootstrap)
$key.values.private_key_openssh | to text | save --force .ssh_key
chmod 0600 .ssh_key
# Save bulk credentials from select commit
save-bulk-creds ENV COMMIT="HEAD":
#!/usr/bin/env bash
mkdir -p workbench/custom/rundir
DATE=$(git show --no-patch --format=%cs {{COMMIT}})
for i in 1 2 3; do
sops --config /dev/null --input-type binary --output-type binary --decrypt \
<(git cat-file blob {{COMMIT}}:secrets/groups/{{ENV}}$i/no-deploy/bulk.creds.pools.json) \
| jq -r '.[]'
done \
| jq -s \
> "workbench/custom/rundir/bulk.creds.secret.{{ENV}}.{{COMMIT}}.$DATE.pools.json"
echo
echo "Bulk credentials file for environment {{ENV}}, commit {{COMMIT}} has been saved at:"
echo " workbench/custom/rundir/bulk.creds.secret.{{ENV}}.{{COMMIT}}.$DATE.pools.json"
echo
echo "Do not commit this file and delete it when local workbench work is completed."
# Save ssh config
save-ssh-config:
#!/usr/bin/env nu
print "Retrieving ssh config from tofu..."
nix build ".#opentofu.cluster" --out-link terraform.tf.json
tofu init -reconfigure
tofu workspace select -or-create cluster
let tf = (tofu show -json | from json)
let key = ($tf.values.root_module.resources | where type == local_file and name == ssh_config)
$key.values.content | to text | save --force .ssh_config
chmod 0600 .ssh_config
# Set the shell's default node env
set-default-cardano-env ENV TESTNET_MAGIC=null PPID=null:
#!/usr/bin/env bash
set -euo pipefail
{{checkEnv}}
{{stateDir}}
# The log and socket file may not exist immediately upon node startup, so only check for the pid file
if ! [ -s "$STATEDIR/node-{{ENV}}.pid" ]; then
echo "Environment {{ENV}} does not appear to be running as $STATEDIR/node-{{ENV}}.pid does not exist"
exit 1
fi
echo "Linking: $(ln -sfv "$STATEDIR/node-{{ENV}}.socket" node.socket)"
echo "Linking: $(ln -sfv "$STATEDIR/node-{{ENV}}.log" node.log)"
echo
if [ -n "{{PPID}}" ]; then
PARENTID="{{PPID}}"
else
PARENTID="$PPID"
fi
SHELLPID=$(cat /proc/$PARENTID/status | awk '/PPid/ {print $2}')
DEFAULT_PATH=$(pwd)/node.socket
echo "Updating shell env vars:"
echo " CARDANO_NODE_SOCKET_PATH=$DEFAULT_PATH"
echo " CARDANO_NODE_NETWORK_ID=$MAGIC"
echo " TESTNET_MAGIC=$MAGIC"
SH=$(cat /proc/$SHELLPID/comm)
if [[ "$SH" =~ bash$|zsh$ ]]; then
# Modifying a parent shells env vars is generally not done
# This is a hacky way to accomplish it in bash and zsh
gdb -iex "set auto-load no" /proc/$SHELLPID/exe $SHELLPID <<END >/dev/null
call (int) setenv("CARDANO_NODE_SOCKET_PATH", "$DEFAULT_PATH", 1)
call (int) setenv("CARDANO_NODE_NETWORK_ID", "$MAGIC", 1)
call (int) setenv("TESTNET_MAGIC", "$MAGIC", 1)
END
# Zsh env vars get updated, but the shell doesn't reflect this
if [ "$SH" = "zsh" ]; then
echo
echo "Cardano env vars have been updated as seen by \`env\`, but zsh \`echo \$VAR\` will not reflect this."
echo "To sync zsh shell vars with env vars:"
echo " source scripts/sync-env-vars.sh"
fi
else
echo
echo "Unexpected shell: $SH"
echo "The following vars will need to be manually exported, or the equivalent operation for your shell:"
echo " export CARDANO_NODE_SOCKET_PATH=$DEFAULT_PATH"
echo " export CARDANO_NODE_NETWORK_ID=$MAGIC"
echo " export TESTNET_MAGIC=$MAGIC"
fi
# Show nix flake details
show-flake *ARGS:
nix flake show --allow-import-from-derivation {{ARGS}}
# Show DNS nameservers
show-nameservers:
#!/usr/bin/env nu
let domain = (nix eval --raw '.#cardano-parts.cluster.infra.aws.domain')
let zones = (aws route53 list-hosted-zones-by-name | from json).HostedZones
let id = ($zones | where Name == $"($domain).").Id.0
let sets = (aws route53 list-resource-record-sets --hosted-zone-id $id | from json).ResourceRecordSets
let ns = ($sets | where Type == "NS").ResourceRecords.0.Value
print "Nameservers for the following hosted zone need to be added to the NS record of the delegating authority"
print $"Nameservers for domain: ($domain) \(hosted zone id: ($id)) are:"
print ($ns | to text)
# Decrypt a file to stdout
sops-decrypt-binary FILE:
#!/usr/bin/env bash
set -euo pipefail
{{sopsConfigSetup}}
[ -n "${DEBUG:-}" ] && set -x
# Default to stdout decrypted output.
# This supports the common use case of obtaining decrypted state for cmd arg input while leaving the encrypted file intact on disk.
sops --config "$(sops_config {{FILE}})" --input-type binary --output-type binary --decrypt {{FILE}}
# Encrypt a file in place
sops-encrypt-binary FILE:
#!/usr/bin/env bash
set -euo pipefail
{{sopsConfigSetup}}
[ -n "${DEBUG:-}" ] && set -x
# Default to in-place encrypted output.
# This supports the common use case of first time encrypting plaintext state for public storage, ex: git repo commit.
sops --config "$(sops_config {{FILE}})" --input-type binary --output-type binary --encrypt {{FILE}} | sponge {{FILE}}
# Rotate sops encryption
sops-rotate-binary FILE:
#!/usr/bin/env bash
set -euo pipefail
{{sopsConfigSetup}}
[ -n "${DEBUG:-}" ] && set -x
# Default to in-place encryption rotation.
# This supports the common use case of rekeying, for example if recipient keys have changed.
just sops-decrypt-binary {{FILE}} | sponge {{FILE}}
just sops-encrypt-binary {{FILE}}
# Scp using repo ssh config
scp *ARGS:
#!/usr/bin/env nu
{{checkSshConfig}}
scp -o LogLevel=ERROR -F .ssh_config {{ARGS}}
# Ssh using repo ssh config
ssh HOSTNAME *ARGS:
#!/usr/bin/env nu
{{checkSshConfig}}
ssh -o LogLevel=ERROR -F .ssh_config {{HOSTNAME}} {{ARGS}}
# Ssh using cluster bootstrap key
ssh-bootstrap HOSTNAME *ARGS:
#!/usr/bin/env nu
{{checkSshConfig}}
{{checkSshKey}}
ssh -o LogLevel=ERROR -F .ssh_config -i .ssh_key {{HOSTNAME}} {{ARGS}}
# Ssh to all
ssh-for-all *ARGS:
#!/usr/bin/env nu
let nodes = (nix eval --json '.#nixosConfigurations' --apply builtins.attrNames | from json)
$nodes | par-each {|node| just ssh -q $node {{ARGS}}}
# Ssh for select
ssh-for-each HOSTNAMES *ARGS:
colmena exec --verbose --parallel 0 --on {{HOSTNAMES}} {{ARGS}}
# List machine ips based on regex pattern
ssh-list-ips PATTERN:
#!/usr/bin/env nu
scj dump /dev/stdout -c .ssh_config | from json | default "" Host | default "" HostName | where Host =~ "{{PATTERN}}" | get HostName | str join " "
# List machine names based on regex pattern
ssh-list-names PATTERN:
#!/usr/bin/env nu
scj dump /dev/stdout -c .ssh_config | from json | default "" Host | default "" HostName | where Host =~ "{{PATTERN}}" | get Host | str join " "
# Start a fork to conway demo
start-demo:
#!/usr/bin/env bash
set -euo pipefail
just stop-node demo
{{stateDir}}
echo "Cleaning state-demo..."
if [ -d state-demo ]; then
chmod -R +w state-demo
rm -rf state-demo
fi
echo "Generating state-demo config..."
export ENV=custom
export GENESIS_DIR=state-demo
export KEY_DIR=state-demo/envs/custom
export DATA_DIR=state-demo/rundir
export CARDANO_NODE_SOCKET_PATH="$STATEDIR/node-demo.socket"
export TESTNET_MAGIC=42
export NUM_GENESIS_KEYS=3
export POOL_NAMES="sp-1 sp-2 sp-3"
export STAKE_POOL_DIR=state-demo/groups/stake-pools
export BULK_CREDS=state-demo/bulk.creds.all.json
export PAYMENT_KEY=state-demo/envs/custom/utxo-keys/rich-utxo
export UNSTABLE=true
export UNSTABLE_LIB=true
export USE_ENCRYPTION=true
export USE_DECRYPTION=true
export USE_NODE_CONFIG_BP=false
export USE_CREATE_TESTNET_DATA=false
export DEBUG=true
export SECURITY_PARAM=8
export SLOT_LENGTH=100
export START_TIME=$(date --utc +"%Y-%m-%dT%H:%M:%SZ" --date " now + 30 seconds")
if [ "$USE_CREATE_TESTNET_DATA" = true ]; then
ERA_CMD="alonzo" \
nix run .#job-gen-custom-node-config-data
else
nix run .#job-gen-custom-node-config
fi
nix run .#job-create-stake-pool-keys
if [ "$USE_DECRYPTION" = true ]; then
BFT_CREDS=$(just sops-decrypt-binary "$KEY_DIR"/delegate-keys/bulk.creds.bft.json)
POOL_CREDS=$(just sops-decrypt-binary "$STAKE_POOL_DIR"/no-deploy/bulk.creds.pools.json)
else
BFT_CREDS=$(cat "$KEY_DIR"/delegate-keys/bulk.creds.bft.json)
POOL_CREDS=$(cat "$STAKE_POOL_DIR"/no-deploy/bulk.creds.pools.json)
fi
(
jq -r '.[]' <<< "$BFT_CREDS"
jq -r '.[]' <<< "$POOL_CREDS"
) | jq -s > "$BULK_CREDS"
echo "Start cardano-node in the background. Run \"just stop-node demo\" to stop"
NODE_CONFIG="$DATA_DIR/node-config.json" \
NODE_TOPOLOGY="$DATA_DIR/topology.json" \
SOCKET_PATH="$STATEDIR/node-demo.socket" \
nohup setsid nix run .#run-cardano-node &> "$STATEDIR/node-demo.log" & echo $! > "$STATEDIR/node-demo.pid" &
just set-default-cardano-env demo "" "$PPID"
echo "Sleeping 30 seconds until $(date -d @$(($(date +%s) + 30)))"
sleep 30
echo
if [ "$USE_CREATE_TESTNET_DATA" = false ]; then
echo "Moving genesis utxo..."
BYRON_SIGNING_KEY="$KEY_DIR"/utxo-keys/shelley.000.skey \
ERA_CMD="alonzo" \
nix run .#job-move-genesis-utxo
echo "Sleeping 7 seconds until $(date -d @$(($(date +%s) + 7)))"
sleep 7
echo
fi
echo "Registering stake pools..."
POOL_RELAY=demo.local \
POOL_RELAY_PORT=3001 \
ERA_CMD="alonzo" \
nix run .#job-register-stake-pools
echo "Sleeping 7 seconds until $(date -d @$(($(date +%s) + 7)))"
sleep 7
echo
echo "Delegating rewards stake key..."
ERA_CMD="alonzo" \
nix run .#job-delegate-rewards-stake-key
echo "Sleeping 160 seconds until $(date -d @$(($(date +%s) + 160)))"
sleep 160
echo
echo "Forking to babbage..."
just query-tip demo
MAJOR_VERSION=7 \
ERA_CMD="alonzo" \
nix run .#job-update-proposal-hard-fork
echo "Sleeping 160 seconds until $(date -d @$(($(date +%s) + 160)))"
sleep 160
echo
echo "Forking to babbage (intra-era)..."
just query-tip demo
MAJOR_VERSION=8 \
ERA_CMD="babbage" \
nix run .#job-update-proposal-hard-fork
echo "Sleeping 160 seconds until $(date -d @$(($(date +%s) + 160)))"
sleep 160
echo
echo "Forking to conway..."
just query-tip demo
MAJOR_VERSION=9 \
ERA_CMD="babbage" \
nix run .#job-update-proposal-hard-fork
echo "Sleeping 160 seconds until $(date -d @$(($(date +%s) + 160)))"
sleep 160
echo
just query-tip demo
echo "Finished sequence..."
echo
# Start a node for specific env
start-node ENV:
#!/usr/bin/env bash
set -euo pipefail
{{stateDir}}
if ! [[ "{{ENV}}" =~ mainnet$|preprod$|preview$|private$|sanchonet$|shelley-qa$ ]]; then
echo "Error: only node environments for mainnet, preprod, preview, private, sanchonet and shelley-qa are supported for start-node recipe"
exit 1
fi
# Stop any existing running node env for a clean restart
just stop-node {{ENV}}
echo "Starting cardano-node for envrionment {{ENV}}"
mkdir -p "$STATEDIR"
if [[ "{{ENV}}" =~ mainnet$|preprod$|preview$ ]]; then
UNSTABLE=false
UNSTABLE_LIB=false
UNSTABLE_MITHRIL=false
USE_NODE_CONFIG_BP=false
else
UNSTABLE=true
UNSTABLE_LIB=true
UNSTABLE_MITHRIL=true
USE_NODE_CONFIG_BP=false
fi
# Set required entrypoint vars and run node in a new nohup background session
ENVIRONMENT="{{ENV}}" \
UNSTABLE="$UNSTABLE" \
UNSTABLE_LIB="$UNSTABLE_LIB" \
UNSTABLE_MITHRIL="$UNSTABLE_MITHRIL" \
USE_NODE_CONFIG_BP="$USE_NODE_CONFIG_BP" \
DATA_DIR="$STATEDIR" \
SOCKET_PATH="$STATEDIR/node-{{ENV}}.socket" \
nohup setsid nix run .#run-cardano-node &> "$STATEDIR/node-{{ENV}}.log" & echo $! > "$STATEDIR/node-{{ENV}}.pid" &
just set-default-cardano-env {{ENV}} "" "$PPID"
# Stop all nodes
stop-all:
#!/usr/bin/env bash
set -euo pipefail
for i in mainnet preprod preview private shelley-qa sanchonet demo; do
just stop-node $i
done
# Stop node for a specific env
stop-node ENV:
#!/usr/bin/env bash
set -euo pipefail
{{stateDir}}
if [ -f "$STATEDIR/node-{{ENV}}.pid" ]; then
echo "Stopping cardano-node for envrionment {{ENV}}"
kill $(< "$STATEDIR/node-{{ENV}}.pid") 2> /dev/null || true
rm -f "$STATEDIR/node-{{ENV}}.pid" "$STATEDIR/node-{{ENV}}.socket"
fi
# Diff against cardano-parts template
template-diff FILE *ARGS:
#!/usr/bin/env bash
set -euo pipefail
if ! [ -f {{FILE}} ]; then
FILE="<(echo '')"
else
FILE="{{FILE}}"
fi
if [ "{{templatePath}}" = "no-path-given" ]; then
SRC_FILE="<(curl -sL \"{{templateUrl}}/{{FILE}}\")"
SRC_NAME="{{templateUrl}}/{{FILE}}"
else
SRC_FILE="{{templatePath}}/{{FILE}}"
SRC_NAME="$SRC_FILE"
fi
eval "icdiff -L {{FILE}} -L \"$SRC_NAME\" {{ARGS}} $FILE $SRC_FILE"
# Patch against cardano-parts template
template-patch FILE:
#!/usr/bin/env bash
set -euo pipefail
if git status --porcelain "{{FILE}}" | grep -q "{{FILE}}"; then
echo "Git file {{FILE}} is dirty. Please revert or commit changes to clean state and try again."
exit 1
fi
if [ "{{templatePath}}" = "no-path-given" ]; then
SRC_FILE="<(curl -sL \"{{templateUrl}}/{{FILE}}\")"
else
SRC_FILE="{{templatePath}}/{{FILE}}"
fi
PATCH_FILE=$(eval "diff -Naru \"{{FILE}}\" $SRC_FILE || true")
patch "{{FILE}}" < <(echo "$PATCH_FILE")
git add -p "{{FILE}}"
# Run tofu for cluster or grafana workspace
tofu *ARGS:
#!/usr/bin/env bash
set -euo pipefail
IGREEN='\033[1;92m'
IRED='\033[1;91m'
NC='\033[0m'
SOPS=("sops" "--input-type" "binary" "--output-type" "binary" "--decrypt")
read -r -a ARGS <<< "{{ARGS}}"
if [[ ${ARGS[0]} =~ cluster|grafana ]]; then
WORKSPACE="${ARGS[0]}"
ARGS=("${ARGS[@]:1}")
else
WORKSPACE="cluster"
fi
unset VAR_FILE
if [ -s "secrets/tf/$WORKSPACE.tfvars" ]; then
VAR_FILE="secrets/tf/$WORKSPACE.tfvars"
fi
echo -e "Running tofu in the ${IGREEN}$WORKSPACE${NC} workspace..."
rm --force terraform.tf.json
nix build ".#opentofu.$WORKSPACE" --out-link terraform.tf.json
tofu init -reconfigure
tofu workspace select -or-create "$WORKSPACE"
tofu ${ARGS[@]} ${VAR_FILE:+-var-file=<("${SOPS[@]}" "$VAR_FILE")}
# Truncate a select chain after slot
truncate-chain ENV SLOT:
#!/usr/bin/env bash
set -euo pipefail
[ -n "${DEBUG:-}" ] && set -x
{{stateDir}}
if ! [[ "{{ENV}}" =~ mainnet$|preprod$|preview$|private$|sanchonet$|shelley-qa$ ]]; then
echo "Error: only node environments for mainnet, preprod, preview, private, sanchonet and shelley-qa are supported for truncate-chain recipe"
exit 1
fi
if ! [ -d "$STATEDIR/db-{{ENV}}/node" ]; then
echo "Error: no chain state appears to exist for {{ENV}} at: $STATEDIR/db-{{ENV}}/node"
exit 1
fi
echo "Truncating cardano-node chain for envrionment {{ENV}} to slot {{SLOT}}"
just stop-node {{ENV}}
mkdir -p "$STATEDIR"
SYNTH_ARGS=(
"--db" "$STATEDIR/db-{{ENV}}/node/"
"cardano"
"--config" "$STATEDIR/config/{{ENV}}/config.json"
)
TRUNC_ARGS=(
"${SYNTH_ARGS[@]}"
"--truncate-after-slot" "{{SLOT}}"
)
nix run .#job-gen-env-config &> /dev/null
if [[ "{{ENV}}" =~ mainnet$|preprod$|preview$ ]]; then
cp result/environments/config/{{ENV}}/* "$STATEDIR/config/{{ENV}}/"
chmod -R +w "$STATEDIR/config/{{ENV}}/"
db-truncater "${TRUNC_ARGS[@]}"
echo "Truncation finished."
echo "Analyzing to confirm truncation. This may take some time..."
db-analyser "${SYNTH_ARGS[@]}"
else
cp result/environments-pre/config/{{ENV}}/* "$STATEDIR/config/{{ENV}}/"
chmod -R +w "$STATEDIR/config/{{ENV}}/"
db-truncater-ng "${TRUNC_ARGS[@]}"
echo "Truncation finished."
echo "Analyzing to confirm truncation. This may take some time..."
db-analyser-ng "${SYNTH_ARGS[@]}"
fi
echo
echo "Note that db-truncater does not produce exact results."
echo "If the chain is still longer then you want, try reducing the truncation slot more."
# Update cluster ips from tofu
update-ips:
#!/usr/bin/env nu
nix build ".#opentofu.cluster" --out-link terraform.tf.json
tofu init -reconfigure
tofu workspace select -or-create cluster
echo
let nodeCount = nix eval .#nixosConfigurations --raw --apply 'let f = x: toString (builtins.length (builtins.attrNames x)); in f'
echo $"Processing ip information for ($nodeCount) nixos machine configurations..."
let eipRecords = (tofu show -json
| from json
| get values.root_module.resources
| where type == "aws_eip"
)
($eipRecords
| reduce --fold ["
let
all = {
"]
{|eip, all|
$all | append $"
($eip.name) = {
privateIpv4 = "($eip.values.private_ip)";
publicIpv4 = "($eip.values.public_ip)";
};"
}
| append "
};
in {
flake.nixosModules.ips = all;
flake.nixosModules.ip-module = {
name,
lib,
...
}: {
options.ips = {
privateIpv4 = lib.mkOption {
type = lib.types.str;
default = all.${name}.privateIpv4 or "";
};
publicIpv4 = lib.mkOption {
type = lib.types.str;
default = all.${name}.publicIpv4 or "";
};
};
};
}
"
| str join "\n"
| alejandra --quiet -
| save --force flake/nixosModules/ips-DONT-COMMIT.nix
)
# This is required for flake builds to find the nix module.
# The pre-push git hook will complain if this file has been committed accidently.
git add --intent-to-add flake/nixosModules/ips-DONT-COMMIT.nix
echo $"Ips were written for a machine count of: ($eipRecords | length)"
if $nodeCount != ($eipRecords | length | into string) {
echo
echo $"(ansi bg_red)WARNING:(ansi reset) There are ($nodeCount) nixos machine configurations but ($eipRecords | length) ip record sets were written."
echo
}
echo "Ips have been written to: flake/nixosModules/ips-DONT-COMMIT.nix"
echo "Obviously, don't commit this file."
# Generate example ip-module code
update-ips-example:
#!/usr/bin/env bash
set -euo pipefail
echo "The following code example shows the expected struct of the \"ips\" and \"ip-module\" nixosModules."
echo "This may be used as a template if a custom flake/nixosModules/ips-DONT-COMMIT.nix file"
echo "needs to be created and managed manually, for example, in a non-aws environment."
echo
cat <<"EOF"
let
all = {
machine-example-1 = {
privateIpv4 = "172.16.0.1";
publicIpv4 = "1.2.3.4";
};
};
in {
flake.nixosModules.ips = all;
flake.nixosModules.ip-module = {
name,
lib,
...
}: {
options.ips = {
privateIpv4 = lib.mkOption {
type = lib.types.str;
default = all.${name}.privateIpv4 or "";
};
publicIpv4 = lib.mkOption {
type = lib.types.str;
default = all.${name}.publicIpv4 or "";
};
};
};
}
EOF