From 528f68789801e04a6f123130583b0bd3ba6d5955 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?J=C3=A9r=C3=B4me=20Petazzoni?= Date: Thu, 30 Apr 2026 10:50:01 +0200 Subject: [PATCH] =?UTF-8?q?=E2=99=BB=EF=B8=8F=20Update=20konk=20machine=20?= =?UTF-8?q?type?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit vcluster deployment mode needs 2 volumes per cluster (one for the control plane, one for shpod), so we're switching to a smaller machine type since that's quickly becoming the limiting factor rather than CPU/RAM. --- prepare-labs/konk.sh | 9 +++++---- 1 file changed, 5 insertions(+), 4 deletions(-) diff --git a/prepare-labs/konk.sh b/prepare-labs/konk.sh index 751cedc0..dfa2c34f 100755 --- a/prepare-labs/konk.sh +++ b/prepare-labs/konk.sh @@ -4,20 +4,21 @@ # 500 MB RAM # 10% CPU # (See https://docs.google.com/document/d/1n0lwp6rQKQUIuo_A5LQ1dgCzrmjkDjmDtNj1Jn92UrI) -# PRO2-XS = 4 core, 16 gb # Note that we also need 2 volumes per vcluster (one for vcluster itself, one for shpod), # so we might hit the maximum number of volumes per node! -# (TODO: check what that limit is on Scaleway and Linode) +# (the limit on Scaleway is 15 volumes per node; 8 on Linode apparently? seems low, to check!) # # With vspod: # 800 MB RAM # 33% CPU # +# PRO2-XS = 4 core, 16 GB +# DEV1-XL = 4 core, 12 GB set -e KONKTAG=konk -PROVIDER=linode +PROVIDER=scaleway STUDENTS=2 case "$PROVIDER" in @@ -26,7 +27,7 @@ linode) export TF_VAR_location=fr-par ;; scaleway) - export TF_VAR_node_size=PRO2-XS + export TF_VAR_node_size=DEV1-XL # For tiny testing purposes, these are okay too: #export TF_VAR_node_size=PLAY2-NANO export TF_VAR_location=fr-par-2