Trainer tools to create and prepare VMs for Docker workshops on AWS or Azure
Prerequisites
- Docker
- Docker Compose
- Parallel SSH (on a Mac:
brew install pssh) - the configuration scripts require this
And if you want to generate printable cards:
General Workflow
- fork/clone repo
- set required environment variables
- create your own setting file from
settings/example.yaml - if necessary, increase allowed open files:
ulimit -Sn 10000 - run
./workshopctlcommands to create instances, install docker, setup each users environment in node1, other management tasks - run
./workshopctl cardscommand to generate PDF for printing handouts of each users host IP's and login info
Clone/Fork the Repo, and Build the Tools Image
The Docker Compose file here is used to build a image with all the dependencies to run the ./workshopctl commands and optional tools. Each run of the script will check if you have those dependencies locally on your host, and will only use the container if you're missing a dependency.
$ git clone https://github.com/jpetazzo/orchestration-workshop.git
$ cd orchestration-workshop/prepare-vms
$ docker-compose build
Preparing to Run ./workshopctl
Required AWS Permissions/Info
- Initial assumptions are you're using a root account. If you'd like to use a IAM user, it will need
AmazonEC2FullAccessandIAMReadOnlyAccess. - Using a non-default VPC or Security Group isn't supported out of box yet, so you will have to customize
lib/commands.shif you want to change that. - These instances will assign the default VPC Security Group, which does not open any ports from Internet by default. So you'll need to add Inbound rules for
SSH | TCP | 22 | 0.0.0.0/0andCustom TCP Rule | TCP | 8000 - 8002 | 0.0.0.0/0, or run./workshopctl opensgwhich opens up all ports.
Required Environment Variables
AWS_ACCESS_KEY_IDAWS_SECRET_ACCESS_KEYAWS_DEFAULT_REGION
If you're not using AWS, set these to placeholder values:
export AWS_ACCESS_KEY_ID="foo"
export AWS_SECRET_ACCESS_KEY="foo"
export AWS_DEFAULT_REGION="foo"
If you don't have the aws CLI installed, you will get a warning that it's a missing dependency. If you're not using AWS you can ignore this.
Update/copy settings/example.yaml
Then pass settings/YOUR_WORKSHOP_NAME-settings.yaml as an argument to ./workshopctl deploy, ./workshopctl cards, etc.
./workshopctl cards 2016-09-28-00-33-bret settings/orchestration.yaml
./workshopctl Usage
workshopctl - the orchestration workshop swiss army knife
Commands:
ami Show the AMI that will be used for deployment
amis List Ubuntu AMIs in the current region
build Build the Docker image to run this program in a container
cards Generate ready-to-print cards for a batch of VMs
deploy Install Docker on a bunch of running VMs
ec2quotas Check our EC2 quotas (max instances)
help Show available commands
ids List the instance IDs belonging to a given tag or token
ips List the IP addresses of the VMs for a given tag or token
kube Setup kubernetes clusters with kubeadm (must be run AFTER deploy)
kubetest Check that all notes are reporting as Ready
list List available batches in the current region
opensg Open the default security group to ALL ingress traffic
pull_images Pre-pull a bunch of Docker images
retag Apply a new tag to a batch of VMs
start Start a batch of VMs
status List instance status for a given batch
stop Stop (terminate, shutdown, kill, remove, destroy...) instances
test Run tests (pre-flight checks) on a batch of VMs
wrap Run this program in a container
Summary of What ./workshopctl Does For You
- Used to manage bulk AWS instances for you without needing to use AWS cli or gui.
- Can manage multiple "tags" or groups of instances, which are tracked in
prepare-vms/tags/ - Can also create PDF/HTML for printing student info for instance IP's and login.
- The
./workshopctlscript can be executed directly. - It will run locally if all its dependencies are fulfilled; otherwise it will run in the Docker container you created with
docker-compose build(preparevms_prepare-vms). - During
startit will add your default local SSH key to all instances under theubuntuuser. - During
deployit will create thedockeruser with passwordtraining, which is printing on the cards for students. This can be configured with thedocker_user_passwordproperty in the settings file.
Example Steps to Launch a Batch of AWS Instances for a Workshop
- Run
./workshopctl start NCreatesNEC2 instances- Your local SSH key will be synced to instances under
ubuntuuser - AWS instances will be created and tagged based on date, and IP's stored in
prepare-vms/tags/
- Your local SSH key will be synced to instances under
- Run
./workshopctl deploy TAG settings/somefile.yamlto runlib/postprep.pyvia parallel-ssh- If it errors or times out, you should be able to rerun
- Requires good connection to run all the parallel SSH connections, up to 100 parallel (ProTip: create dedicated management instance in same AWS region where you run all these utils from)
- Run
./workshopctl pull_images TAGto pre-pull a bunch of Docker images to the instances - Run
./workshopctl cards TAG settings/somefile.yamlgenerates PDF/HTML files to print and cut and hand out to students - Have a great workshop
- Run
./workshopctl stop TAGto terminate instances.
Example Steps to Launch Azure Instances
- Install the Azure CLI and authenticate with a valid account
- Customize
azuredeploy.parameters.json- Required:
- Provide the SSH public key you plan to use for instance configuration
- Optional:
- Choose a name for the workshop (default is "workshop")
- Choose the number of instances (default is 3)
- Customize the desired instance size (default is Standard_D1_v2)
- Required:
- Launch instances with your chosen resource group name and your preferred region; the examples are "workshop" and "eastus":
az group create --name workshop --location eastus
az group deployment create --resource-group workshop --template-file azuredeploy.json --parameters @azuredeploy.parameters.json
The az group deployment create command can take several minutes and will only say - Running .. until it completes, unless you increase the verbosity with --verbose or --debug.
To display the IPs of the instances you've launched:
az vm list-ip-addresses --resource-group workshop --output table
If you want to put the IPs into prepare-vms/tags/<tag>/ips.txt for a tag of "myworkshop":
- If you haven't yet installed
jqand/or created your event's tags directory inprepare-vms:
brew install jq
mkdir -p tags/myworkshop
- And then generate the IP list:
az vm list-ip-addresses --resource-group workshop --output json | jq -r '.[].virtualMachine.network.publicIpAddresses[].ipAddress' > tags/myworkshop/ips.txt
After the workshop is over, remove the instances:
az group delete --resource-group workshop
Example Steps to Configure Instances from a non-AWS Source
- Launch instances via your preferred method. You'll need to get the instance IPs and be able to ssh into them.
- Set placeholder values for AWS environment variable settings.
- Choose a tag. It could be an event name, datestamp, etc. Ensure you have created a directory for your tag:
prepare-vms/tags/<tag>/ - If you have not already generated a file with the IPs to be configured:
- The file should be named
prepare-vms/tags/<tag>/ips.txt - Format is one IP per line, no other info needed.
- The file should be named
- Ensure the settings file is as desired (especially the number of nodes):
prepare-vms/settings/kube101.yaml - For a tag called
myworkshop, configure instances:workshopctl deploy myworkshop settings/kube101.yaml - Optionally, configure Kubernetes clusters of the size in the settings:
workshopctl kube myworkshop - Optionally, test your Kubernetes clusters. They may take a little time to become ready:
workshopctl kubetest myworkshop - Generate cards to print and hand out:
workshopctl cards myworkshop settings/kube101.yaml - Print the cards file:
prepare-vms/tags/myworkshop/ips.html
Other Tools
Deploying your SSH key to all the machines
- Make sure that you have SSH keys loaded (
ssh-add -l). - Source
rc. - Run
pcopykey.
Even More Details
Sync of SSH keys
When the start command is run, your local RSA SSH public key will be added to your AWS EC2 keychain.
To see which local key will be uploaded, run ssh-add -l | grep RSA.
Instance + tag creation
10 VMs will be started, with an automatically generated tag (timestamp + your username).
Your SSH key will be added to the authorized_keys of the ubuntu user.
Creation of ./$TAG/ directory and contents
Following the creation of the VMs, a text file will be created containing a list of their IPs.
This ips.txt file will be created in the $TAG/ directory and a symlink will be placed in the working directory of the script.
If you create new VMs, the symlinked file will be overwritten.
Deployment
Instances can be deployed manually using the deploy command:
$ ./workshopctl deploy TAG settings/somefile.yaml
The postprep.py file will be copied via parallel-ssh to all of the VMs and executed.
Pre-pull images
$ ./workshopctl pull_images TAG
Generate cards
$ ./workshopctl cards TAG settings/somefile.yaml
If you want to generate both HTML and PDF cards, install wkhtmltopdf; without that installed, only HTML cards will be generated.
If you don't have wkhtmltopdf installed, you will get a warning that it is a missing dependency. If you plan to just print the HTML cards, you can ignore this.
List tags
$ ./workshopctl list
List VMs
$ ./workshopctl list TAG
This will print a human-friendly list containing some information about each instance.
Stop and destroy VMs
$ ./workshopctl stop TAG
ToDo
- Don't write to bash history in system() in postprep
- compose, etc version inconsistent (int vs str)