diff --git a/docs/index.html b/docs/index.html index da776bd6..f137798f 100644 --- a/docs/index.html +++ b/docs/index.html @@ -1662,7 +1662,7 @@ Can you see how? - They are using a different kind of ID, reflecting the fact that they are SwarmKit objects instead of "classic" Docker Engine objects. -- They're *scope* is "swarm" instead of "local". +- Their *scope* is `swarm` instead of `local`. - They are using the overlay driver. @@ -1859,7 +1859,7 @@ Note: if the hash rate goes to zero and doesn't climb back up, try to `rm` and ` ## Checkpoint -- We've seen how to setup a Swarm +- We've seen how to set up a Swarm - We've used it to host our own registry @@ -1869,7 +1869,7 @@ Note: if the hash rate goes to zero and doesn't climb back up, try to `rm` and ` - We've deployed and scaled our application -Let's treat ourselves with a nice pat in the back! +Let's treat ourselves with a nice pat on the back! -- @@ -2499,7 +2499,7 @@ What we will do: - Manually send a few log entries using one-shot containers -- Setup our containers to send their logs to Logstash +- Set our containers up to send their logs to Logstash --- @@ -2780,7 +2780,7 @@ After ~15 seconds, you should see the log messages in Kibana. **This is not a "production-grade" setup.** -It is just an educational example. We did setup a single +It is just an educational example. We did set up a single ElasticSearch instance and a single Logstash instance. In a production setup, you need an ElasticSearch cluster @@ -2789,7 +2789,7 @@ need multiple Logstash instances. And if you want to withstand bursts of logs, you need some kind of message queue: -Redis if you're cheap, Kafka is you want to make sure +Redis if you're cheap, Kafka if you want to make sure that you don't drop messages on the floor. Good luck. --- @@ -3500,7 +3500,7 @@ the task (it will delete+re-create on all nodes). --- -## Setup Grafana +## Set up Grafana .exercise[