The chapter 'Smoke Test' is updated.
parent
bfd7a1e69c
commit
2669322613
|
@ -9,15 +9,14 @@ In this section you will verify the ability to [encrypt secret data at rest](htt
|
||||||
Create a generic secret:
|
Create a generic secret:
|
||||||
|
|
||||||
```
|
```
|
||||||
kubectl create secret generic kubernetes-the-hard-way \
|
$ kubectl create secret generic kubernetes-the-hard-way \
|
||||||
--from-literal="mykey=mydata"
|
--from-literal="mykey=mydata"
|
||||||
```
|
```
|
||||||
|
|
||||||
Print a hexdump of the `kubernetes-the-hard-way` secret stored in etcd:
|
Print a hexdump of the `kubernetes-the-hard-way` secret stored in etcd:
|
||||||
|
|
||||||
```
|
```
|
||||||
gcloud compute ssh controller-0 \
|
$ ssh -i ~/.ssh/id_rsa-k8s.pub 10.240.0.21 "sudo ETCDCTL_API=3 etcdctl get \
|
||||||
--command "sudo ETCDCTL_API=3 etcdctl get \
|
|
||||||
--endpoints=https://127.0.0.1:2379 \
|
--endpoints=https://127.0.0.1:2379 \
|
||||||
--cacert=/etc/etcd/ca.pem \
|
--cacert=/etc/etcd/ca.pem \
|
||||||
--cert=/etc/etcd/kubernetes.pem \
|
--cert=/etc/etcd/kubernetes.pem \
|
||||||
|
@ -48,6 +47,7 @@ gcloud compute ssh controller-0 \
|
||||||
|
|
||||||
The etcd key should be prefixed with `k8s:enc:aescbc:v1:key1`, which indicates the `aescbc` provider was used to encrypt the data with the `key1` encryption key.
|
The etcd key should be prefixed with `k8s:enc:aescbc:v1:key1`, which indicates the `aescbc` provider was used to encrypt the data with the `key1` encryption key.
|
||||||
|
|
||||||
|
|
||||||
## Deployments
|
## Deployments
|
||||||
|
|
||||||
In this section you will verify the ability to create and manage [Deployments](https://kubernetes.io/docs/concepts/workloads/controllers/deployment/).
|
In this section you will verify the ability to create and manage [Deployments](https://kubernetes.io/docs/concepts/workloads/controllers/deployment/).
|
||||||
|
@ -55,13 +55,13 @@ In this section you will verify the ability to create and manage [Deployments](h
|
||||||
Create a deployment for the [nginx](https://nginx.org/en/) web server:
|
Create a deployment for the [nginx](https://nginx.org/en/) web server:
|
||||||
|
|
||||||
```
|
```
|
||||||
kubectl run nginx --image=nginx
|
$ kubectl run nginx --image=nginx
|
||||||
```
|
```
|
||||||
|
|
||||||
List the pod created by the `nginx` deployment:
|
List the pod created by the `nginx` deployment:
|
||||||
|
|
||||||
```
|
```
|
||||||
kubectl get pods -l run=nginx
|
$ kubectl get pods -l run=nginx
|
||||||
```
|
```
|
||||||
|
|
||||||
> output
|
> output
|
||||||
|
@ -78,13 +78,13 @@ In this section you will verify the ability to access applications remotely usin
|
||||||
Retrieve the full name of the `nginx` pod:
|
Retrieve the full name of the `nginx` pod:
|
||||||
|
|
||||||
```
|
```
|
||||||
POD_NAME=$(kubectl get pods -l run=nginx -o jsonpath="{.items[0].metadata.name}")
|
$ POD_NAME=$(kubectl get pods -l run=nginx -o jsonpath="{.items[0].metadata.name}")
|
||||||
```
|
```
|
||||||
|
|
||||||
Forward port `8080` on your local machine to port `80` of the `nginx` pod:
|
Forward port `8080` on your local machine to port `80` of the `nginx` pod:
|
||||||
|
|
||||||
```
|
```
|
||||||
kubectl port-forward $POD_NAME 8080:80
|
$ kubectl port-forward $POD_NAME 8080:80
|
||||||
```
|
```
|
||||||
|
|
||||||
> output
|
> output
|
||||||
|
@ -97,7 +97,7 @@ Forwarding from [::1]:8080 -> 80
|
||||||
In a new terminal make an HTTP request using the forwarding address:
|
In a new terminal make an HTTP request using the forwarding address:
|
||||||
|
|
||||||
```
|
```
|
||||||
curl --head http://127.0.0.1:8080
|
$ curl --head http://127.0.0.1:8080
|
||||||
```
|
```
|
||||||
|
|
||||||
> output
|
> output
|
||||||
|
@ -123,6 +123,7 @@ Handling connection for 8080
|
||||||
^C
|
^C
|
||||||
```
|
```
|
||||||
|
|
||||||
|
|
||||||
### Logs
|
### Logs
|
||||||
|
|
||||||
In this section you will verify the ability to [retrieve container logs](https://kubernetes.io/docs/concepts/cluster-administration/logging/).
|
In this section you will verify the ability to [retrieve container logs](https://kubernetes.io/docs/concepts/cluster-administration/logging/).
|
||||||
|
@ -130,7 +131,7 @@ In this section you will verify the ability to [retrieve container logs](https:/
|
||||||
Print the `nginx` pod logs:
|
Print the `nginx` pod logs:
|
||||||
|
|
||||||
```
|
```
|
||||||
kubectl logs $POD_NAME
|
$ kubectl logs $POD_NAME
|
||||||
```
|
```
|
||||||
|
|
||||||
> output
|
> output
|
||||||
|
@ -139,6 +140,7 @@ kubectl logs $POD_NAME
|
||||||
127.0.0.1 - - [30/Sep/2018:19:23:10 +0000] "HEAD / HTTP/1.1" 200 0 "-" "curl/7.58.0" "-"
|
127.0.0.1 - - [30/Sep/2018:19:23:10 +0000] "HEAD / HTTP/1.1" 200 0 "-" "curl/7.58.0" "-"
|
||||||
```
|
```
|
||||||
|
|
||||||
|
|
||||||
### Exec
|
### Exec
|
||||||
|
|
||||||
In this section you will verify the ability to [execute commands in a container](https://kubernetes.io/docs/tasks/debug-application-cluster/get-shell-running-container/#running-individual-commands-in-a-container).
|
In this section you will verify the ability to [execute commands in a container](https://kubernetes.io/docs/tasks/debug-application-cluster/get-shell-running-container/#running-individual-commands-in-a-container).
|
||||||
|
@ -146,15 +148,16 @@ In this section you will verify the ability to [execute commands in a container]
|
||||||
Print the nginx version by executing the `nginx -v` command in the `nginx` container:
|
Print the nginx version by executing the `nginx -v` command in the `nginx` container:
|
||||||
|
|
||||||
```
|
```
|
||||||
kubectl exec -ti $POD_NAME -- nginx -v
|
$ kubectl exec -ti $POD_NAME -- nginx -v
|
||||||
```
|
```
|
||||||
|
|
||||||
> output
|
> output
|
||||||
|
|
||||||
```
|
```
|
||||||
nginx version: nginx/1.15.4
|
$ nginx version: nginx/1.15.4
|
||||||
```
|
```
|
||||||
|
|
||||||
|
|
||||||
## Services
|
## Services
|
||||||
|
|
||||||
In this section you will verify the ability to expose applications using a [Service](https://kubernetes.io/docs/concepts/services-networking/service/).
|
In this section you will verify the ability to expose applications using a [Service](https://kubernetes.io/docs/concepts/services-networking/service/).
|
||||||
|
@ -162,7 +165,7 @@ In this section you will verify the ability to expose applications using a [Serv
|
||||||
Expose the `nginx` deployment using a [NodePort](https://kubernetes.io/docs/concepts/services-networking/service/#type-nodeport) service:
|
Expose the `nginx` deployment using a [NodePort](https://kubernetes.io/docs/concepts/services-networking/service/#type-nodeport) service:
|
||||||
|
|
||||||
```
|
```
|
||||||
kubectl expose deployment nginx --port 80 --type NodePort
|
$ kubectl expose deployment nginx --port 80 --type NodePort
|
||||||
```
|
```
|
||||||
|
|
||||||
> The LoadBalancer service type can not be used because your cluster is not configured with [cloud provider integration](https://kubernetes.io/docs/getting-started-guides/scratch/#cloud-provider). Setting up cloud provider integration is out of scope for this tutorial.
|
> The LoadBalancer service type can not be used because your cluster is not configured with [cloud provider integration](https://kubernetes.io/docs/getting-started-guides/scratch/#cloud-provider). Setting up cloud provider integration is out of scope for this tutorial.
|
||||||
|
@ -170,29 +173,20 @@ kubectl expose deployment nginx --port 80 --type NodePort
|
||||||
Retrieve the node port assigned to the `nginx` service:
|
Retrieve the node port assigned to the `nginx` service:
|
||||||
|
|
||||||
```
|
```
|
||||||
NODE_PORT=$(kubectl get svc nginx \
|
$ NODE_PORT=$(kubectl get svc nginx \
|
||||||
--output=jsonpath='{range .spec.ports[0]}{.nodePort}')
|
--output=jsonpath='{range .spec.ports[0]}{.nodePort}')
|
||||||
```
|
```
|
||||||
|
|
||||||
Create a firewall rule that allows remote access to the `nginx` node port:
|
Retrieve the IP address of a worker instance:
|
||||||
|
|
||||||
```
|
```
|
||||||
gcloud compute firewall-rules create kubernetes-the-hard-way-allow-nginx-service \
|
WORKER_IP=$ kubectl get nodes $(kubectl get pods -o wide | grep nginx | awk '{ print $7 }') -o wide | tail -1 | awk '{ print $6 }'
|
||||||
--allow=tcp:${NODE_PORT} \
|
|
||||||
--network kubernetes-the-hard-way
|
|
||||||
```
|
```
|
||||||
|
|
||||||
Retrieve the external IP address of a worker instance:
|
Make an HTTP request using the IP address and the `nginx` node port:
|
||||||
|
|
||||||
```
|
```
|
||||||
EXTERNAL_IP=$(gcloud compute instances describe worker-0 \
|
curl -I http://${WORKER_IP}:${NODE_PORT}
|
||||||
--format 'value(networkInterfaces[0].accessConfigs[0].natIP)')
|
|
||||||
```
|
|
||||||
|
|
||||||
Make an HTTP request using the external IP address and the `nginx` node port:
|
|
||||||
|
|
||||||
```
|
|
||||||
curl -I http://${EXTERNAL_IP}:${NODE_PORT}
|
|
||||||
```
|
```
|
||||||
|
|
||||||
> output
|
> output
|
||||||
|
@ -216,7 +210,7 @@ This section will verify the ability to run untrusted workloads using [gVisor](h
|
||||||
Create the `untrusted` pod:
|
Create the `untrusted` pod:
|
||||||
|
|
||||||
```
|
```
|
||||||
cat <<EOF | kubectl apply -f -
|
$ cat <<EOF | kubectl apply -f -
|
||||||
apiVersion: v1
|
apiVersion: v1
|
||||||
kind: Pod
|
kind: Pod
|
||||||
metadata:
|
metadata:
|
||||||
|
@ -237,8 +231,9 @@ In this section you will verify the `untrusted` pod is running under gVisor (run
|
||||||
Verify the `untrusted` pod is running:
|
Verify the `untrusted` pod is running:
|
||||||
|
|
||||||
```
|
```
|
||||||
kubectl get pods -o wide
|
$ kubectl get pods -o wide
|
||||||
```
|
```
|
||||||
|
|
||||||
```
|
```
|
||||||
NAME READY STATUS RESTARTS AGE IP NODE
|
NAME READY STATUS RESTARTS AGE IP NODE
|
||||||
busybox-68654f944b-djjjb 1/1 Running 0 5m 10.200.0.2 worker-0
|
busybox-68654f944b-djjjb 1/1 Running 0 5m 10.200.0.2 worker-0
|
||||||
|
@ -247,23 +242,25 @@ untrusted 1/1 Running 0 10s 10.200.0.3
|
||||||
```
|
```
|
||||||
|
|
||||||
|
|
||||||
Get the node name where the `untrusted` pod is running:
|
Get the node's IP address where the `untrusted` pod is running:
|
||||||
|
|
||||||
```
|
```
|
||||||
INSTANCE_NAME=$(kubectl get pod untrusted --output=jsonpath='{.spec.nodeName}')
|
$ INSTANCE_NAME=$(kubectl get pod untrusted --output=jsonpath='{.spec.nodeName}')
|
||||||
|
$ INSTANCE_IP_ADDRESS=$(kubectl get nodes ${INSTANCE_NAME} -o wide | tail -1 | awk '{ print $6 }')
|
||||||
```
|
```
|
||||||
|
|
||||||
SSH into the worker node:
|
SSH into the worker node:
|
||||||
|
|
||||||
```
|
```
|
||||||
gcloud compute ssh ${INSTANCE_NAME}
|
$ ssh -i ~/.ssh/id_rsa-k8s.pub ${INSTANCE_IP_ADDRESS}
|
||||||
```
|
```
|
||||||
|
|
||||||
List the containers running under gVisor:
|
List the containers running under gVisor:
|
||||||
|
|
||||||
```
|
```
|
||||||
sudo runsc --root /run/containerd/runsc/k8s.io list
|
$ sudo runsc --root /run/containerd/runsc/k8s.io list
|
||||||
```
|
```
|
||||||
|
|
||||||
```
|
```
|
||||||
I0930 19:27:13.255142 20832 x:0] ***************************
|
I0930 19:27:13.255142 20832 x:0] ***************************
|
||||||
I0930 19:27:13.255326 20832 x:0] Args: [runsc --root /run/containerd/runsc/k8s.io list]
|
I0930 19:27:13.255326 20832 x:0] Args: [runsc --root /run/containerd/runsc/k8s.io list]
|
||||||
|
@ -286,21 +283,21 @@ I0930 19:27:13.259733 20832 x:0] Exiting with status: 0
|
||||||
Get the ID of the `untrusted` pod:
|
Get the ID of the `untrusted` pod:
|
||||||
|
|
||||||
```
|
```
|
||||||
POD_ID=$(sudo crictl -r unix:///var/run/containerd/containerd.sock \
|
$ POD_ID=$(sudo crictl -r unix:///var/run/containerd/containerd.sock \
|
||||||
pods --name untrusted -q)
|
pods --name untrusted -q)
|
||||||
```
|
```
|
||||||
|
|
||||||
Get the ID of the `webserver` container running in the `untrusted` pod:
|
Get the ID of the `webserver` container running in the `untrusted` pod:
|
||||||
|
|
||||||
```
|
```
|
||||||
CONTAINER_ID=$(sudo crictl -r unix:///var/run/containerd/containerd.sock \
|
$ CONTAINER_ID=$(sudo crictl -r unix:///var/run/containerd/containerd.sock \
|
||||||
ps -p ${POD_ID} -q)
|
ps -p ${POD_ID} -q)
|
||||||
```
|
```
|
||||||
|
|
||||||
Use the gVisor `runsc` command to display the processes running inside the `webserver` container:
|
Use the gVisor `runsc` command to display the processes running inside the `webserver` container:
|
||||||
|
|
||||||
```
|
```
|
||||||
sudo runsc --root /run/containerd/runsc/k8s.io ps ${CONTAINER_ID}
|
$ sudo runsc --root /run/containerd/runsc/k8s.io ps ${CONTAINER_ID}
|
||||||
```
|
```
|
||||||
|
|
||||||
> output
|
> output
|
||||||
|
|
Loading…
Reference in New Issue