# Smoke Test This lab walks you through a quick smoke test to make sure things are working. ## Test ``` kubectl run nginx --image=nginx --port=80 --replicas=3 ``` ``` deployment "nginx" created ``` ``` kubectl get pods -o wide ``` ``` NAME READY STATUS RESTARTS AGE IP NODE nginx-2032906785-ms8hw 1/1 Running 0 21s 10.200.2.2 worker2 nginx-2032906785-sokxz 1/1 Running 0 21s 10.200.1.2 worker1 nginx-2032906785-u8rzc 1/1 Running 0 21s 10.200.0.2 worker0 ``` ``` kubectl expose deployment nginx --type NodePort ``` ``` service "nginx" exposed ``` > Note that --type=LoadBalancer will not work because we did not configure a cloud provider when bootstrapping this cluster. Grab the `NodePort` that was setup for the nginx service: ``` NODE_PORT=$(kubectl get svc nginx --output=jsonpath='{range .spec.ports[0]}{.nodePort}') ``` ### Create the Node Port Firewall Rule #### GCP ``` gcloud compute firewall-rules create kubernetes-nginx-service \ --allow=tcp:${NODE_PORT} \ --network kubernetes ``` Grab the `EXTERNAL_IP` for one of the worker nodes: ``` NODE_PUBLIC_IP=$(gcloud compute instances describe worker0 \ --format 'value(networkInterfaces[0].accessConfigs[0].natIP)') ``` #### AWS ``` SECURITY_GROUP_ID=$(aws ec2 describe-security-groups \ --filters "Name=tag:Name,Values=kubernetes" | \ jq -r '.SecurityGroups[].GroupId') ``` ``` aws ec2 authorize-security-group-ingress \ --group-id ${SECURITY_GROUP_ID} \ --protocol tcp \ --port ${NODE_PORT} \ --cidr 0.0.0.0/0 ``` Grab the `EXTERNAL_IP` for one of the worker nodes: ``` NODE_PUBLIC_IP=$(aws ec2 describe-instances \ --filters "Name=tag:Name,Values=worker0" | \ jq -j '.Reservations[].Instances[].PublicIpAddress') ``` #### Azure ``` # Get the fqdn for the public worker ingestion load balancer NODE_PUBLIC_IP=$(azure network public-ip show \ --resource-group the-hard-way \ --name the-hard-way-workers \ --json | jq -r '.dnsSettings.fqdn') # Add NSG rule to enable traffic to node ports azure network nsg rule create \ --resource-group the-hard-way \ --nsg-name the-hard-way-nsg \ --name allow-internet-$NODE_PORT \ --protocol tcp \ --access allow \ --source-address-prefix Internet \ --destination-address-prefix 10.240.0.0/16 \ --destination-port-range $NODE_PORT \ --priority 110 \ --direction inbound # Create balancing rules NODE_PORT:NODE_PORT on the load balancer azure network lb probe create \ --resource-group the-hard-way \ --lb-name the-hard-way-lb \ --name nginx-app-health \ --interval 5 \ --port $NODE_PORT \ --protocol tcp azure network lb rule create \ --resource-group the-hard-way \ --lb-name the-hard-way-lb \ --name nginx-app \ --frontend-port $NODE_PORT \ --backend-port $NODE_PORT \ --frontend-ip-name the-hard-way-fe \ --backend-address-pool-name backend-pool\ --probe-name nginx-app-health ``` --- Test the nginx service using cURL: ``` curl http://${NODE_PUBLIC_IP}:${NODE_PORT} ``` ```
If you see this page, the nginx web server is successfully installed and working. Further configuration is required.
For online documentation and support please refer to
nginx.org.
Commercial support is available at
nginx.com.
Thank you for using nginx.
```