In order to access the ArgoCD web interface within our Kubernetes cluster, we’ll need to know the public IP address of our workshop environment.
Clicking the link below, will load the CloudFormation stacks page. Select bridgecrew-workshop
and select the Outputs tab, where we will see the public IP and ArgoCD URL.
The URL will be in the form https://<CLUSTER IP>:32443
click this link and you will be prompted with the ArgoCD login screen.
You can cat .bcworkshop/.argo-password
within our Cloud9 terminal to reveal the password for login. The username is admin
.
Now let’s log into the ArgoCD web interface to see the current state of our environment.
We can see that one of our environments, kustomizegoat-prod
, is healthy. It is “synced” between what we’ve declared in our IaC configuration and what is running on our cluster.
Let’s take a closer look at the visualizations for the healthy prod environment.
Select the “kustomizegoat-prod” box for a deeper look.
In this view, we can see a pretty simple deployment, a Kubernetes deployment, with a single ReplicaSet running two Pods. Then a sService loadbalancer linking to the pods.
Now let’s do the same with the kustomizegoat-dev environment to see what is going on there!
In the development environment, we see a different story. The service loadbalancer is created, but the Deployment has no children, ReplicaSet or Pods. We also see an error in the ArgoCD interface:
Clicking the “App Conditions: 1 Error” message, we see why this deployment wasn’t successful.
It looks like the deployment was blocked from running on the cluster because Bridgecrew’s Kubernetes admission controller identified a number of security issues. But thats not the nicest way to view the error.
Follow the complete details
link at the bottom of the error message for a clearer view.