This job view page is being replaced by Spyglass soon. Check out the new job view.
ResultFAILURE
Tests 1 failed / 2 succeeded
Started2022-08-15 07:31
Elapsed11m25s
Revisionmaster

Test Failures


kubetest2 Test 2.00s

exit status 255
				from junit_runner.xml

Filter through log files | View test history on testgrid


Show 2 Passed Tests

Error lines from build-log.txt

ERROR: (gcloud.auth.activate-service-account) There was a problem refreshing your current auth tokens: ('invalid_grant: Invalid JWT Signature.', {'error': 'invalid_grant', 'error_description': 'Invalid JWT Signature.'})
Please run:

  $ gcloud auth login

to obtain new credentials.

... skipping 171 lines ...
I0815 07:32:44.771461    6135 http.go:37] curl https://storage.googleapis.com/kops-ci/bin/latest-ci-updown-green.txt
I0815 07:32:44.773819    6135 http.go:37] curl https://storage.googleapis.com/kops-ci/bin/1.25.0-alpha.3+v1.25.0-alpha.2-85-g429ebecdca/linux/amd64/kops
I0815 07:32:45.442916    6135 local.go:42] ⚙️ ssh-keygen -t ed25519 -N  -q -f /tmp/kops/e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io/id_ed25519
I0815 07:32:45.449726    6135 up.go:44] Cleaning up any leaked resources from previous cluster
I0815 07:32:45.449832    6135 dumplogs.go:45] /home/prow/go/src/k8s.io/kops/_rundir/369eae17-1c6c-11ed-9820-1e463d3c6692/kops toolbox dump --name e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io --dir /logs/artifacts --private-key /tmp/kops/e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io/id_ed25519 --ssh-user ubuntu
I0815 07:32:45.449854    6135 local.go:42] ⚙️ /home/prow/go/src/k8s.io/kops/_rundir/369eae17-1c6c-11ed-9820-1e463d3c6692/kops toolbox dump --name e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io --dir /logs/artifacts --private-key /tmp/kops/e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io/id_ed25519 --ssh-user ubuntu
W0815 07:32:45.904914    6135 down.go:34] Dumping cluster logs at the start of Down() failed: exit status 1
I0815 07:32:45.904969    6135 down.go:48] /home/prow/go/src/k8s.io/kops/_rundir/369eae17-1c6c-11ed-9820-1e463d3c6692/kops delete cluster --name e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io --yes
I0815 07:32:45.904982    6135 local.go:42] ⚙️ /home/prow/go/src/k8s.io/kops/_rundir/369eae17-1c6c-11ed-9820-1e463d3c6692/kops delete cluster --name e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io --yes
I0815 07:32:45.939112    6166 featureflag.go:164] FeatureFlag "SpecOverrideFlag"=true
Error: error reading cluster configuration: Cluster.kops.k8s.io "e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io" not found
I0815 07:32:46.368836    6135 http.go:37] curl http://metadata.google.internal/computeMetadata/v1/instance/network-interfaces/0/access-configs/0/external-ip
2022/08/15 07:32:46 failed to get external ip from metadata service: http://metadata.google.internal/computeMetadata/v1/instance/network-interfaces/0/access-configs/0/external-ip returned 404
I0815 07:32:46.383981    6135 http.go:37] curl https://ip.jsb.workers.dev
I0815 07:32:46.513829    6135 up.go:159] /home/prow/go/src/k8s.io/kops/_rundir/369eae17-1c6c-11ed-9820-1e463d3c6692/kops create cluster --name e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io --cloud aws --kubernetes-version https://storage.googleapis.com/kubernetes-release/release/v1.24.3 --ssh-public-key /tmp/kops/e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io/id_ed25519.pub --override cluster.spec.nodePortAccess=0.0.0.0/0 --image=099720109477/ubuntu/images/hvm-ssd/ubuntu-focal-20.04-amd64-server-20220810 --channel=alpha --networking=kube-router --container-runtime=containerd --node-size=t3.large --discovery-store=s3://k8s-kops-prow/discovery --admin-access 34.69.195.104/32 --master-count 1 --master-volume-size 48 --node-count 4 --node-volume-size 48 --zones sa-east-1a --master-size c5.large
I0815 07:32:46.513865    6135 local.go:42] ⚙️ /home/prow/go/src/k8s.io/kops/_rundir/369eae17-1c6c-11ed-9820-1e463d3c6692/kops create cluster --name e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io --cloud aws --kubernetes-version https://storage.googleapis.com/kubernetes-release/release/v1.24.3 --ssh-public-key /tmp/kops/e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io/id_ed25519.pub --override cluster.spec.nodePortAccess=0.0.0.0/0 --image=099720109477/ubuntu/images/hvm-ssd/ubuntu-focal-20.04-amd64-server-20220810 --channel=alpha --networking=kube-router --container-runtime=containerd --node-size=t3.large --discovery-store=s3://k8s-kops-prow/discovery --admin-access 34.69.195.104/32 --master-count 1 --master-volume-size 48 --node-count 4 --node-volume-size 48 --zones sa-east-1a --master-size c5.large
I0815 07:32:46.545565    6174 featureflag.go:164] FeatureFlag "SpecOverrideFlag"=true
I0815 07:32:46.565068    6174 create_cluster.go:862] Using SSH public key: /tmp/kops/e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io/id_ed25519.pub
I0815 07:32:47.056138    6174 new_cluster.go:1168]  Cloud Provider ID = aws
... skipping 590 lines ...

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 07:33:36.135728    6211 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-sa-east-1a	Master	c5.large	1	1	sa-east-1a
nodes-sa-east-1a	Node	t3.large	4	4	sa-east-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 07:33:46.188024    6211 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-sa-east-1a	Master	c5.large	1	1	sa-east-1a
nodes-sa-east-1a	Node	t3.large	4	4	sa-east-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 07:33:56.223400    6211 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-sa-east-1a	Master	c5.large	1	1	sa-east-1a
nodes-sa-east-1a	Node	t3.large	4	4	sa-east-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 07:34:06.276227    6211 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-sa-east-1a	Master	c5.large	1	1	sa-east-1a
nodes-sa-east-1a	Node	t3.large	4	4	sa-east-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 07:34:16.311557    6211 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-sa-east-1a	Master	c5.large	1	1	sa-east-1a
nodes-sa-east-1a	Node	t3.large	4	4	sa-east-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 07:34:26.344470    6211 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-sa-east-1a	Master	c5.large	1	1	sa-east-1a
nodes-sa-east-1a	Node	t3.large	4	4	sa-east-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 07:34:36.379488    6211 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-sa-east-1a	Master	c5.large	1	1	sa-east-1a
nodes-sa-east-1a	Node	t3.large	4	4	sa-east-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 07:34:46.415527    6211 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-sa-east-1a	Master	c5.large	1	1	sa-east-1a
nodes-sa-east-1a	Node	t3.large	4	4	sa-east-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 07:34:56.457877    6211 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-sa-east-1a	Master	c5.large	1	1	sa-east-1a
nodes-sa-east-1a	Node	t3.large	4	4	sa-east-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 07:35:06.507374    6211 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-sa-east-1a	Master	c5.large	1	1	sa-east-1a
nodes-sa-east-1a	Node	t3.large	4	4	sa-east-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 07:35:16.540459    6211 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-sa-east-1a	Master	c5.large	1	1	sa-east-1a
nodes-sa-east-1a	Node	t3.large	4	4	sa-east-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 07:35:26.573628    6211 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-sa-east-1a	Master	c5.large	1	1	sa-east-1a
nodes-sa-east-1a	Node	t3.large	4	4	sa-east-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 07:35:36.608410    6211 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-sa-east-1a	Master	c5.large	1	1	sa-east-1a
nodes-sa-east-1a	Node	t3.large	4	4	sa-east-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 07:35:46.649662    6211 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-sa-east-1a	Master	c5.large	1	1	sa-east-1a
nodes-sa-east-1a	Node	t3.large	4	4	sa-east-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 07:35:56.699564    6211 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-sa-east-1a	Master	c5.large	1	1	sa-east-1a
nodes-sa-east-1a	Node	t3.large	4	4	sa-east-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 07:36:06.746970    6211 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-sa-east-1a	Master	c5.large	1	1	sa-east-1a
nodes-sa-east-1a	Node	t3.large	4	4	sa-east-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 07:36:16.790712    6211 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-sa-east-1a	Master	c5.large	1	1	sa-east-1a
nodes-sa-east-1a	Node	t3.large	4	4	sa-east-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 07:36:26.827091    6211 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-sa-east-1a	Master	c5.large	1	1	sa-east-1a
nodes-sa-east-1a	Node	t3.large	4	4	sa-east-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 07:36:36.865552    6211 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-sa-east-1a	Master	c5.large	1	1	sa-east-1a
nodes-sa-east-1a	Node	t3.large	4	4	sa-east-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 07:36:46.916938    6211 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-sa-east-1a	Master	c5.large	1	1	sa-east-1a
nodes-sa-east-1a	Node	t3.large	4	4	sa-east-1a

... skipping 14 lines ...
Pod	kube-system/ebs-csi-controller-74594ff879-zdm78	system-cluster-critical pod "ebs-csi-controller-74594ff879-zdm78" is pending
Pod	kube-system/ebs-csi-node-cz26d			system-node-critical pod "ebs-csi-node-cz26d" is pending
Pod	kube-system/ebs-csi-node-jp6dn			system-node-critical pod "ebs-csi-node-jp6dn" is pending
Pod	kube-system/kube-router-bgpfr			system-node-critical pod "kube-router-bgpfr" is pending
Pod	kube-system/kube-router-vsq4c			system-node-critical pod "kube-router-vsq4c" is pending

Validation Failed
W0815 07:37:00.625797    6211 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-sa-east-1a	Master	c5.large	1	1	sa-east-1a
nodes-sa-east-1a	Node	t3.large	4	4	sa-east-1a

... skipping 19 lines ...
Pod	kube-system/ebs-csi-node-cz26d			system-node-critical pod "ebs-csi-node-cz26d" is pending
Pod	kube-system/ebs-csi-node-jp6dn			system-node-critical pod "ebs-csi-node-jp6dn" is pending
Pod	kube-system/ebs-csi-node-p5rmm			system-node-critical pod "ebs-csi-node-p5rmm" is pending
Pod	kube-system/kube-router-7jb4h			system-node-critical pod "kube-router-7jb4h" is pending
Pod	kube-system/kube-router-ndr7b			system-node-critical pod "kube-router-ndr7b" is pending

Validation Failed
W0815 07:37:13.237527    6211 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-sa-east-1a	Master	c5.large	1	1	sa-east-1a
nodes-sa-east-1a	Node	t3.large	4	4	sa-east-1a

... skipping 12 lines ...
Pod	kube-system/ebs-csi-controller-74594ff879-fl8jl	system-cluster-critical pod "ebs-csi-controller-74594ff879-fl8jl" is pending
Pod	kube-system/ebs-csi-controller-74594ff879-zdm78	system-cluster-critical pod "ebs-csi-controller-74594ff879-zdm78" is pending
Pod	kube-system/ebs-csi-node-525p4			system-node-critical pod "ebs-csi-node-525p4" is pending
Pod	kube-system/ebs-csi-node-cz26d			system-node-critical pod "ebs-csi-node-cz26d" is pending
Pod	kube-system/ebs-csi-node-p5rmm			system-node-critical pod "ebs-csi-node-p5rmm" is pending

Validation Failed
W0815 07:37:25.783429    6211 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-sa-east-1a	Master	c5.large	1	1	sa-east-1a
nodes-sa-east-1a	Node	t3.large	4	4	sa-east-1a

... skipping 8 lines ...
VALIDATION ERRORS
KIND	NAME						MESSAGE
Pod	kube-system/coredns-5c44b6cf7d-g4lq5		system-cluster-critical pod "coredns-5c44b6cf7d-g4lq5" is not ready (coredns)
Pod	kube-system/ebs-csi-controller-74594ff879-fl8jl	system-cluster-critical pod "ebs-csi-controller-74594ff879-fl8jl" is pending
Pod	kube-system/ebs-csi-controller-74594ff879-zdm78	system-cluster-critical pod "ebs-csi-controller-74594ff879-zdm78" is pending

Validation Failed
W0815 07:37:38.341380    6211 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-sa-east-1a	Master	c5.large	1	1	sa-east-1a
nodes-sa-east-1a	Node	t3.large	4	4	sa-east-1a

... skipping 6 lines ...
i-0d9edbe3ac3099734	node	True

VALIDATION ERRORS
KIND	NAME						MESSAGE
Pod	kube-system/ebs-csi-controller-74594ff879-fl8jl	system-cluster-critical pod "ebs-csi-controller-74594ff879-fl8jl" is not ready (ebs-plugin)

Validation Failed
W0815 07:37:50.818406    6211 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-sa-east-1a	Master	c5.large	1	1	sa-east-1a
nodes-sa-east-1a	Node	t3.large	4	4	sa-east-1a

... skipping 6 lines ...
i-0d9edbe3ac3099734	node	True

VALIDATION ERRORS
KIND	NAME						MESSAGE
Pod	kube-system/ebs-csi-controller-74594ff879-fl8jl	system-cluster-critical pod "ebs-csi-controller-74594ff879-fl8jl" is not ready (ebs-plugin)

Validation Failed
W0815 07:38:03.286045    6211 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-sa-east-1a	Master	c5.large	1	1	sa-east-1a
nodes-sa-east-1a	Node	t3.large	4	4	sa-east-1a

... skipping 141 lines ...
i-0d9edbe3ac3099734	node	True

Your cluster e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io is ready
I0815 07:40:08.355173    6135 up.go:105] cluster reported as up
I0815 07:40:08.355237    6135 local.go:42] ⚙️ /home/prow/go/bin/kubetest2-tester-kops --ginkgo-args=--debug --test-args=-test.timeout=60m -num-nodes=0 --test-package-marker=stable.txt --parallel=25
I0815 07:40:08.374702    6221 featureflag.go:164] FeatureFlag "SpecOverrideFlag"=true
F0815 07:40:10.349706    6221 tester.go:482] failed to run ginkgo tester: failed to get kubectl package from published releases: failed to get latest release name: exit status 1
I0815 07:40:10.353003    6135 dumplogs.go:45] /home/prow/go/src/k8s.io/kops/_rundir/369eae17-1c6c-11ed-9820-1e463d3c6692/kops toolbox dump --name e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io --dir /logs/artifacts --private-key /tmp/kops/e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io/id_ed25519 --ssh-user ubuntu
I0815 07:40:10.353091    6135 local.go:42] ⚙️ /home/prow/go/src/k8s.io/kops/_rundir/369eae17-1c6c-11ed-9820-1e463d3c6692/kops toolbox dump --name e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io --dir /logs/artifacts --private-key /tmp/kops/e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io/id_ed25519 --ssh-user ubuntu
I0815 07:41:00.437373    6135 dumplogs.go:78] /home/prow/go/src/k8s.io/kops/_rundir/369eae17-1c6c-11ed-9820-1e463d3c6692/kops get cluster --name e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io -o yaml
I0815 07:41:00.437473    6135 local.go:42] ⚙️ /home/prow/go/src/k8s.io/kops/_rundir/369eae17-1c6c-11ed-9820-1e463d3c6692/kops get cluster --name e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io -o yaml
I0815 07:41:00.962407    6135 dumplogs.go:78] /home/prow/go/src/k8s.io/kops/_rundir/369eae17-1c6c-11ed-9820-1e463d3c6692/kops get instancegroups --name e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io -o yaml
I0815 07:41:00.962449    6135 local.go:42] ⚙️ /home/prow/go/src/k8s.io/kops/_rundir/369eae17-1c6c-11ed-9820-1e463d3c6692/kops get instancegroups --name e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io -o yaml
... skipping 273 lines ...
route-table:rtb-0f948e4bdebcee67f	ok
vpc:vpc-092e44211e393e543	ok
dhcp-options:dopt-0863aae6266234422	ok
Deleted kubectl config for e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io

Deleted cluster: "e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io"
Error: exit status 255
+ EXIT_VALUE=1
+ set +o xtrace