This job view page is being replaced by Spyglass soon. Check out the new job view.
ResultFAILURE
Tests 1 failed / 2 succeeded
Started2022-08-17 19:24
Elapsed11m35s
Revisionmaster

Test Failures


kubetest2 Test 1.83s

exit status 255
				from junit_runner.xml

Filter through log files | View test history on testgrid


Show 2 Passed Tests

Error lines from build-log.txt

ERROR: (gcloud.auth.activate-service-account) There was a problem refreshing your current auth tokens: ('invalid_grant: Invalid JWT Signature.', {'error': 'invalid_grant', 'error_description': 'Invalid JWT Signature.'})
Please run:

  $ gcloud auth login

to obtain new credentials.

... skipping 171 lines ...
I0817 19:25:16.276347    6157 http.go:37] curl https://storage.googleapis.com/kops-ci/bin/latest-ci-updown-green.txt
I0817 19:25:16.280296    6157 http.go:37] curl https://storage.googleapis.com/kops-ci/bin/1.25.0-alpha.3+v1.25.0-alpha.2-99-g7035526204/linux/amd64/kops
I0817 19:25:16.967511    6157 local.go:42] ⚙️ ssh-keygen -t ed25519 -N  -q -f /tmp/kops/e2e-e2e-kops-aws-k8s-1-22.test-cncf-aws.k8s.io/id_ed25519
I0817 19:25:16.974332    6157 up.go:44] Cleaning up any leaked resources from previous cluster
I0817 19:25:16.974429    6157 dumplogs.go:45] /home/prow/go/src/k8s.io/kops/_rundir/171e15a8-1e62-11ed-a9bf-0eb48dfbe9bd/kops toolbox dump --name e2e-e2e-kops-aws-k8s-1-22.test-cncf-aws.k8s.io --dir /logs/artifacts --private-key /tmp/kops/e2e-e2e-kops-aws-k8s-1-22.test-cncf-aws.k8s.io/id_ed25519 --ssh-user ubuntu
I0817 19:25:16.974458    6157 local.go:42] ⚙️ /home/prow/go/src/k8s.io/kops/_rundir/171e15a8-1e62-11ed-a9bf-0eb48dfbe9bd/kops toolbox dump --name e2e-e2e-kops-aws-k8s-1-22.test-cncf-aws.k8s.io --dir /logs/artifacts --private-key /tmp/kops/e2e-e2e-kops-aws-k8s-1-22.test-cncf-aws.k8s.io/id_ed25519 --ssh-user ubuntu
W0817 19:25:17.501874    6157 down.go:34] Dumping cluster logs at the start of Down() failed: exit status 1
I0817 19:25:17.501937    6157 down.go:48] /home/prow/go/src/k8s.io/kops/_rundir/171e15a8-1e62-11ed-a9bf-0eb48dfbe9bd/kops delete cluster --name e2e-e2e-kops-aws-k8s-1-22.test-cncf-aws.k8s.io --yes
I0817 19:25:17.501951    6157 local.go:42] ⚙️ /home/prow/go/src/k8s.io/kops/_rundir/171e15a8-1e62-11ed-a9bf-0eb48dfbe9bd/kops delete cluster --name e2e-e2e-kops-aws-k8s-1-22.test-cncf-aws.k8s.io --yes
I0817 19:25:17.534114    6193 featureflag.go:164] FeatureFlag "SpecOverrideFlag"=true
Error: error reading cluster configuration: Cluster.kops.k8s.io "e2e-e2e-kops-aws-k8s-1-22.test-cncf-aws.k8s.io" not found
I0817 19:25:17.972927    6157 http.go:37] curl http://metadata.google.internal/computeMetadata/v1/instance/network-interfaces/0/access-configs/0/external-ip
2022/08/17 19:25:17 failed to get external ip from metadata service: http://metadata.google.internal/computeMetadata/v1/instance/network-interfaces/0/access-configs/0/external-ip returned 404
I0817 19:25:17.990189    6157 http.go:37] curl https://ip.jsb.workers.dev
I0817 19:25:18.140314    6157 up.go:159] /home/prow/go/src/k8s.io/kops/_rundir/171e15a8-1e62-11ed-a9bf-0eb48dfbe9bd/kops create cluster --name e2e-e2e-kops-aws-k8s-1-22.test-cncf-aws.k8s.io --cloud aws --kubernetes-version https://storage.googleapis.com/kubernetes-release/release/v1.22.12 --ssh-public-key /tmp/kops/e2e-e2e-kops-aws-k8s-1-22.test-cncf-aws.k8s.io/id_ed25519.pub --override cluster.spec.nodePortAccess=0.0.0.0/0 --image=099720109477/ubuntu/images/hvm-ssd/ubuntu-focal-20.04-amd64-server-20220810 --channel=alpha --networking=calico --container-runtime=containerd --discovery-store=s3://k8s-kops-prow/discovery --admin-access 35.224.49.127/32 --master-count 1 --master-volume-size 48 --node-count 4 --node-volume-size 48 --zones us-west-1a --master-size c5.large
I0817 19:25:18.140361    6157 local.go:42] ⚙️ /home/prow/go/src/k8s.io/kops/_rundir/171e15a8-1e62-11ed-a9bf-0eb48dfbe9bd/kops create cluster --name e2e-e2e-kops-aws-k8s-1-22.test-cncf-aws.k8s.io --cloud aws --kubernetes-version https://storage.googleapis.com/kubernetes-release/release/v1.22.12 --ssh-public-key /tmp/kops/e2e-e2e-kops-aws-k8s-1-22.test-cncf-aws.k8s.io/id_ed25519.pub --override cluster.spec.nodePortAccess=0.0.0.0/0 --image=099720109477/ubuntu/images/hvm-ssd/ubuntu-focal-20.04-amd64-server-20220810 --channel=alpha --networking=calico --container-runtime=containerd --discovery-store=s3://k8s-kops-prow/discovery --admin-access 35.224.49.127/32 --master-count 1 --master-volume-size 48 --node-count 4 --node-volume-size 48 --zones us-west-1a --master-size c5.large
I0817 19:25:18.175666    6201 featureflag.go:164] FeatureFlag "SpecOverrideFlag"=true
I0817 19:25:18.193619    6201 create_cluster.go:862] Using SSH public key: /tmp/kops/e2e-e2e-kops-aws-k8s-1-22.test-cncf-aws.k8s.io/id_ed25519.pub
I0817 19:25:18.686827    6201 new_cluster.go:1168]  Cloud Provider ID = aws
... skipping 557 lines ...

I0817 19:25:58.758412    6157 up.go:243] /home/prow/go/src/k8s.io/kops/_rundir/171e15a8-1e62-11ed-a9bf-0eb48dfbe9bd/kops validate cluster --name e2e-e2e-kops-aws-k8s-1-22.test-cncf-aws.k8s.io --count 10 --wait 15m0s
I0817 19:25:58.758501    6157 local.go:42] ⚙️ /home/prow/go/src/k8s.io/kops/_rundir/171e15a8-1e62-11ed-a9bf-0eb48dfbe9bd/kops validate cluster --name e2e-e2e-kops-aws-k8s-1-22.test-cncf-aws.k8s.io --count 10 --wait 15m0s
I0817 19:25:58.791663    6241 featureflag.go:164] FeatureFlag "SpecOverrideFlag"=true
Validating cluster e2e-e2e-kops-aws-k8s-1-22.test-cncf-aws.k8s.io

W0817 19:25:59.839045    6241 validate_cluster.go:184] (will retry): unexpected error during validation: unable to resolve Kubernetes cluster API URL dns: lookup api.e2e-e2e-kops-aws-k8s-1-22.test-cncf-aws.k8s.io on 10.63.240.10:53: no such host
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.medium	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0817 19:26:09.877914    6241 validate_cluster.go:232] (will retry): cluster not yet healthy
W0817 19:26:19.914807    6241 validate_cluster.go:184] (will retry): unexpected error during validation: unable to resolve Kubernetes cluster API URL dns: lookup api.e2e-e2e-kops-aws-k8s-1-22.test-cncf-aws.k8s.io on 10.63.240.10:53: no such host
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.medium	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0817 19:26:29.954506    6241 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.medium	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0817 19:26:39.992826    6241 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.medium	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0817 19:26:50.049947    6241 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.medium	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0817 19:27:00.084803    6241 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.medium	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0817 19:27:10.120561    6241 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.medium	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0817 19:27:20.164538    6241 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.medium	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0817 19:27:30.208907    6241 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.medium	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0817 19:27:40.244802    6241 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.medium	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0817 19:27:50.293682    6241 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.medium	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0817 19:28:00.330456    6241 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.medium	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0817 19:28:10.379552    6241 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.medium	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0817 19:28:20.440002    6241 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.medium	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0817 19:28:30.483375    6241 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.medium	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0817 19:28:40.519656    6241 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.medium	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0817 19:28:50.554950    6241 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.medium	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0817 19:29:00.595049    6241 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.medium	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0817 19:29:10.649637    6241 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.medium	4	4	us-west-1a

... skipping 15 lines ...
Pod	kube-system/coredns-autoscaler-6658b4bf85-8s28r	system-cluster-critical pod "coredns-autoscaler-6658b4bf85-8s28r" is pending
Pod	kube-system/ebs-csi-controller-65ddb8876b-dnxpb	system-cluster-critical pod "ebs-csi-controller-65ddb8876b-dnxpb" is pending
Pod	kube-system/ebs-csi-controller-65ddb8876b-nzzrp	system-cluster-critical pod "ebs-csi-controller-65ddb8876b-nzzrp" is pending
Pod	kube-system/ebs-csi-node-48dl4			system-node-critical pod "ebs-csi-node-48dl4" is pending
Pod	kube-system/ebs-csi-node-hc8gc			system-node-critical pod "ebs-csi-node-hc8gc" is pending

Validation Failed
W0817 19:29:22.357174    6241 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.medium	4	4	us-west-1a

... skipping 18 lines ...
Pod	kube-system/ebs-csi-controller-65ddb8876b-dnxpb	system-cluster-critical pod "ebs-csi-controller-65ddb8876b-dnxpb" is pending
Pod	kube-system/ebs-csi-controller-65ddb8876b-nzzrp	system-cluster-critical pod "ebs-csi-controller-65ddb8876b-nzzrp" is pending
Pod	kube-system/ebs-csi-node-48dl4			system-node-critical pod "ebs-csi-node-48dl4" is pending
Pod	kube-system/ebs-csi-node-hc8gc			system-node-critical pod "ebs-csi-node-hc8gc" is pending
Pod	kube-system/ebs-csi-node-trgn8			system-node-critical pod "ebs-csi-node-trgn8" is pending

Validation Failed
W0817 19:29:33.563668    6241 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.medium	4	4	us-west-1a

... skipping 17 lines ...
Pod	kube-system/ebs-csi-controller-65ddb8876b-nzzrp	system-cluster-critical pod "ebs-csi-controller-65ddb8876b-nzzrp" is pending
Pod	kube-system/ebs-csi-node-48dl4			system-node-critical pod "ebs-csi-node-48dl4" is pending
Pod	kube-system/ebs-csi-node-hc8gc			system-node-critical pod "ebs-csi-node-hc8gc" is pending
Pod	kube-system/ebs-csi-node-trgn8			system-node-critical pod "ebs-csi-node-trgn8" is pending
Pod	kube-system/ebs-csi-node-x8jqr			system-node-critical pod "ebs-csi-node-x8jqr" is pending

Validation Failed
W0817 19:29:44.849330    6241 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.medium	4	4	us-west-1a

... skipping 11 lines ...
Pod	kube-system/calico-node-6gf86			system-node-critical pod "calico-node-6gf86" is pending
Pod	kube-system/calico-node-mwncl			system-node-critical pod "calico-node-mwncl" is not ready (calico-node)
Pod	kube-system/ebs-csi-controller-65ddb8876b-dnxpb	system-cluster-critical pod "ebs-csi-controller-65ddb8876b-dnxpb" is pending
Pod	kube-system/ebs-csi-node-trgn8			system-node-critical pod "ebs-csi-node-trgn8" is pending
Pod	kube-system/ebs-csi-node-x8jqr			system-node-critical pod "ebs-csi-node-x8jqr" is pending

Validation Failed
W0817 19:29:56.210494    6241 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.medium	4	4	us-west-1a

... skipping 9 lines ...
KIND	NAME									MESSAGE
Pod	kube-system/calico-node-6gf86						system-node-critical pod "calico-node-6gf86" is not ready (calico-node)
Pod	kube-system/calico-node-mwncl						system-node-critical pod "calico-node-mwncl" is not ready (calico-node)
Pod	kube-system/ebs-csi-node-x8jqr						system-node-critical pod "ebs-csi-node-x8jqr" is pending
Pod	kube-system/kube-proxy-ip-172-20-37-91.us-west-1.compute.internal	system-node-critical pod "kube-proxy-ip-172-20-37-91.us-west-1.compute.internal" is pending

Validation Failed
W0817 19:30:07.475075    6241 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.medium	4	4	us-west-1a

... skipping 141 lines ...
ip-172-20-61-193.us-west-1.compute.internal	master	True

Your cluster e2e-e2e-kops-aws-k8s-1-22.test-cncf-aws.k8s.io is ready
I0817 19:32:00.080979    6157 up.go:105] cluster reported as up
I0817 19:32:00.081107    6157 local.go:42] ⚙️ /home/prow/go/bin/kubetest2-tester-kops --ginkgo-args=--debug --test-args=-test.timeout=60m -num-nodes=0 --test-package-marker=stable-1.22.txt --parallel=25
I0817 19:32:00.103202    6251 featureflag.go:164] FeatureFlag "SpecOverrideFlag"=true
F0817 19:32:01.905020    6251 tester.go:482] failed to run ginkgo tester: failed to get kubectl package from published releases: failed to get latest release name: exit status 1
I0817 19:32:01.909091    6157 dumplogs.go:45] /home/prow/go/src/k8s.io/kops/_rundir/171e15a8-1e62-11ed-a9bf-0eb48dfbe9bd/kops toolbox dump --name e2e-e2e-kops-aws-k8s-1-22.test-cncf-aws.k8s.io --dir /logs/artifacts --private-key /tmp/kops/e2e-e2e-kops-aws-k8s-1-22.test-cncf-aws.k8s.io/id_ed25519 --ssh-user ubuntu
I0817 19:32:01.909137    6157 local.go:42] ⚙️ /home/prow/go/src/k8s.io/kops/_rundir/171e15a8-1e62-11ed-a9bf-0eb48dfbe9bd/kops toolbox dump --name e2e-e2e-kops-aws-k8s-1-22.test-cncf-aws.k8s.io --dir /logs/artifacts --private-key /tmp/kops/e2e-e2e-kops-aws-k8s-1-22.test-cncf-aws.k8s.io/id_ed25519 --ssh-user ubuntu
I0817 19:32:30.645920    6157 dumplogs.go:78] /home/prow/go/src/k8s.io/kops/_rundir/171e15a8-1e62-11ed-a9bf-0eb48dfbe9bd/kops get cluster --name e2e-e2e-kops-aws-k8s-1-22.test-cncf-aws.k8s.io -o yaml
I0817 19:32:30.645959    6157 local.go:42] ⚙️ /home/prow/go/src/k8s.io/kops/_rundir/171e15a8-1e62-11ed-a9bf-0eb48dfbe9bd/kops get cluster --name e2e-e2e-kops-aws-k8s-1-22.test-cncf-aws.k8s.io -o yaml
I0817 19:32:31.155439    6157 dumplogs.go:78] /home/prow/go/src/k8s.io/kops/_rundir/171e15a8-1e62-11ed-a9bf-0eb48dfbe9bd/kops get instancegroups --name e2e-e2e-kops-aws-k8s-1-22.test-cncf-aws.k8s.io -o yaml
I0817 19:32:31.155479    6157 local.go:42] ⚙️ /home/prow/go/src/k8s.io/kops/_rundir/171e15a8-1e62-11ed-a9bf-0eb48dfbe9bd/kops get instancegroups --name e2e-e2e-kops-aws-k8s-1-22.test-cncf-aws.k8s.io -o yaml
... skipping 453 lines ...
route-table:rtb-07326ddc987b23e6a	ok
vpc:vpc-04443c35962ed516f	ok
dhcp-options:dopt-064e0aff382427329	ok
Deleted kubectl config for e2e-e2e-kops-aws-k8s-1-22.test-cncf-aws.k8s.io

Deleted cluster: "e2e-e2e-kops-aws-k8s-1-22.test-cncf-aws.k8s.io"
Error: exit status 255
+ EXIT_VALUE=1
+ set +o xtrace