This job view page is being replaced by Spyglass soon. Check out the new job view.
ResultFAILURE
Tests 1 failed / 2 succeeded
Started2022-08-15 15:31
Elapsed9m47s
Revisionmaster

Test Failures


kubetest2 Test 2.09s

exit status 255
				from junit_runner.xml

Filter through log files | View test history on testgrid


Show 2 Passed Tests

Error lines from build-log.txt

ERROR: (gcloud.auth.activate-service-account) There was a problem refreshing your current auth tokens: ('invalid_grant: Invalid JWT Signature.', {'error': 'invalid_grant', 'error_description': 'Invalid JWT Signature.'})
Please run:

  $ gcloud auth login

to obtain new credentials.

... skipping 171 lines ...
I0815 15:32:53.181884    6201 http.go:37] curl https://storage.googleapis.com/kops-ci/bin/latest-ci-updown-green.txt
I0815 15:32:53.206273    6201 http.go:37] curl https://storage.googleapis.com/kops-ci/bin/1.25.0-alpha.3+v1.25.0-alpha.2-88-gf442cc2d0a/linux/amd64/kops
I0815 15:32:54.537654    6201 local.go:42] ⚙️ ssh-keygen -t ed25519 -N  -q -f /tmp/kops/e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io/id_ed25519
I0815 15:32:54.551324    6201 up.go:44] Cleaning up any leaked resources from previous cluster
I0815 15:32:54.551605    6201 dumplogs.go:45] /home/prow/go/src/k8s.io/kops/_rundir/44cafb4e-1caf-11ed-9820-1e463d3c6692/kops toolbox dump --name e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io --dir /logs/artifacts --private-key /tmp/kops/e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io/id_ed25519 --ssh-user ubuntu
I0815 15:32:54.551761    6201 local.go:42] ⚙️ /home/prow/go/src/k8s.io/kops/_rundir/44cafb4e-1caf-11ed-9820-1e463d3c6692/kops toolbox dump --name e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io --dir /logs/artifacts --private-key /tmp/kops/e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io/id_ed25519 --ssh-user ubuntu
W0815 15:32:55.062471    6201 down.go:34] Dumping cluster logs at the start of Down() failed: exit status 1
I0815 15:32:55.062526    6201 down.go:48] /home/prow/go/src/k8s.io/kops/_rundir/44cafb4e-1caf-11ed-9820-1e463d3c6692/kops delete cluster --name e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io --yes
I0815 15:32:55.062539    6201 local.go:42] ⚙️ /home/prow/go/src/k8s.io/kops/_rundir/44cafb4e-1caf-11ed-9820-1e463d3c6692/kops delete cluster --name e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io --yes
I0815 15:32:55.098478    6235 featureflag.go:164] FeatureFlag "SpecOverrideFlag"=true
Error: error reading cluster configuration: Cluster.kops.k8s.io "e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io" not found
I0815 15:32:55.571672    6201 http.go:37] curl http://metadata.google.internal/computeMetadata/v1/instance/network-interfaces/0/access-configs/0/external-ip
2022/08/15 15:32:55 failed to get external ip from metadata service: http://metadata.google.internal/computeMetadata/v1/instance/network-interfaces/0/access-configs/0/external-ip returned 404
I0815 15:32:55.588537    6201 http.go:37] curl https://ip.jsb.workers.dev
I0815 15:32:55.728275    6201 up.go:159] /home/prow/go/src/k8s.io/kops/_rundir/44cafb4e-1caf-11ed-9820-1e463d3c6692/kops create cluster --name e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io --cloud aws --kubernetes-version https://storage.googleapis.com/kubernetes-release/release/v1.24.3 --ssh-public-key /tmp/kops/e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io/id_ed25519.pub --override cluster.spec.nodePortAccess=0.0.0.0/0 --image=099720109477/ubuntu/images/hvm-ssd/ubuntu-focal-20.04-amd64-server-20220810 --channel=alpha --networking=kube-router --container-runtime=containerd --node-size=t3.large --discovery-store=s3://k8s-kops-prow/discovery --admin-access 34.133.95.99/32 --master-count 1 --master-volume-size 48 --node-count 4 --node-volume-size 48 --zones us-west-1a --master-size c5.large
I0815 15:32:55.728321    6201 local.go:42] ⚙️ /home/prow/go/src/k8s.io/kops/_rundir/44cafb4e-1caf-11ed-9820-1e463d3c6692/kops create cluster --name e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io --cloud aws --kubernetes-version https://storage.googleapis.com/kubernetes-release/release/v1.24.3 --ssh-public-key /tmp/kops/e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io/id_ed25519.pub --override cluster.spec.nodePortAccess=0.0.0.0/0 --image=099720109477/ubuntu/images/hvm-ssd/ubuntu-focal-20.04-amd64-server-20220810 --channel=alpha --networking=kube-router --container-runtime=containerd --node-size=t3.large --discovery-store=s3://k8s-kops-prow/discovery --admin-access 34.133.95.99/32 --master-count 1 --master-volume-size 48 --node-count 4 --node-volume-size 48 --zones us-west-1a --master-size c5.large
I0815 15:32:55.761767    6244 featureflag.go:164] FeatureFlag "SpecOverrideFlag"=true
I0815 15:32:55.780094    6244 create_cluster.go:862] Using SSH public key: /tmp/kops/e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io/id_ed25519.pub
I0815 15:32:56.321097    6244 new_cluster.go:1168]  Cloud Provider ID = aws
... skipping 580 lines ...

I0815 15:33:36.371012    6201 up.go:243] /home/prow/go/src/k8s.io/kops/_rundir/44cafb4e-1caf-11ed-9820-1e463d3c6692/kops validate cluster --name e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io --count 10 --wait 15m0s
I0815 15:33:36.371098    6201 local.go:42] ⚙️ /home/prow/go/src/k8s.io/kops/_rundir/44cafb4e-1caf-11ed-9820-1e463d3c6692/kops validate cluster --name e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io --count 10 --wait 15m0s
I0815 15:33:36.405833    6284 featureflag.go:164] FeatureFlag "SpecOverrideFlag"=true
Validating cluster e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io

W0815 15:33:37.464617    6284 validate_cluster.go:184] (will retry): unexpected error during validation: unable to resolve Kubernetes cluster API URL dns: lookup api.e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io on 10.63.240.10:53: no such host
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.large	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 15:33:47.500521    6284 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.large	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 15:33:57.544409    6284 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.large	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 15:34:07.584536    6284 validate_cluster.go:232] (will retry): cluster not yet healthy
W0815 15:34:17.627622    6284 validate_cluster.go:184] (will retry): unexpected error during validation: unable to resolve Kubernetes cluster API URL dns: lookup api.e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io on 10.63.240.10:53: no such host
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.large	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 15:34:27.676966    6284 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.large	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 15:34:37.727597    6284 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.large	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 15:34:47.766306    6284 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.large	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 15:34:57.815516    6284 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.large	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 15:35:07.855896    6284 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.large	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 15:35:17.909813    6284 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.large	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 15:35:27.951652    6284 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.large	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 15:35:37.988665    6284 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.large	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 15:35:48.032389    6284 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.large	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 15:35:58.082405    6284 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.large	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 15:36:08.121641    6284 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.large	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 15:36:18.176865    6284 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.large	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 15:36:28.227995    6284 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.large	4	4	us-west-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0815 15:36:38.279867    6284 validate_cluster.go:232] (will retry): cluster not yet healthy
W0815 15:37:18.316324    6284 validate_cluster.go:184] (will retry): unexpected error during validation: error listing nodes: Get "https://api.e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io/api/v1/nodes": dial tcp 203.0.113.123:443: i/o timeout
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-us-west-1a	Master	c5.large	1	1	us-west-1a
nodes-us-west-1a	Node	t3.large	4	4	us-west-1a

NODE STATUS
... skipping 140 lines ...
i-06fd58c093520c26d	node	True

Your cluster e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io is ready
I0815 15:39:11.184181    6201 up.go:105] cluster reported as up
I0815 15:39:11.184328    6201 local.go:42] ⚙️ /home/prow/go/bin/kubetest2-tester-kops --ginkgo-args=--debug --test-args=-test.timeout=60m -num-nodes=0 --test-package-marker=stable.txt --parallel=25
I0815 15:39:11.205074    6294 featureflag.go:164] FeatureFlag "SpecOverrideFlag"=true
F0815 15:39:13.270158    6294 tester.go:482] failed to run ginkgo tester: failed to get kubectl package from published releases: failed to get latest release name: exit status 1
I0815 15:39:13.272941    6201 dumplogs.go:45] /home/prow/go/src/k8s.io/kops/_rundir/44cafb4e-1caf-11ed-9820-1e463d3c6692/kops toolbox dump --name e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io --dir /logs/artifacts --private-key /tmp/kops/e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io/id_ed25519 --ssh-user ubuntu
I0815 15:39:13.273001    6201 local.go:42] ⚙️ /home/prow/go/src/k8s.io/kops/_rundir/44cafb4e-1caf-11ed-9820-1e463d3c6692/kops toolbox dump --name e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io --dir /logs/artifacts --private-key /tmp/kops/e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io/id_ed25519 --ssh-user ubuntu
I0815 15:39:39.621574    6201 dumplogs.go:78] /home/prow/go/src/k8s.io/kops/_rundir/44cafb4e-1caf-11ed-9820-1e463d3c6692/kops get cluster --name e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io -o yaml
I0815 15:39:39.621628    6201 local.go:42] ⚙️ /home/prow/go/src/k8s.io/kops/_rundir/44cafb4e-1caf-11ed-9820-1e463d3c6692/kops get cluster --name e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io -o yaml
I0815 15:39:40.141930    6201 dumplogs.go:78] /home/prow/go/src/k8s.io/kops/_rundir/44cafb4e-1caf-11ed-9820-1e463d3c6692/kops get instancegroups --name e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io -o yaml
I0815 15:39:40.141964    6201 local.go:42] ⚙️ /home/prow/go/src/k8s.io/kops/_rundir/44cafb4e-1caf-11ed-9820-1e463d3c6692/kops get instancegroups --name e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io -o yaml
... skipping 263 lines ...
route-table:rtb-0d031449364dd6c17	ok
vpc:vpc-08d55af98818f1562	ok
dhcp-options:dopt-08644366af64d900a	ok
Deleted kubectl config for e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io

Deleted cluster: "e2e-e2e-kops-aws-cni-kuberouter.test-cncf-aws.k8s.io"
Error: exit status 255
+ EXIT_VALUE=1
+ set +o xtrace