This job view page is being replaced by Spyglass soon. Check out the new job view.
ResultFAILURE
Tests 8 failed / 186 succeeded
Started2020-01-14 20:26
Elapsed4h35m
Revision
Buildergke-prow-default-pool-cf4891d4-6stf
links{u'resultstore': {u'url': u'https://source.cloud.google.com/results/invocations/5617ea0f-a58e-4eec-bc32-f9c8726f7ada/targets/test'}}
pod0438039c-370c-11ea-9f20-3687633bf296
resultstorehttps://source.cloud.google.com/results/invocations/5617ea0f-a58e-4eec-bc32-f9c8726f7ada/targets/test
infra-commitdd3ff6d1a
job-versionv1.17.1-beta.0.52+d224476cd0730b
master_os_imagecos-77-12371-89-0
node_os_imageubuntu-gke-1804-d1809-0-v20200114
pod0438039c-370c-11ea-9f20-3687633bf296
revisionv1.17.1-beta.0.52+d224476cd0730b

Test Failures


Kubernetes e2e suite [sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node 1m2s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[sig\-storage\]\sCSI\sVolumes\s\[Driver\:\spd\.csi\.storage\.gke\.io\]\[Serial\]\s\[Testpattern\:\sDynamic\sPV\s\(filesystem\svolmode\)\]\smultiVolume\s\[Slow\]\sshould\sconcurrently\saccess\sthe\ssingle\svolume\sfrom\spods\son\sthe\ssame\snode$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/multivolume.go:293
Jan 14 22:59:42.231: waiting for csi driver node registration on: timed out waiting for the condition
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/drivers/csi.go:452
				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (xfs)][Slow] volumes should allow exec of files on the volume 5m10s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[sig\-storage\]\sCSI\sVolumes\s\[Driver\:\spd\.csi\.storage\.gke\.io\]\[Serial\]\s\[Testpattern\:\sDynamic\sPV\s\(xfs\)\]\[Slow\]\svolumes\sshould\sallow\sexec\sof\sfiles\son\sthe\svolume$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/volumes.go:191
Jan 14 23:18:26.867: Unexpected error:
    <*errors.errorString | 0xc00455aa80>: {
        s: "expected pod \"exec-volume-test-pd-csi-storage-gke-io-dynamicpv-fprq\" success: Gave up after waiting 5m0s for pod \"exec-volume-test-pd-csi-storage-gke-io-dynamicpv-fprq\" to be \"success or failure\"",
    }
    expected pod "exec-volume-test-pd-csi-storage-gke-io-dynamicpv-fprq" success: Gave up after waiting 5m0s for pod "exec-volume-test-pd-csi-storage-gke-io-dynamicpv-fprq" to be "success or failure"
occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/util.go:894
				
				Click to see stdout/stderrfrom junit_01.xml

Find exec-volume-test-pd-csi-storage-gke-io-dynamicpv-fprq mentions in log files | View test history on testgrid


Kubernetes e2e suite [sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (xfs)][Slow] volumes should store data 10m5s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[sig\-storage\]\sCSI\sVolumes\s\[Driver\:\spd\.csi\.storage\.gke\.io\]\[Serial\]\s\[Testpattern\:\sDynamic\sPV\s\(xfs\)\]\[Slow\]\svolumes\sshould\sstore\sdata$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/volumes.go:150
Jan 15 00:49:11.687: Failed to create injector pod: timed out waiting for the condition
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/volume/fixtures.go:597
				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly] 2.62s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[sig\-storage\]\sIn\-tree\sVolumes\s\[Driver\:\shostPathSymlink\]\s\[Testpattern\:\sInline\-volume\s\(default\sfs\)\]\ssubPath\sshould\sunmount\sif\spod\sis\sforce\sdeleted\swhile\skubelet\sis\sdown\s\[Disruptive\]\[Slow\]\[LinuxOnly\]$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:347
Jan 14 21:19:21.038: while waiting for hostPath init pod to succeed
Unexpected error:
    <*errors.errorString | 0xc001d54920>: {
        s: "pod \"hostpath-symlink-prep-provisioning-7968\" failed with status: {Phase:Failed Conditions:[] Message:Pod Node didn't have enough resource: pods, requested: 1, used: 110, capacity: 110 Reason:OutOfpods NominatedNodeName: HostIP: PodIP: PodIPs:[] StartTime:2020-01-14 21:19:20 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[] QOSClass: EphemeralContainerStatuses:[]}",
    }
    pod "hostpath-symlink-prep-provisioning-7968" failed with status: {Phase:Failed Conditions:[] Message:Pod Node didn't have enough resource: pods, requested: 1, used: 110, capacity: 110 Reason:OutOfpods NominatedNodeName: HostIP: PodIP: PodIPs:[] StartTime:2020-01-14 21:19:20 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[] QOSClass: EphemeralContainerStatuses:[]}
occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/drivers/in_tree.go:883
				
				Click to see stdout/stderrfrom junit_01.xml

Find to mentions in log files | View test history on testgrid


Kubernetes e2e suite [sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. 2.59s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[sig\-storage\]\sIn\-tree\sVolumes\s\[Driver\:\slocal\]\[LocalVolumeType\:\sblock\]\s\[Testpattern\:\sPre\-provisioned\sPV\s\(block\svolmode\)\]\sdisruptive\[Disruptive\]\sShould\stest\sthat\spv\sused\sin\sa\spod\sthat\sis\sforce\sdeleted\swhile\sthe\skubelet\sis\sdown\scleans\sup\swhen\sthe\skubelet\sreturns\.$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/disruptive.go:149
Jan 14 21:19:29.099: Unexpected error:
    <*errors.errorString | 0xc000440570>: {
        s: "pod ran to completion",
    }
    pod ran to completion
occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/host_exec.go:103
				
				Click to see stdout/stderrfrom junit_01.xml

Find ran mentions in log files | View test history on testgrid


Kubernetes e2e suite [sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] 3.97s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[sig\-storage\]\sIn\-tree\sVolumes\s\[Driver\:\slocal\]\[LocalVolumeType\:\sgce\-localssd\-scsi\-fs\]\s\[Serial\]\s\[Testpattern\:\sPre\-provisioned\sPV\s\(default\sfs\)\]\ssubPath\sshould\sfail\sif\ssubpath\swith\sbackstepping\sis\soutside\sthe\svolume\s\[Slow\]$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:273
Jan 14 21:19:26.601: Unexpected error:
    <*errors.errorString | 0xc000440570>: {
        s: "pod ran to completion",
    }
    pod ran to completion
occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/host_exec.go:103
				
				Click to see stdout/stderrfrom junit_01.xml

Find ran mentions in log files | View test history on testgrid


Kubernetes e2e suite [sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly] 1.70s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[sig\-storage\]\sIn\-tree\sVolumes\s\[Driver\:\slocal\]\[LocalVolumeType\:\sgce\-localssd\-scsi\-fs\]\s\[Serial\]\s\[Testpattern\:\sPre\-provisioned\sPV\s\(default\sfs\)\]\ssubPath\sshould\sunmount\sif\spod\sis\sgracefully\sdeleted\swhile\skubelet\sis\sdown\s\[Disruptive\]\[Slow\]\[LinuxOnly\]$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:340
Jan 14 21:19:22.867: Unexpected error:
    <*errors.errorString | 0xc000440570>: {
        s: "pod ran to completion",
    }
    pod ran to completion
occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/host_exec.go:103
				
				Click to see stdout/stderrfrom junit_01.xml

Find ran mentions in log files | View test history on testgrid


Test 4h18m

error during ./hack/ginkgo-e2e.sh --ginkgo.focus=\[Serial\]|\[Disruptive\] --ginkgo.skip=\[Flaky\]|\[Feature:.+\] --minStartupPods=8 --report-dir=/workspace/_artifacts --disable-log-dump=true: exit status 1
				from junit_runner.xml

Filter through log files | View test history on testgrid


Show 186 Passed Tests

Show 4662 Skipped Tests